2 #============================================================= -*-perl-*-
4 # BackupPC_tarIncCreate: create a tar archive of an existing incremental dump
9 # Usage: BackupPC_tarIncCreate [options] files/directories...
14 # -h host Host from which the tar archive is created.
15 # -n dumpNum Dump number from which the tar archive is created.
16 # A negative number means relative to the end (eg -1
17 # means the most recent dump, -2 2nd most recent etc).
18 # -s shareName Share name from which the tar archive is created.
21 # -t print summary totals
22 # -r pathRemove path prefix that will be replaced with pathAdd
23 # -p pathAdd new path prefix
24 # -b BLOCKS BLOCKS x 512 bytes per record (default 20; same as tar)
25 # -w writeBufSz write buffer size (default 1MB)
27 # The -h, -n and -s options specify which dump is used to generate
28 # the tar archive. The -r and -p options can be used to relocate
29 # the paths in the tar archive so extracted files can be placed
30 # in a location different from their original location.
33 # Craig Barratt <cbarratt@users.sourceforge.net>
36 # Copyright (C) 2001-2003 Craig Barratt
38 # This program is free software; you can redistribute it and/or modify
39 # it under the terms of the GNU General Public License as published by
40 # the Free Software Foundation; either version 2 of the License, or
41 # (at your option) any later version.
43 # This program is distributed in the hope that it will be useful,
44 # but WITHOUT ANY WARRANTY; without even the implied warranty of
45 # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
46 # GNU General Public License for more details.
48 # You should have received a copy of the GNU General Public License
49 # along with this program; if not, write to the Free Software
50 # Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
52 #========================================================================
54 # Version 2.1.0, released 20 Jun 2004.
56 # See http://backuppc.sourceforge.net.
58 #========================================================================
62 use lib "__INSTALLDIR__/lib";
67 use BackupPC::Attrib qw(:all);
68 use BackupPC::FileZIO;
70 use BackupPC::SearchLib;
73 die("BackupPC::Lib->new failed\n") if ( !(my $bpc = BackupPC::Lib->new) );
74 my $TopDir = $bpc->TopDir();
75 my $BinDir = $bpc->BinDir();
76 my %Conf = $bpc->Conf();
81 if ( !getopts("th:n:p:r:s:b:w:", \%opts) || @ARGV < 1 ) {
83 usage: $0 [options] files/directories...
85 -h host host from which the tar archive is created
86 -n dumpNum dump number from which the tar archive is created
87 A negative number means relative to the end (eg -1
88 means the most recent dump, -2 2nd most recent etc).
89 -s shareName share name from which the tar archive is created
92 -t print summary totals
93 -r pathRemove path prefix that will be replaced with pathAdd
94 -p pathAdd new path prefix
95 -b BLOCKS BLOCKS x 512 bytes per record (default 20; same as tar)
96 -w writeBufSz write buffer size (default 1048576 = 1MB)
101 if ( $opts{h} !~ /^([\w\.\s-]+)$/ ) {
102 print(STDERR "$0: bad host name '$opts{h}'\n");
107 if ( $opts{n} !~ /^(-?\d+)$/ ) {
108 print(STDERR "$0: bad dump number '$opts{n}'\n");
113 my @Backups = $bpc->BackupInfoRead($Host);
121 $Num = $Backups[@Backups + $Num]{num} if ( -@Backups <= $Num && $Num < 0 );
122 for ( $i = 0 ; $i < @Backups ; $i++ ) {
123 last if ( $Backups[$i]{num} == $Num );
125 if ( $i >= @Backups ) {
126 print(STDERR "$0: bad backup number $Num for host $Host\n");
130 my $PathRemove = $1 if ( $opts{r} =~ /(.+)/ );
131 my $PathAdd = $1 if ( $opts{p} =~ /(.+)/ );
132 if ( $opts{s} !~ /^([\w\s\.\/\$-]+)$/ && $opts{s} ne "*" ) {
133 print(STDERR "$0: bad share name '$opts{s}'\n");
136 our $ShareName = $opts{s};
137 our $view = BackupPC::View->new($bpc, $Host, \@Backups);
140 # This constant and the line of code below that uses it are borrowed
141 # from Archive::Tar. Thanks to Calle Dybedahl and Stephen Zander.
144 # Archive::Tar is Copyright 1997 Calle Dybedahl. All rights reserved.
145 # Copyright 1998 Stephen Zander. All rights reserved.
148 = 'a100 a8 a8 a8 a12 a12 A8 a1 a100 a6 a2 a32 a32 a8 a8 a155 x12';
149 my $tar_header_length = 512;
151 my $BufSize = $opts{w} || 1048576; # 1MB or 2^20
153 my $WriteBufSz = ($opts{b} || 20) * $tar_header_length;
155 my(%UidCache, %GidCache);
156 my(%HardLinkExtraFiles, @HardLinks);
159 # Write out all the requested files/directories
163 if ( $ShareName eq "*" ) {
164 my $PathRemoveOrig = $PathRemove;
165 my $PathAddOrig = $PathAdd;
166 foreach $ShareName ( $view->shareList($Num) ) {
167 #print(STDERR "Doing share ($ShareName)\n");
168 $PathRemove = "/" if ( !defined($PathRemoveOrig) );
169 ($PathAdd = "/$ShareName/$PathAddOrig") =~ s{//+}{/}g;
170 foreach my $dir ( @ARGV ) {
171 archiveWrite($fh, $dir);
173 archiveWriteHardLinks($fh);
176 foreach my $dir ( @ARGV ) {
177 archiveWrite($fh, $dir);
179 archiveWriteHardLinks($fh);
183 # Finish with two null 512 byte headers, and then round out a full
186 my $data = "\0" x ($tar_header_length * 2);
187 TarWrite($fh, \$data);
188 TarWrite($fh, undef);
191 # print out totals if requested
194 print STDERR "Done: $FileCnt files, $ByteCnt bytes, $DirCnt dirs,",
195 " $SpecialCnt specials, $ErrorCnt errors\n";
197 if ( $ErrorCnt && !$FileCnt && !$DirCnt ) {
199 # Got errors, with no files or directories; exit with non-zero
206 ###########################################################################
208 ###########################################################################
212 my($fh, $dir, $tarPathOverride) = @_;
214 if ( $dir =~ m{(^|/)\.\.(/|$)} ) {
215 print(STDERR "$0: bad directory '$dir'\n");
219 $dir = "/" if ( $dir eq "." );
220 #print(STDERR "calling find with $Num, $ShareName, $dir\n");
222 if ( $view->find($Num, $ShareName, $dir, 0, \&TarWriteFile,
223 $fh, $tarPathOverride) < 0 ) {
224 print(STDERR "$0: bad share or directory '$ShareName/$dir'\n");
231 # Write out any hardlinks (if any)
233 sub archiveWriteHardLinks
236 foreach my $hdr ( @HardLinks ) {
238 if ( defined($PathRemove)
239 && substr($hdr->{linkname}, 0, length($PathRemove)+1)
240 eq ".$PathRemove" ) {
241 substr($hdr->{linkname}, 0, length($PathRemove)+1) = ".$PathAdd";
243 TarWriteFileInfo($fh, $hdr);
246 %HardLinkExtraFiles = ();
253 $UidCache{$uid} = (getpwuid($uid))[0] if ( !exists($UidCache{$uid}) );
254 return $UidCache{$uid};
261 $GidCache{$gid} = (getgrgid($gid))[0] if ( !exists($GidCache{$gid}) );
262 return $GidCache{$gid};
267 my($fh, $dataRef) = @_;
269 if ( !defined($dataRef) ) {
271 # do flush by padding to a full $WriteBufSz
273 my $data = "\0" x ($WriteBufSz - length($WriteBuf));
276 if ( length($WriteBuf) + length($$dataRef) < $WriteBufSz ) {
278 # just buffer and return
280 $WriteBuf .= $$dataRef;
283 my $done = $WriteBufSz - length($WriteBuf);
284 if ( syswrite($fh, $WriteBuf . substr($$dataRef, 0, $done))
286 print(STDERR "Unable to write to output file ($!)\n");
289 while ( $done + $WriteBufSz <= length($$dataRef) ) {
290 if ( syswrite($fh, substr($$dataRef, $done, $WriteBufSz))
292 print(STDERR "Unable to write to output file ($!)\n");
295 $done += $WriteBufSz;
297 $WriteBuf = substr($$dataRef, $done);
304 if ( $size % $tar_header_length ) {
305 my $data = "\0" x ($tar_header_length - ($size % $tar_header_length));
306 TarWrite($fh, \$data);
314 $hdr->{uname} = UidLookup($hdr->{uid}) if ( !defined($hdr->{uname}) );
315 $hdr->{gname} = GidLookup($hdr->{gid}) if ( !defined($hdr->{gname}) );
316 my $devmajor = defined($hdr->{devmajor}) ? sprintf("%07o", $hdr->{devmajor})
318 my $devminor = defined($hdr->{devminor}) ? sprintf("%07o", $hdr->{devminor})
321 if ( $hdr->{size} >= 2 * 65536 * 65536 ) {
323 # GNU extension for files >= 8GB: send size in big-endian binary
325 $sizeStr = pack("c4 N N", 0x80, 0, 0, 0,
326 $hdr->{size} / (65536 * 65536),
327 $hdr->{size} % (65536 * 65536));
328 } elsif ( $hdr->{size} >= 1 * 65536 * 65536 ) {
330 # sprintf octal only handles up to 2^32 - 1
332 $sizeStr = sprintf("%03o", $hdr->{size} / (1 << 24))
333 . sprintf("%08o", $hdr->{size} % (1 << 24));
335 $sizeStr = sprintf("%011o", $hdr->{size});
337 my $data = pack($tar_pack_header,
338 substr($hdr->{name}, 0, 99),
339 sprintf("%07o", $hdr->{mode}),
340 sprintf("%07o", $hdr->{uid}),
341 sprintf("%07o", $hdr->{gid}),
343 sprintf("%011o", $hdr->{mtime}),
344 "", #checksum field - space padded by pack("A8")
346 substr($hdr->{linkname}, 0, 99),
347 $hdr->{magic} || 'ustar ',
348 $hdr->{version} || ' ',
355 substr($data, 148, 7) = sprintf("%06o\0", unpack("%16C*",$data));
356 TarWrite($fh, \$data);
364 # Handle long link names (symbolic links)
366 if ( length($hdr->{linkname}) > 99 ) {
368 my $data = $hdr->{linkname} . "\0";
369 $h{name} = "././\@LongLink";
371 $h{size} = length($data);
372 TarWriteHeader($fh, \%h);
373 TarWrite($fh, \$data);
374 TarWritePad($fh, length($data));
377 # Handle long file names
379 if ( length($hdr->{name}) > 99 ) {
381 my $data = $hdr->{name} . "\0";
382 $h{name} = "././\@LongLink";
384 $h{size} = length($data);
385 TarWriteHeader($fh, \%h);
386 TarWrite($fh, \$data);
387 TarWritePad($fh, length($data));
389 TarWriteHeader($fh, $hdr);
393 # returns 1 if a given directory has files somewhere under it
394 # in a given dump of a given share
396 sub checkSubDirs($$$$) {
397 my ($dir, $share, $host, $dumpNo) = @_;
399 my $dsn = $Conf{SearchDSN};
400 my $db_user = $Conf{SearchUser} || '';
405 if (substr($dir, 0, 1) == '.')
407 $dir = substr($dir, 1, length($dir));
410 if (substr($dir, 0, 1) == '/')
412 $dir = substr($dir, 1, length($dir));
415 if (substr($dir, length($dir)-1, 1) == '/')
417 $dir = substr($dir, 0, length($dir)-1);
422 print STDERR "doing db...";
424 SELECT hosts.name, shares.name, startfiles.name, COUNT(files.*) AS subfiles
425 FROM files startfiles
426 INNER JOIN shares ON (shares.id=startfiles.shareid)
427 INNER JOIN hosts ON (hosts.id=shares.hostid)
428 INNER JOIN backups ON (
429 backups.num=startfiles.backupnum AND
430 backups.hostid=hosts.id AND backups.shareid=shares.id
433 files.backupnum=startfiles.backupnum AND
434 files.shareid=startfiles.shareid AND
435 files.path LIKE startfiles.path || '/%' AND
436 files.type<>startfiles.type AND
437 files.id <> startfiles.id
442 startfiles.type=? AND
443 startfiles.backupnum=?
444 GROUP BY hosts.name, shares.name, startfiles.name, startfiles.backupnum;
446 my $dbh = DBI->connect($dsn, $db_user, "", { RaiseError => 1, AutoCommit => 1} );
447 my $sth = $dbh->prepare($search_sql);
448 $sth->execute($host, $share, BPC_FTYPE_DIR, $dumpNo);
449 print STDERR "done\n";
450 while (my @r_data = $sth->fetchrow_array())
452 $DBCache[$r_data[0]][$r_data[1]][$r_data[2]] = 1;
457 $DBCache[$host][$share][$dir] = $ret;
462 if ($DBCache[$host][$share][$dir] != undef && $DBCache[$host][$share][$dir] == 1)
474 my($hdr, $fh, $tarPathOverride) = @_;
477 my $tarPath = $hdr->{relPath};
478 $tarPath = $tarPathOverride if ( defined($tarPathOverride) );
480 $tarPath =~ s{//+}{/}g;
481 if ( defined($PathRemove)
482 && substr($tarPath, 0, length($PathRemove)) eq $PathRemove ) {
483 substr($tarPath, 0, length($PathRemove)) = $PathAdd;
485 $tarPath = "./" . $tarPath if ( $tarPath !~ /^\.\// );
486 $tarPath =~ s{//+}{/}g;
487 $hdr->{name} = $tarPath;
489 if ( $hdr->{type} == BPC_FTYPE_DIR ) {
491 # Directory: just write the header
495 $hdr->{name} .= "/" if ( $hdr->{name} !~ m{/$} );
496 # check if it has files under it in the database
497 if ( checkSubDirs($hdr->{path}, $ShareName, $Host, $Num) != 0 )
499 TarWriteFileInfo($fh, $hdr);
502 } elsif ( $hdr->{type} == BPC_FTYPE_FILE ) {
504 # Regular file: write the header and file
506 my $f = BackupPC::FileZIO->open($hdr->{fullPath}, 0, $hdr->{compress});
507 if ( !defined($f) ) {
508 print(STDERR "Unable to open file $hdr->{fullPath}\n");
512 TarWriteFileInfo($fh, $hdr);
514 while ( $f->read(\$data, $BufSize) > 0 ) {
515 TarWrite($fh, \$data);
516 $size += length($data);
519 TarWritePad($fh, $size);
522 } elsif ( $hdr->{type} == BPC_FTYPE_HARDLINK ) {
524 # Hardlink file: either write a hardlink or the complete file
525 # depending upon whether the linked-to file will be written
528 # Start by reading the contents of the link.
530 my $f = BackupPC::FileZIO->open($hdr->{fullPath}, 0, $hdr->{compress});
531 if ( !defined($f) ) {
532 print(STDERR "Unable to open file $hdr->{fullPath}\n");
537 while ( $f->read(\$data, $BufSize) > 0 ) {
538 $hdr->{linkname} .= $data;
542 # Check @ARGV and the list of hardlinked files we have explicity
543 # dumped to see if we have dumped this file or not
546 my $name = $hdr->{linkname};
548 if ( $HardLinkExtraFiles{$name} ) {
551 foreach my $arg ( @ARGV ) {
554 $done = 1 if ( $name eq $arg || $name =~ /^\Q$arg\// );
559 # Target file will be or was written, so just remember
560 # the hardlink so we can dump it later.
562 push(@HardLinks, $hdr);
566 # Have to dump the original file. Just call the top-level
567 # routine, so that we save the hassle of dealing with
568 # mangling, merging and attributes.
570 $HardLinkExtraFiles{$hdr->{linkname}} = 1;
571 archiveWrite($fh, $hdr->{linkname}, $hdr->{name});
573 } elsif ( $hdr->{type} == BPC_FTYPE_SYMLINK ) {
575 # Symbolic link: read the symbolic link contents into the header
576 # and write the header.
578 my $f = BackupPC::FileZIO->open($hdr->{fullPath}, 0, $hdr->{compress});
579 if ( !defined($f) ) {
580 print(STDERR "Unable to open symlink file $hdr->{fullPath}\n");
585 while ( $f->read(\$data, $BufSize) > 0 ) {
586 $hdr->{linkname} .= $data;
590 TarWriteFileInfo($fh, $hdr);
592 } elsif ( $hdr->{type} == BPC_FTYPE_CHARDEV
593 || $hdr->{type} == BPC_FTYPE_BLOCKDEV
594 || $hdr->{type} == BPC_FTYPE_FIFO ) {
596 # Special files: for char and block special we read the
597 # major and minor numbers from a plain file.
599 if ( $hdr->{type} != BPC_FTYPE_FIFO ) {
600 my $f = BackupPC::FileZIO->open($hdr->{fullPath}, 0,
603 if ( !defined($f) || $f->read(\$data, $BufSize) < 0 ) {
604 print(STDERR "Unable to open/read char/block special file"
605 . " $hdr->{fullPath}\n");
606 $f->close if ( defined($f) );
611 if ( $data =~ /(\d+),(\d+)/ ) {
612 $hdr->{devmajor} = $1;
613 $hdr->{devminor} = $2;
617 TarWriteFileInfo($fh, $hdr);
620 print(STDERR "Got unknown type $hdr->{type} for $hdr->{name}\n");