X-Git-Url: http://git.rot13.org/?p=BackupPC.git;a=blobdiff_plain;f=bin%2FBackupPC_ASA_PostArchive_Update;h=50d9cc902a94dc3579a7da16b0dbff2e9d86c237;hp=33527a0ab84be8aff3532b029b5713c4fc23b768;hb=1ad5ae30debf935221d2a2bb36289a87cb604a18;hpb=acf1846a475e06c0926e729046d128ae770b20c7 diff --git a/bin/BackupPC_ASA_PostArchive_Update b/bin/BackupPC_ASA_PostArchive_Update index 33527a0..50d9cc9 100755 --- a/bin/BackupPC_ASA_PostArchive_Update +++ b/bin/BackupPC_ASA_PostArchive_Update @@ -13,8 +13,8 @@ use POSIX qw/strftime/; use Cwd qw/abs_path/; use Archive::Tar::Streamed; use Algorithm::Diff; -use Getopt::Std; use File::Slurp; +use Getopt::Long::Descriptive; =head1 NAME @@ -22,43 +22,28 @@ BackupPC_ASA_PostArchive_Update =head1 DESCRIPTION - # /etc/BackupPC/pc/dvd_tar.pl + # /etc/BackupPC/pc/_search_archive.pl =cut -# FIXME -my $debug = $ENV{DEBUG} || 1; -my $check = $ENV{CHECK} || 1; - my $bpc = BackupPC::Lib->new || die "can't create BackupPC::Lib"; +$bpc->ConfigRead('_search_archive'); # read our configuration my %Conf = $bpc->Conf(); -warn "## ARGV=",dump @ARGV; - - -my $args; -my $name; -foreach ( @ARGV ) { - my $v = $_; - if ( m/(\w+)=(.+)/ ) { - $name = $1; - $v = $2; - } - if ( $name =~ m/List/ ) { - push @{ $args->{$name} }, $v; - } else { - $args->{$name} = $v; - } -} - -warn "args = ",dump($args); - use BackupPC::Search; %BackupPC::Search::Conf = %Conf; -my $path = abs_path($0); -$path =~ s{/[^/]+$}{/}; # FIXME remove? +my ($opt,$usage) = describe_options( +"%c %o", +[ 'host|h=s@', "import just host(s)" ], +[ 'num|n=s@', "import just backup number(s)" ], +[ 'check|c', "check archives on disk and sync", { default => 1 } ], +[ 'debug|d', "debug", { default => 1 } ], +[ 'help', "show help" ], +); + +print($usage->text), exit if $opt->help; $|=1; @@ -66,7 +51,7 @@ my $start_t = time(); my $t_fmt = '%Y-%m-%d %H:%M:%S'; -warn "## Conf = ",dump( \%Conf ); +#warn "## Conf = ",dump( \%Conf ); my $dbh = DBI->connect($Conf{SearchDSN}, $Conf{SearchUser}, "", { RaiseError => 1, AutoCommit => 0 }); @@ -107,7 +92,7 @@ sub get_backup_id($$) { $hsn_cache->{"$host $num"} = $id; - print STDERR "# $host $num == $id\n" if $debug; + print STDERR "# $host $num == $id\n" if $opt->debug; return $id; } @@ -137,15 +122,18 @@ my $sth_inc_size = $dbh->prepare(qq{ sub check_archive { my ($host,$num) = @_; + warn "# check_archive $host $num"; my $t = time(); - my @tar_parts = - sort map { s/^\Q$Conf{ArchiveDest}\E\/*//; $_ } - glob "$Conf{ArchiveDest}/$host.$num.*" - ; + my $glob = "$Conf{ArchiveDest}/$host.$num.*"; - return unless @tar_parts; + my @tar_parts = sort map { s/^\Q$Conf{ArchiveDest}\E\/*//; $_ } glob $glob ; + + if ( ! @tar_parts ) { + warn "ERROR: no files for $glob"; + return; + } print curr_time, " check $host $num"; @@ -155,7 +143,7 @@ sub check_archive { if ( ! -e $md5_path ) { system_ok "cd $Conf{ArchiveDest} && /usr/bin/md5sum $host.$num.* > $md5_path"; } else { - system_ok "cd $Conf{ArchiveDest} && /usr/bin/md5sum -c $md5_path" if $check; + system_ok "cd $Conf{ArchiveDest} && /usr/bin/md5sum -c $md5_path" if $opt->check; } my $md5sum; @@ -176,34 +164,35 @@ sub check_archive { } open(my $gzip, $Conf{GzipPath}." -l $filename |") || die "can't gzip -l $filename: $!"; + local $/ = undef; my $line = <$gzip>; - chomp($line); - $line = <$gzip> if ($line =~ /^\s+compressed/); + close($gzip); my ($comp, $uncomp) = (0,0); - if ($line =~ m/^\s+(\d+)\s+(\d+)\s+\d+\.\d+/) { + if ($line =~ m/\s+(\d+)\s+(\d+)\s+\d+\.\d+/s) { if (wantarray) { return [ $1, $2 ]; } else { return $2; } } else { - die "can't find size in line: $line"; + warn "ERROR can't parse: $line"; + return -s $filename; } } sub check_part { - my ($host, $num, $part_nr, $tar_size, $size, $md5, $items) = @_; + my ($host, $num, $part_nr, $tar_size, $size, $md5, $items, $filename) = @_; my $backup_id = get_backup_id($host, $num); my $sth_md5 = $dbh->prepare(qq{ select - id, tar_size, size, md5, items + id, tar_size, size, md5, items, filename from backup_parts - where backup_id = ? and part_nr = ? + where backup_id = ? and part_nr = ? and filename = ? }); - $sth_md5->execute($backup_id, $part_nr); + $sth_md5->execute($backup_id, $part_nr, $filename); if (my $row = $sth_md5->fetchrow_hashref) { return if ( @@ -223,21 +212,22 @@ sub check_archive { tar_size, size, md5, - items - ) values (?,?,?,?,?,?) + items, + filename + ) values (?,?,?,?,?,?,?) }); - $sth_insert->execute($backup_id, $part_nr, $tar_size, $size, $md5, $items); + $sth_insert->execute($backup_id, $part_nr, $tar_size, $size, $md5, $items, $filename); $dbh->commit; } - print " [parts: ",join(", ", @tar_parts),"]" if $debug; + print " [parts: ",join(", ", @tar_parts),"]" if $opt->debug; my @tar_files; my $backup_part; - print " reading" if $debug; + print " reading" if $opt->debug; my $part_nr = 0; my $inc_size = 0; @@ -253,8 +243,8 @@ sub check_archive { my $size = (stat( $path ))[7] || die "can't stat $path: $!"; - if ($size > $Conf{MaxArchiveSize}) { - print ", part bigger than media $size > $Conf{MaxArchiveSize}\n"; + if ($size > $Conf{ArchiveMediaSize}) { + print ", part bigger than media $size > $Conf{ArchiveMediaSize}\n"; return 0; } @@ -274,8 +264,8 @@ sub check_archive { $items++; $tar_size_inarc += $entry->size; - if ($tar_size_inarc > $Conf{MaxArchiveFileSize}) { - print ", part $filename is too big $tar_size_inarc > $Conf{MaxArchiveFileSize}\n"; + if ($tar_size_inarc > $Conf{ArchiveChunkSize}) { + print ", part $filename is too big $tar_size_inarc > $Conf{ArchiveChunkSize}\n"; return 0; } @@ -309,10 +299,10 @@ sub check_archive { my $items = 1; $part_nr++; - check_part($host, $num, $part_nr, $tar_size, $size, $md5, $items); + check_part($host, $num, $part_nr, $tar_size, $size, $md5, $items, $filename); # round increment size to 2k block size - $inc_size += int(($size + 2048) / 2048); + $inc_size += int((($size + 2048) / 2048 ) * 2048); } $sth_inc_size->execute( @@ -370,16 +360,16 @@ sub check_archive { #----- main -foreach ( 0 .. $#{ $args->{HostList} } ) { +foreach ( 0 .. $#{ $opt->host } ) { - my $host = $args->{'HostList'}->[$_]; - my $num = $args->{'BackupList'}->[$_]; + my $host = $opt->host->[$_]; + my $num = $opt->num->[$_]; check_archive $host => $num; } -exit; +exit; # FIXME my $sth = $dbh->prepare( qq{ @@ -416,10 +406,10 @@ while (my $row = $sth->fetchrow_hashref) { # this will return -1 if file doesn't exist my $size = BackupPC::Search::get_tgz_size_by_name($tar_file); - print "# host: ".$row->{host}.", share: ".$row->{'share'}.", backup_num:".$row->{num}." size: $size backup.size: ", $row->{inc_size},"\n" if $debug; + print "# host: ".$row->{host}.", share: ".$row->{'share'}.", backup_num:".$row->{num}." size: $size backup.size: ", $row->{inc_size},"\n" if $opt->debug; if ( $row->{'inc_size'} != -1 && $size != -1 && $row->{'inc_size'} >= $size && $row->{parts} == $row->{backup_parts}) { - if ($check) { + if ($opt->check) { tar_check($row->{'host'}, $row->{'share'}, $row->{'num'}, $tar_file) && next; } else { next; @@ -434,7 +424,7 @@ while (my $row = $sth->fetchrow_hashref) { =for later # re-create archive? my $cmd = qq[ $tarIncCreate -h "$row->{host}" -s "$row->{share}" -n $row->{num} -f ]; - print STDERR "## $cmd\n" if ($debug); + print STDERR "## $cmd\n" if ($opt->debug); if (system($cmd) != 0) { print STDERR " FAILED, marking this backup deleted";