| # Copyright (C) all contributors <meta@public-inbox.org> |
| # License: AGPL-3.0+ <https://www.gnu.org/licenses/agpl-3.0.txt> |
| package PublicInbox::Xapcmd; |
| use v5.12; |
| use autodie qw(chmod opendir rename syswrite); |
| use PublicInbox::Spawn qw(which popen_rd); |
| use PublicInbox::Syscall; |
| use PublicInbox::Admin qw(setup_signals); |
| use PublicInbox::Over; |
| use PublicInbox::Search qw(xap_terms); |
| use PublicInbox::SearchIdx; |
| use File::Temp 0.19 (); # ->newdir |
| use File::Path qw(remove_tree); |
| use POSIX qw(WNOHANG _exit); |
| use PublicInbox::DS; |
| |
| # support testing with dev versions of Xapian which installs |
| # commands with a version number suffix (e.g. "xapian-compact-1.5") |
| our $XAPIAN_COMPACT = $ENV{XAPIAN_COMPACT} || 'xapian-compact'; |
| our @COMPACT_OPT = qw(jobs|j=i quiet|q blocksize|b=s no-full|n fuller|F); |
| |
| sub commit_changes ($$$$) { |
| my ($ibx, $im, $tmp, $opt) = @_; |
| my $reshard = $opt->{reshard}; |
| |
| $SIG{INT} or die 'BUG: $SIG{INT} not handled'; |
| my (@old_shard, $over_chg); |
| |
| # Sort shards highest-to-lowest, since ->xdb_shards_flat |
| # determines the number of shards to load based on the max; |
| # and we'd rather xdb_shards_flat to momentarily fail rather |
| # than load out-of-date shards |
| my @order = sort { |
| my ($x) = ($a =~ m!/([0-9]+)/*\z!); |
| my ($y) = ($b =~ m!/([0-9]+)/*\z!); |
| ($y // -1) <=> ($x // -1) # we may have non-shards |
| } keys %$tmp; |
| |
| my ($dname) = ($order[0] =~ m!(.*/)[^/]+/*\z!); |
| my $mode = (stat($dname))[2]; |
| for my $old (@order) { |
| next if $old eq ''; # no invalid paths |
| my $newdir = $tmp->{$old}; |
| my $have_old = -e $old; |
| if (!$have_old && !defined($opt->{reshard})) { |
| die "failed to stat($old): $!"; |
| } |
| |
| my $new = $newdir->dirname if defined($newdir); |
| my $over = "$old/over.sqlite3"; |
| if (-f $over) { # only for v1, v2 over is untouched |
| defined $new or die "BUG: $over exists when culling v2"; |
| $over = PublicInbox::Over->new($over); |
| my $tmp_over = "$new/over.sqlite3"; |
| $over->dbh->sqlite_backup_to_file($tmp_over); |
| $over = undef; |
| $over_chg = 1; |
| } |
| |
| if (!defined($new)) { # culled shard |
| push @old_shard, $old; |
| next; |
| } |
| |
| chmod $mode & 07777, $new; |
| rename $old, "$new/old" if $have_old; |
| rename $new, $old; |
| push @old_shard, "$old/old" if $have_old; |
| } |
| |
| # trigger ->check_inodes in read-only daemons |
| syswrite($im->{lockfh}, '.') if $over_chg && $im; |
| |
| remove_tree(@old_shard); |
| $tmp = undef; |
| if (!$opt->{-coarse_lock}) { |
| $opt->{-skip_lock} = 1; |
| $im //= $ibx if $ibx->can('eidx_sync') || $ibx->can('cidx_run'); |
| if ($im->can('count_shards')) { # v2w, eidx, cidx |
| my $pr = $opt->{-progress}; |
| my $n = $im->count_shards; |
| if (defined $reshard && $n != $reshard) { |
| die |
| "BUG: counted $n shards after resharding to $reshard"; |
| } |
| my $prev = $im->{shards} // $ibx->{nshard}; |
| if ($pr && $prev != $n) { |
| $pr->("shard count changed: $prev => $n\n"); |
| $im->{shards} = $n; |
| } |
| } |
| my $env = $opt->{-idx_env}; |
| local %ENV = (%ENV, %$env) if $env; |
| if ($ibx->can('eidx_sync')) { |
| $ibx->eidx_sync($opt); |
| } elsif (!$ibx->can('cidx_run')) { |
| PublicInbox::Admin::index_inbox($ibx, $im, $opt); |
| } |
| } |
| } |
| |
| sub cb_spawn { |
| my ($cb, $args, $opt) = @_; # $cb = cpdb() or compact() |
| my $pid = PublicInbox::DS::fork_persist; |
| return $pid if $pid > 0; |
| $SIG{__DIE__} = sub { warn @_; _exit(1) }; # don't jump up stack |
| $cb->($args, $opt); |
| _exit(0); |
| } |
| |
| sub runnable_or_die ($) { |
| my ($exe) = @_; |
| which($exe) or die "$exe not found in PATH\n"; |
| } |
| |
| sub prepare_reindex ($$) { |
| my ($ibx, $opt) = @_; |
| if ($ibx->can('eidx_sync') || $ibx->can('cidx_run')) { |
| # no prep needed for ExtSearchIdx nor CodeSearchIdx |
| } elsif ($ibx->version == 1) { |
| my $dir = $ibx->search->xdir(1); |
| my $xdb = $PublicInbox::Search::X{Database}->new($dir); |
| if (my $lc = $xdb->get_metadata('last_commit')) { |
| $opt->{reindex}->{from} = $lc; |
| } |
| } else { # v2 |
| my $max = $ibx->max_git_epoch // return; |
| my $from = $opt->{reindex}->{from}; |
| my $mm = $ibx->mm; |
| my $v = PublicInbox::Search::SCHEMA_VERSION(); |
| foreach my $i (0..$max) { |
| $from->[$i] = $mm->last_commit_xap($v, $i); |
| } |
| } |
| } |
| |
| sub same_fs_or_die ($$) { |
| my ($x, $y) = @_; |
| return if ((stat($x))[0] == (stat($y))[0]); # 0 - st_dev |
| die "$x and $y reside on different filesystems\n"; |
| } |
| |
| sub kill_pids { |
| my ($sig, $pids) = @_; |
| kill($sig, keys %$pids); # pids may be empty |
| } |
| |
| sub process_queue { |
| my ($queue, $task, $opt) = @_; |
| my $max = $opt->{jobs} // scalar(@$queue); |
| my $cb = \&$task; |
| if ($max <= 1) { |
| while (defined(my $args = shift @$queue)) { |
| $cb->($args, $opt); |
| } |
| return; |
| } |
| |
| # run in parallel: |
| my %pids; |
| local @SIG{keys %SIG} = values %SIG; |
| setup_signals(\&kill_pids, \%pids); |
| while (@$queue) { |
| while (scalar(keys(%pids)) < $max && scalar(@$queue)) { |
| my $args = shift @$queue; |
| $pids{cb_spawn($cb, $args, $opt)} = $args; |
| } |
| |
| my $flags = 0; |
| while (scalar keys %pids) { |
| my $pid = waitpid(-1, $flags) or last; |
| last if $pid < 0; |
| my $args = delete $pids{$pid}; |
| if ($args) { |
| die join(' ', @$args)." failed: $?\n" if $?; |
| } else { |
| warn "unknown PID($pid) reaped: $?\n"; |
| } |
| $flags = WNOHANG if scalar(@$queue); |
| } |
| } |
| } |
| |
| sub prepare_run { |
| my ($ibx, $opt) = @_; |
| my $tmp = {}; # old shard dir => File::Temp->newdir object or undef |
| my @queue; # ([old//src,newdir]) - list of args for cpdb() or compact() |
| my ($old, $misc_ok); |
| if ($ibx->can('cidx_run')) { |
| $old = $ibx->xdir(1); |
| } elsif ($ibx->can('eidx_sync')) { |
| $misc_ok = 1; |
| $old = $ibx->xdir(1); |
| } elsif (my $srch = $ibx->search) { |
| $old = $srch->xdir(1); |
| } |
| if (defined $old) { |
| -d $old or die "$old does not exist\n"; |
| } |
| my $reshard = $opt->{reshard}; |
| if (defined $reshard && $reshard <= 0) { |
| die "--reshard must be a positive number\n"; |
| } |
| |
| # we want temporary directories to be as deep as possible, |
| # so v2 shards can keep "xap$SCHEMA_VERSION" on a separate FS. |
| if (defined($old) && $ibx->can('version') && $ibx->version == 1) { |
| if (defined $reshard) { |
| warn |
| "--reshard=$reshard ignored for v1 $ibx->{inboxdir}\n"; |
| } |
| my ($dir) = ($old =~ m!(.*?/)[^/]+/*\z!); |
| same_fs_or_die($dir, $old); |
| my $v = PublicInbox::Search::SCHEMA_VERSION(); |
| my $wip = File::Temp->newdir("xapian$v-XXXX", DIR => $dir); |
| $tmp->{$old} = $wip; |
| PublicInbox::Syscall::nodatacow_dir($wip->dirname); |
| push @queue, [ $old, $wip ]; |
| } elsif (defined $old) { |
| opendir(my $dh, $old); |
| my @old_shards; |
| while (defined(my $dn = readdir($dh))) { |
| if ($dn =~ /\A[0-9]+\z/) { |
| push(@old_shards, $dn + 0); |
| } elsif ($dn eq '.' || $dn eq '..') { |
| } elsif ($dn =~ /\Aover\.sqlite3/) { |
| } elsif ($dn eq 'misc' && $misc_ok) { |
| } else { |
| warn "W: skipping unknown dir: $old/$dn\n" |
| } |
| } |
| die "No Xapian shards found in $old\n" unless @old_shards; |
| @old_shards = sort { $a <=> $b } @old_shards; |
| my ($src, $max_shard); |
| if (!defined($reshard) || $reshard == scalar(@old_shards)) { |
| # 1:1 copy |
| $max_shard = scalar(@old_shards) - 1; |
| } else { |
| # M:N copy |
| $max_shard = $reshard - 1; |
| $src = [ map { "$old/$_" } @old_shards ]; |
| } |
| foreach my $dn (0..$max_shard) { |
| my $wip = File::Temp->newdir("$dn-XXXX", DIR => $old); |
| same_fs_or_die($old, $wip->dirname); |
| my $cur = "$old/$dn"; |
| push @queue, [ $src // $cur , $wip ]; |
| PublicInbox::Syscall::nodatacow_dir($wip->dirname); |
| $tmp->{$cur} = $wip; |
| } |
| # mark old shards to be unlinked |
| if ($src) { |
| $tmp->{$_} ||= undef for @$src; |
| } |
| } |
| ($tmp, \@queue); |
| } |
| |
| sub check_compact () { runnable_or_die($XAPIAN_COMPACT) } |
| |
| sub run { |
| my ($ibx, $task, $opt) = @_; # task = 'cpdb' or 'compact' |
| PublicInbox::Admin::progress_prepare($opt ||= {}); |
| my $dir; |
| for my $fld (qw(inboxdir topdir cidx_dir)) { |
| my $d = $ibx->{$fld} // next; |
| -d $d or die "$fld=$d does not exist\n"; |
| $dir = $d; |
| last; |
| } |
| check_compact() if $opt->{compact} && |
| ($ibx->can('cidx_run') || $ibx->search); |
| |
| if (!$ibx->can('eidx_sync') && $ibx->can('version') && |
| !$opt->{-coarse_lock}) { |
| # per-epoch ranges for v2 |
| # v1:{ from => $OID }, v2:{ from => [ $OID, $OID, $OID ] } } |
| $opt->{reindex} = { from => $ibx->version == 1 ? '' : [] }; |
| PublicInbox::SearchIdx::load_xapian_writable(); |
| } |
| |
| local @SIG{keys %SIG} = values %SIG; |
| setup_signals(); |
| my $restore = $ibx->with_umask; |
| |
| my $im = $ibx->can('importer') ? $ibx->importer(0) : undef; |
| ($im // $ibx)->lock_acquire; |
| my ($tmp, $queue) = prepare_run($ibx, $opt); |
| |
| # fine-grained locking if we prepare for reindex |
| if (!$opt->{-coarse_lock}) { |
| prepare_reindex($ibx, $opt); |
| ($im // $ibx)->lock_release; |
| } |
| |
| $ibx->cleanup if $ibx->can('cleanup'); |
| if ($task eq 'cpdb' && $opt->{reshard} && $ibx->can('cidx_run')) { |
| cidx_reshard($ibx, $queue, $opt); |
| } else { |
| process_queue($queue, $task, $opt); |
| } |
| ($im // $ibx)->lock_acquire if !$opt->{-coarse_lock}; |
| commit_changes($ibx, $im, $tmp, $opt); |
| } |
| |
| sub cpdb_retryable ($$) { |
| my ($src, $pfx) = @_; |
| if (ref($@) =~ /\bDatabaseModifiedError\b/) { |
| warn "$pfx Xapian DB modified, reopening and retrying\n"; |
| $src->reopen; |
| return 1; |
| } |
| if ($@) { |
| warn "$pfx E: ", ref($@), "\n"; |
| die; |
| } |
| 0; |
| } |
| |
| sub progress_pfx ($) { |
| my ($wip) = @_; # tempdir v2: ([0-9])+-XXXX |
| my @p = split(m'/', $wip); |
| |
| # "basename(inboxdir)/xap15/0" for v2, |
| # "basename(inboxdir)/xapian15" for v1: |
| ($p[-1] =~ /\A([0-9]+)/) ? "$p[-3]/$p[-2]/$1" : "$p[-2]/$p[-1]"; |
| } |
| |
| sub kill_compact { # setup_signals callback |
| my ($sig, $ioref) = @_; |
| kill($sig, $$ioref->attached_pid // return) if defined($$ioref); |
| } |
| |
| # xapian-compact wrapper |
| sub compact ($$) { # cb_spawn callback |
| my ($args, $opt) = @_; |
| my ($src, $newdir) = @$args; |
| my $dst = ref($newdir) ? $newdir->dirname : $newdir; |
| my $pfx = $opt->{-progress_pfx} ||= progress_pfx($src); |
| my $pr = $opt->{-progress}; |
| my $rdr = {}; |
| |
| foreach my $fd (0..2) { |
| defined(my $dfd = $opt->{$fd}) or next; |
| $rdr->{$fd} = $dfd; |
| } |
| |
| # we rely on --no-renumber to keep docids synched to NNTP |
| my $cmd = [ $XAPIAN_COMPACT, '--no-renumber' ]; |
| for my $sw (qw(no-full fuller)) { |
| push @$cmd, "--$sw" if $opt->{$sw}; |
| } |
| for my $sw (qw(blocksize)) { |
| defined(my $v = $opt->{$sw}) or next; |
| push @$cmd, "--$sw", $v; |
| } |
| $pr->("$pfx `".join(' ', @$cmd)."'\n") if $pr; |
| push @$cmd, $src, $dst; |
| local @SIG{keys %SIG} = values %SIG; |
| setup_signals(\&kill_compact, \my $rd); |
| $rd = popen_rd($cmd, undef, $rdr); |
| while (<$rd>) { |
| if ($pr) { |
| s/\r/\r$pfx /g; |
| $pr->("$pfx $_"); |
| } |
| } |
| $rd->close or die "@$cmd failed: \$?=$?\n"; |
| } |
| |
| sub cpdb_loop ($$$;$$) { |
| my ($src, $dst, $pr_data, $cur_shard, $reshard) = @_; |
| my ($pr, $fmt, $nr, $pfx); |
| if ($pr_data) { |
| $pr = $pr_data->{pr}; |
| $fmt = $pr_data->{fmt}; |
| $nr = \($pr_data->{nr}); |
| $pfx = $pr_data->{pfx}; |
| } |
| |
| my ($it, $end); |
| do { |
| eval { |
| $it = $src->postlist_begin(''); |
| $end = $src->postlist_end(''); |
| }; |
| } while (cpdb_retryable($src, $pfx)); |
| |
| do { |
| eval { |
| for (; $it != $end; $it++) { |
| my $docid = $it->get_docid; |
| if (defined $reshard) { |
| my $dst_shard = $docid % $reshard; |
| next if $dst_shard != $cur_shard; |
| } |
| my $doc = $src->get_document($docid); |
| $dst->replace_document($docid, $doc); |
| if ($pr_data && !(++$$nr & 1023)) { |
| $pr->(sprintf($fmt, $$nr)); |
| } |
| } |
| |
| # unlike copydatabase(1), we don't copy spelling |
| # and synonym data (or other user metadata) since |
| # the Perl APIs don't expose iterators for them |
| # (and public-inbox does not use those features) |
| }; |
| } while (cpdb_retryable($src, $pfx)); |
| } |
| |
| sub xapian_write_prep ($) { |
| my ($opt) = @_; |
| PublicInbox::SearchIdx::load_xapian_writable(); |
| my $flag = eval($PublicInbox::Search::Xap.'::DB_CREATE()'); |
| die if $@; |
| $flag |= $PublicInbox::SearchIdx::DB_NO_SYNC if !$opt->{fsync}; |
| (\%PublicInbox::Search::X, $flag); |
| } |
| |
| sub compact_tmp_shard ($) { |
| my ($wip) = @_; |
| my $new = $wip->dirname; |
| my ($dir) = ($new =~ m!(.*?/)[^/]+/*\z!); |
| same_fs_or_die($dir, $new); |
| my $ft = File::Temp->newdir("$new.compact-XXXX", DIR => $dir); |
| PublicInbox::Syscall::nodatacow_dir($ft->dirname); |
| $ft; |
| } |
| |
| sub cidx_reshard { # not docid based |
| my ($cidx, $queue, $opt) = @_; |
| my ($X, $flag) = xapian_write_prep($opt); |
| my $src = $cidx->xdb; |
| delete($cidx->{xdb}) == $src or die "BUG: xdb != $src"; |
| my $pfx = $opt->{-progress_pfx} = progress_pfx($cidx->xdir.'/0'); |
| my $pr = $opt->{-progress}; |
| my $pr_data = { pr => $pr, pfx => $pfx, nr => 0 } if $pr; |
| local @SIG{keys %SIG} = values %SIG; |
| |
| # like copydatabase(1), be sure we don't overwrite anything in case |
| # of other bugs: |
| setup_signals() if $opt->{compact}; |
| my @tmp; |
| my @dst = map { |
| my $wip = $_->[1]; |
| my $tmp = $opt->{compact} ? compact_tmp_shard($wip) : $wip; |
| push @tmp, $tmp; |
| $X->{WritableDatabase}->new($tmp->dirname, $flag); |
| } @$queue; |
| my $l = $src->get_metadata('indexlevel'); |
| $dst[0]->set_metadata('indexlevel', $l) if $l eq 'medium'; |
| my $fmt; |
| if ($pr_data) { |
| my $tot = $src->get_doccount; |
| $fmt = "$pfx % ".length($tot)."u/$tot\n"; |
| $pr->("$pfx copying $tot documents\n"); |
| } |
| my $cur = $src->postlist_begin(''); |
| my $end = $src->postlist_end(''); |
| my $git_dir_hash = $cidx->can('git_dir_hash'); |
| my ($n, $nr); |
| for (; $cur != $end; $cur++) { |
| my $doc = $src->get_document($cur->get_docid); |
| if (my @cmt = xap_terms('Q', $doc)) { |
| $n = hex(substr($cmt[0], 0, 8)) % scalar(@dst); |
| warn "W: multi-commit: @cmt" if scalar(@cmt) != 1; |
| } elsif (my @P = xap_terms('P', $doc)) { |
| $n = $git_dir_hash->($P[0]) % scalar(@dst); |
| warn "W: multi-path @P " if scalar(@P) != 1; |
| } else { |
| warn "W: skipped, no terms in ".$cur->get_docid; |
| next; |
| } |
| $dst[$n]->add_document($doc); |
| $pr->(sprintf($fmt, $nr)) if $pr_data && !(++$nr & 1023); |
| } |
| return if !$opt->{compact}; |
| $src = undef; |
| @dst = (); # flushes and closes |
| my @q; |
| for my $tmp (@tmp) { |
| my $arg = shift @$queue // die 'BUG: $queue empty'; |
| my $wip = $arg->[1] // die 'BUG: no $wip'; |
| push @q, [ "$tmp", $wip ]; |
| } |
| delete $opt->{-progress_pfx}; |
| process_queue(\@q, 'compact', $opt); |
| } |
| |
| # Like copydatabase(1), this is horribly slow; and it doesn't seem due |
| # to the overhead of Perl. |
| sub cpdb ($$) { # cb_spawn callback |
| my ($args, $opt) = @_; |
| my ($old, $wip) = @$args; |
| my ($src, $cur_shard); |
| my $reshard; |
| my ($X, $flag) = xapian_write_prep($opt); |
| if (ref($old) eq 'ARRAY') { |
| my $new = $wip->dirname; |
| ($cur_shard) = ($new =~ m!(?:xap|ei)[0-9]+/([0-9]+)\b!); |
| defined $cur_shard or |
| die "BUG: could not extract shard # from $new"; |
| $reshard = $opt->{reshard}; |
| defined $reshard or die 'BUG: got array src w/o --reshard'; |
| |
| # resharding, M:N copy means have full read access |
| foreach (@$old) { |
| if ($src) { |
| my $sub = $X->{Database}->new($_); |
| $src->add_database($sub); |
| } else { |
| $src = $X->{Database}->new($_); |
| } |
| } |
| } else { |
| $src = $X->{Database}->new($old); |
| } |
| |
| my $tmp = $wip; |
| local @SIG{keys %SIG} = values %SIG; |
| if ($opt->{compact}) { |
| $tmp = compact_tmp_shard($wip); |
| setup_signals(); |
| } |
| |
| # like copydatabase(1), be sure we don't overwrite anything in case |
| # of other bugs: |
| my $new = $wip->dirname; |
| my $dst = $X->{WritableDatabase}->new($tmp->dirname, $flag); |
| my $pr = $opt->{-progress}; |
| my $pfx = $opt->{-progress_pfx} = progress_pfx($new); |
| my $pr_data = { pr => $pr, pfx => $pfx, nr => 0 } if $pr; |
| |
| do { |
| eval { |
| # update the only metadata key for v1: |
| my $lc = $src->get_metadata('last_commit'); |
| $dst->set_metadata('last_commit', $lc) if $lc; |
| |
| # only the first xapian shard (0) gets 'indexlevel' |
| if ($new =~ m!/(?:xapian[0-9]+|(?:ei|xap)[0-9]+/0)\b!) { |
| my $l = $src->get_metadata('indexlevel'); |
| $l eq 'medium' and |
| $dst->set_metadata('indexlevel', $l); |
| } |
| if ($pr_data) { |
| my $tot = $src->get_doccount; |
| |
| # we can only estimate when resharding, |
| # because removed spam causes slight imbalance |
| my $est = ''; |
| if (defined $cur_shard && $reshard > 1) { |
| $tot = int($tot/$reshard); |
| $est = 'around '; |
| } |
| my $fmt = "$pfx % ".length($tot)."u/$tot\n"; |
| $pr->("$pfx copying $est$tot documents\n"); |
| $pr_data->{fmt} = $fmt; |
| $pr_data->{total} = $tot; |
| } |
| }; |
| } while (cpdb_retryable($src, $pfx)); |
| |
| if (defined $reshard) { |
| # we rely on document IDs matching NNTP article number, |
| # so we can't have the Xapian sharding DB support rewriting |
| # document IDs. Thus we iterate through each shard |
| # individually. |
| $src = undef; |
| foreach (@$old) { |
| my $old = $X->{Database}->new($_); |
| cpdb_loop($old, $dst, $pr_data, $cur_shard, $reshard); |
| } |
| } else { |
| cpdb_loop($src, $dst, $pr_data); |
| } |
| |
| $pr->(sprintf($pr_data->{fmt}, $pr_data->{nr})) if $pr; |
| return unless $opt->{compact}; |
| |
| $src = $dst = undef; # flushes and closes |
| |
| # this is probably the best place to do xapian-compact |
| # since $dst isn't readable by HTTP or NNTP clients, yet: |
| compact([ $tmp, $new ], $opt); |
| } |
| |
| 1; |