| #!/usr/bin/env perl |
| use strict; |
| use warnings; |
| use v5.10; |
| use FindBin; |
| BEGIN { unshift @INC, "$FindBin::Bin/../lib" }; |
| use File::Spec::Functions qw/catfile catdir/; |
| use Getopt::Long qw/GetOptions :config no_auto_abbrev/; |
| use Benchmark qw/:hireswallclock/; |
| use IO::Compress::Gzip qw/$GzipError/; |
| use POSIX qw/ceil/; |
| use Log::Any qw($log); |
| use Log::Any::Adapter; |
| use Pod::Usage; |
| use Cache::FastMmap; |
| use Directory::Iterator; |
| use KorAP::XML::Krill qw!get_file_name get_file_name_from_glob!; |
| use KorAP::XML::Archive; |
| use KorAP::XML::Tokenizer; |
| use KorAP::XML::Batch::File; |
| use Config::Simple; |
| use Parallel::ForkManager; |
| use File::Glob ':bsd_glob'; |
| use File::Temp qw/tempdir/; |
| use File::Path qw(remove_tree make_path); |
| use File::Basename; |
| use Mojo::Collection 'c'; |
| use String::Random qw(random_string); |
| use IO::File; |
| use Archive::Tar::Builder; |
| use Fcntl qw(:flock SEEK_END); |
| |
| # use KorAP::XML::ForkPool; |
| # TODO: use Parallel::Loops |
| # TODO: make output files |
| |
| # TODO: Use KorAP::XML::ForkPool! |
| |
| # CHANGES: |
| # ---------------------------------------------------------- |
| # 2013/11/25 |
| # - Initial release |
| # |
| # 2014/10/29 |
| # - Merges foundry data to create indexer friendly documents |
| # |
| # 2016/02/04 |
| # - renamed to korapxml2krill |
| # - added Schreibgebrauch support |
| # |
| # 2016/02/12 |
| # - fixed foundry skipping |
| # - Support overwrite in archive processing |
| # |
| # 2016/02/14 |
| # - Added version information |
| # - Added support for archive files |
| # |
| # 2016/02/15 |
| # - Fixed temporary directory bug |
| # - Improved skipping before unzipping |
| # - Added EXPERIMENTAL concurrency support |
| # |
| # 2016/02/23 |
| # - Merge korapxml2krill and korapxml2krill_dir |
| # |
| # 2016/02/27 |
| # - Added extract function |
| # |
| # 2016/03/17 |
| # - Added meta switch |
| # |
| # 2016/03/18 |
| # - Added meta data caching |
| # |
| # 2016/06/27 |
| # - Added multi archive support |
| # - Added prefix negation support |
| # - Added Malt#Dependency support |
| # |
| # 2016/07/06 |
| # - Added MDParser#Dependency |
| # |
| # 2016/10/15 |
| # - Fixed temporary path issue in script |
| # |
| # 2016/10/24 |
| # - Improved Windows support |
| # |
| # 2016/10/24 |
| # - Added support for document extraction |
| # |
| # 2016/10/27 |
| # - Added wildcard support for document extraction |
| # |
| # 2016/12/21 |
| # - added support for base-sentences and base-tokenizations |
| # |
| # 2017/01/20 |
| # - added support for DRuKoLa annotations |
| # |
| # 2017/02/08 |
| # - added support for pagebreak annotations |
| # |
| # 2017/04/06 |
| # - added support for wildcards in input |
| # |
| # 2017/04/07 |
| # - support configuration option |
| # - support for temporary extraction |
| # |
| # 2017/04/12 |
| # - support serial processing |
| # - support input root |
| # - introduced --sequential-extraction flag |
| # |
| # 2017/06/19 |
| # - added support for DCK |
| # |
| # 2017/06/29 |
| # - Fixed exit codes |
| # |
| # 2017/07/04 |
| # - Fixed tar building process |
| # |
| # 2018/01/16 |
| # - Added LWC support |
| # |
| # 2018/07/19 |
| # - Preliminary support for HNC. |
| # |
| # 2019/01/22 |
| # - Preliminary support for DGD. |
| # - Support for non-word tokens. |
| # |
| # 2019/02/13 |
| # - Support for 'koral:field' array. |
| # - Support for Koral versioning. |
| # - Ignore temporary extract parameter on |
| # directory archiving. |
| # |
| # 2019/08/08 |
| # - Support for Talismane. |
| # |
| # 2019/12/17 |
| # - Added support for DGD pseudo-sentences |
| # based on anchor milestones. |
| # - Support for non-verbal annotations. |
| # |
| # 2020/04/23 |
| # - Added support for Redewiedergabe-Korpus structure |
| # annotations, based on sentence and paragraph milestones |
| # - Added support for Redewiedergabe-Korpus morphology |
| # ---------------------------------------------------------- |
| |
| our $LAST_CHANGE = '2020/10/21'; |
| our $LOCAL = $FindBin::Bin; |
| our $KORAL_VERSION = 0.03; |
| our $VERSION_MSG = <<"VERSION"; |
| Version $KorAP::XML::Krill::VERSION - diewald\@ids-mannheim.de - $LAST_CHANGE |
| VERSION |
| |
| # Parse comand |
| my $cmd; |
| our @ARGV; |
| if ($ARGV[0] && index($ARGV[0], '-') != 0) { |
| $cmd = shift @ARGV; |
| }; |
| my @keep_argv = @ARGV; |
| |
| my (@skip, @sigle, @anno, @input); |
| |
| # Configuration hash |
| my %cfg = (); |
| |
| # Parse options from the command line |
| GetOptions( |
| 'input|i=s' => \@input, |
| 'input-base|ib=s' => \($cfg{input_base}), |
| 'output|o=s' => \($cfg{output}), |
| 'overwrite|w' => \($cfg{overwrite}), |
| 'meta|m=s' => \($cfg{meta}), |
| 'token|t=s' => \($cfg{token}), |
| 'base-sentences|bs=s' => \($cfg{base_sentences}), |
| 'base-paragraphs|bp=s' => \($cfg{base_paragraphs}), |
| 'base-pagebreaks|bpb=s' => \($cfg{base_pagebreaks}), |
| 'gzip|z' => \($cfg{gzip}), |
| 'temporary-extract|te=s' => \($cfg{extract_dir}), |
| 'skip|s=s' => \@skip, |
| 'sigle|sg=s' => \@sigle, |
| 'cache|c=s' => \($cfg{cache_file}), |
| 'config|cfg=s' => \(my $cfg_file), |
| 'log|l=s' => \($cfg{log}), |
| 'anno|a=s' => \@anno, |
| 'primary|p!' => sub { |
| warn 'Primary flag no longer supported!'; |
| }, |
| 'pretty|y' => sub { |
| warn 'Pretty flag no longer supported!'; |
| }, |
| 'jobs|j=i' => \($cfg{jobs}), |
| 'koral|k=f' => \($cfg{koral}), |
| 'to-tar' => \($cfg{to_tar}), |
| 'non-word-tokens|nwt' => \($cfg{non_word_tokens}), |
| 'non-verbal-tokens|nvt' => \($cfg{non_verbal_tokens}), |
| 'sequential-extraction|se' => \($cfg{sequential_extraction}), |
| 'cache-size|cs=s' => \($cfg{cache_size}), |
| 'cache-delete|cd!' => \($cfg{cache_delete}), |
| 'cache-init|ci!' => \($cfg{cache_init}), |
| 'help|h' => sub { |
| pod2usage( |
| -sections => 'NAME|SYNOPSIS|ARGUMENTS|OPTIONS', |
| -verbose => 99, |
| -msg => $VERSION_MSG, |
| -output => '-' |
| ); |
| }, |
| 'version|v' => sub { |
| pod2usage( |
| -verbose => 0, |
| -msg => $VERSION_MSG, |
| -output => '-' |
| ) |
| } |
| ); |
| |
| my %ERROR_HASH = ( |
| -sections => 'NAME|SYNOPSIS|ARGUMENTS|OPTIONS', |
| -verbose => 99, |
| -msg => $VERSION_MSG, |
| -output => '-', |
| -exit => 1 |
| ); |
| |
| # Load from configuration and fill non-given data |
| if ($cfg_file && -e $cfg_file) { |
| my %config; |
| |
| print "Reading config from $cfg_file\n"; |
| |
| Config::Simple->import_from($cfg_file, \%config); |
| |
| foreach (qw!output cache-size input-base token overwrite |
| meta base-sentences base-paragraphs base-pagebreaks |
| gzip to-tar log cache non-word-tokens |
| non-verbal-tokens sequential-extraction cache-init |
| koral extract-dir jobs!) { |
| my $underlined = $_ =~ tr/-/_/r; |
| if (!defined($cfg{$underlined}) && defined $config{$_}) { |
| $cfg{$underlined} = $config{$_}; |
| }; |
| }; |
| |
| # Skip |
| if (!scalar(@skip) && defined $config{'skip'}) { |
| @skip = split /\s*;\s*/, $config{'skip'} ; |
| }; |
| |
| # Sigle |
| if (!scalar(@sigle) && defined $config{'sigle'}) { |
| @sigle = split /\s*;\s*/, $config{'sigle'} ; |
| }; |
| |
| # Anno |
| if (!scalar(@anno) && defined $config{'anno'}) { |
| @anno = split /\s*;\s*/, $config{'anno'} ; |
| }; |
| }; |
| |
| # Init variables and set default values |
| my $output = $cfg{output}; |
| my $input_base = $cfg{input_base}; |
| my $gzip = $cfg{gzip}; |
| my $to_tar = $cfg{to_tar}; |
| my $extract_dir = $cfg{extract_dir}; |
| my $token_base = $cfg{token} // 'OpenNLP#tokens'; |
| my $cache_file = $cfg{cache} // 'korapxml2krill.cache'; |
| my $jobs = $cfg{jobs} // 0; |
| my $cache_delete = $cfg{cache_delete} // 1; |
| my $base_sentences = lc($cfg{base_sentences} // ''); |
| my $base_paragraphs = lc($cfg{base_paragraphs} // ''); |
| my $base_pagebreaks = lc($cfg{base_pagebreaks} // ''); |
| my $sequential_extraction = $cfg{sequential_extraction} // 0; |
| |
| # Get tokenization basis |
| my ($token_base_foundry, $token_base_layer) = split(/#/, $token_base) if $token_base; |
| |
| # Remove file extension |
| $token_base_layer =~ s/\.xml$//i; |
| |
| # Convert sigle to path construct |
| s!^\s*([^_]+?)_([^\.]+?)\.(.+?)\s*$!$1/$2/$3! foreach @sigle; |
| |
| my %skip; |
| $skip{lc($_)} = 1 foreach @skip; |
| |
| Log::Any::Adapter->set( |
| 'Stderr', log_level => uc($cfg{log} // 'ERROR') |
| ); |
| |
| if ($cmd && $output && (!defined($to_tar)) && (!-e $output || !-d $output)) { |
| $log->error("Directory '$output' does not exist."); |
| exit 1; |
| }; |
| |
| # Input has to be defined |
| pod2usage(%ERROR_HASH) unless @input; |
| |
| # Gzip has no effect, if no output is given |
| pod2usage(%ERROR_HASH) if $gzip && !$output; |
| |
| |
| # Start serial processing |
| if ($cmd && $cmd eq 'serial') { |
| |
| # Remove all inputs |
| my $remove_next = 0; |
| @keep_argv = @{c(@keep_argv)->grep( |
| sub { |
| # Input flag |
| if ($_ eq '-i' || $_ eq '--input' || $_ eq '--output' || $_ eq '-o') { |
| $remove_next = 1; |
| return 0; |
| } |
| |
| # input value |
| elsif ($remove_next) { |
| $remove_next = 0; |
| return 0; |
| }; |
| |
| # Pass parameter |
| return 1; |
| } |
| )->to_array}; |
| |
| |
| # Iterate over all inputs |
| foreach (@input) { |
| |
| # This will create a directory |
| my $new_out = catdir($output, get_file_name_from_glob($_)); |
| |
| # Create new path, in case the output is not meant to be tarred |
| unless ($to_tar) { |
| if (make_path($new_out) == 0 && !-d $new_out) { |
| $log->error("Can\'t create path $new_out"); |
| exit 1; |
| }; |
| }; |
| |
| # Create archive command |
| my @archive_cmd = ($^X, $0, 'archive', @keep_argv, '-i', $_, '-o', $new_out); |
| print "Start serial processing of $_ to $new_out\n"; |
| |
| # Start archiving |
| system @archive_cmd; |
| }; |
| |
| exit; |
| }; |
| |
| # Define supported (and preinstalled) transformation modules |
| my @layers = (); |
| push(@layers, ['Base', 'Sentences']) unless $base_sentences; |
| push(@layers, ['Base', 'Paragraphs']) unless $base_paragraphs; |
| |
| # Connexor |
| push(@layers, ['Connexor', 'Morpho'], |
| ['Connexor', 'Syntax'], |
| ['Connexor', 'Phrase'], |
| ['Connexor', 'Sentences']); |
| |
| # CoreNLP |
| push(@layers, |
| ['CoreNLP', 'NamedEntities'], |
| ['CoreNLP', 'Sentences'], |
| ['CoreNLP', 'Morpho'], |
| ['CoreNLP', 'Constituency']); |
| |
| # CMC |
| push(@layers, ['CMC', 'Morpho']); |
| |
| # DeReKo |
| my @dereko_attr = (); |
| if ($base_sentences eq 'dereko#structure') { |
| push @dereko_attr, 'sentences'; |
| }; |
| if ($base_paragraphs eq 'dereko#structure') { |
| push @dereko_attr, 'paragraphs'; |
| }; |
| |
| if ($base_pagebreaks eq 'dereko#structure') { |
| push @dereko_attr, 'pagebreaks'; |
| }; |
| |
| if ($dereko_attr[0]) { |
| push(@layers, ['DeReKo', 'Structure', 'base-' . join('-', @dereko_attr)]); |
| } |
| else { |
| push(@layers, ['DeReKo', 'Structure']); |
| }; |
| |
| # DGD |
| push(@layers, ['DGD', 'Morpho']); |
| if ($base_sentences eq 'dgd#structure') { |
| push(@layers, ['DGD', 'Structure', 'base-sentence']); |
| } |
| |
| # DRuKoLa |
| push(@layers, |
| ['DRuKoLa', 'Morpho']); |
| |
| # Glemm |
| push(@layers, |
| ['Glemm', 'Morpho']); |
| |
| # HNC |
| push(@layers, |
| ['HNC', 'Morpho']); |
| |
| # LWC |
| push(@layers, |
| ['LWC', 'Dependency']); |
| |
| # Malt |
| push(@layers, |
| ['Malt', 'Dependency']); |
| |
| # Marmot |
| push(@layers, |
| ['MarMoT', 'Morpho']); |
| |
| # Mate |
| push(@layers, |
| ['Mate', 'Morpho'], |
| ['Mate', 'Dependency']); |
| |
| # MDParser |
| push(@layers, |
| ['MDParser', 'Dependency']); |
| |
| # OpenNLP |
| push(@layers, |
| ['OpenNLP', 'Morpho'], |
| ['OpenNLP', 'Sentences']); |
| |
| # Redewiedergabe |
| push(@layers, ['RWK', 'Morpho']); |
| if ($base_sentences eq 'rwk#structure') { |
| push(@layers, ['RWK', 'Structure']); |
| }; |
| |
| # Schreibgebrauch |
| push(@layers, |
| ['Sgbr', 'Lemma'], |
| ['Sgbr', 'Morpho']); |
| |
| # Talismane |
| push(@layers, |
| ['Talismane', 'Dependency'], |
| ['Talismane', 'Morpho']); |
| |
| # TreeTagger |
| push(@layers, |
| ['TreeTagger', 'Morpho'], |
| ['TreeTagger', 'Sentences']); |
| |
| # XIP |
| push(@layers, |
| ['XIP', 'Morpho'], |
| ['XIP', 'Constituency'], |
| ['XIP', 'Sentences'], |
| ['XIP', 'Dependency']); |
| |
| |
| # Check filters |
| my @filtered_anno; |
| if ($skip{'#all'}) { |
| foreach (@anno) { |
| push @filtered_anno, [ split('#', $_) ]; |
| }; |
| } |
| |
| # Add all annotations that are not skipped |
| else { |
| # Add to index file - respect skipping |
| foreach my $info (@layers) { |
| # Skip if Foundry or Foundry#Layer should be skipped |
| unless ($skip{lc($info->[0])} || $skip{lc($info->[0]) . '#' . lc($info->[1])}) { |
| push @filtered_anno, $info; |
| }; |
| }; |
| }; |
| |
| |
| # TODO: This should not be initialized for batch |
| my $cache = Cache::FastMmap->new( |
| share_file => $cache_file, |
| cache_size => ($cfg{cache_size} // '50m'), |
| init_file => ($cfg{cache_init} // 1) |
| ); |
| |
| # Create batch object |
| my $batch_file = KorAP::XML::Batch::File->new( |
| cache => $cache, |
| meta_type => $cfg{meta}, |
| overwrite => $cfg{overwrite}, |
| foundry => $token_base_foundry, |
| layer => $token_base_layer, |
| gzip => $gzip, |
| log => $log, |
| koral => ($cfg{koral} // $KORAL_VERSION), |
| anno => \@filtered_anno, |
| non_word_tokens => ($cfg{non_word_tokens} // 0), |
| non_verbal_tokens => ($cfg{non_verbal_tokens} // 0) |
| ); |
| |
| |
| # Auto adjust jobs |
| if ($jobs eq '-1') { |
| my $cores = 1; |
| if (eval("use Sys::Info; 1;") && eval("use Sys::Info::Constants qw( :device_cpu ); 1;")) { |
| $cores = Sys::Info->new->device('CPU')->count; |
| } |
| else { |
| $log->warn("Unable to determine number of cores"); |
| }; |
| |
| $jobs = ceil(5 * $cores); |
| $log->info("Run using $jobs jobs on $cores cores"); |
| }; |
| |
| |
| # Glob and prefix files |
| if (@input > 0) { |
| |
| my @new_input = (); |
| |
| # Iterate over all inputs |
| foreach my $wild_card (@input) { |
| |
| # Prefix with input root |
| $wild_card = $input_base ? catfile($input_base, $wild_card) : $wild_card; |
| |
| push (@new_input, bsd_glob($wild_card)); |
| }; |
| |
| # Sort files by length |
| @input = sort { length($a) <=> length($b) } @new_input; |
| |
| print 'Input is ' . join(', ', @input)."\n"; |
| }; |
| |
| |
| # Process a single file |
| unless ($cmd) { |
| my $input = $input[0]; |
| |
| BEGIN { |
| $main::TIME = Benchmark->new; |
| $main::LAST_STOP = Benchmark->new; |
| }; |
| |
| sub stop_time { |
| my $new = Benchmark->new; |
| $log->info( |
| 'The code took: '. |
| timestr(timediff($new, $main::LAST_STOP)) . |
| ' (overall: ' . timestr(timediff($new, $main::TIME)) . ')' |
| ); |
| $main::LAST_STOP = $new; |
| }; |
| |
| # Create and parse new document |
| $input =~ s{([^/])$}{$1/}o; |
| |
| # Process file |
| $batch_file->process($input, $output); |
| |
| # Delete cache file |
| unlink($cache_file) if $cache_delete; |
| |
| stop_time; |
| exit; |
| }; |
| |
| |
| # Extract XML files |
| if ($cmd eq 'extract') { |
| |
| # Output is required |
| pod2usage(%ERROR_HASH) unless $output; |
| |
| # Create new archive object |
| if (-f($input[0]) && (my $archive = KorAP::XML::Archive->new($input[0]))) { |
| |
| # Check zip capabilities |
| unless ($archive->test_unzip) { |
| $log->error("Unzip is not installed or incompatible."); |
| exit 1; |
| }; |
| |
| # Add further annotation archived |
| $archive->attach($_) foreach @input[1..$#input]; |
| |
| # Will set @sigle |
| my $prefix = set_sigle($archive); |
| |
| # Iterate over all given sigles and extract |
| foreach (@sigle) { |
| |
| print "$_ ...\n"; |
| |
| # TODO: Make this OS independent |
| print '... ' . ( |
| |
| # TODO: |
| # - prefix??? |
| $archive->extract_sigle([$_], $output, $jobs) |
| ? '' : 'not ' |
| ); |
| print "extracted.\n"; |
| }; |
| } |
| |
| # Can't create archive object |
| else { |
| $log->error('Unable to extract from primary archive ' . $input[0]); |
| exit 1; |
| }; |
| } |
| |
| |
| # Process an archive |
| elsif ($cmd eq 'archive') { |
| |
| my $archive_output; |
| |
| # First extract, then archive |
| if (defined $extract_dir && !-d $input[0]) { |
| |
| # Create new archive object |
| if (-f($input[0]) && (my $archive = KorAP::XML::Archive->new($input[0]))) { |
| |
| # Check zip capabilities |
| unless ($archive->test_unzip) { |
| $log->error("Unzip is not installed or incompatible."); |
| exit 1; |
| }; |
| |
| # Add further annotation archived |
| $archive->attach($_) foreach @input[1..$#input]; |
| |
| # Create a temporary directory |
| if ($extract_dir eq ':temp:') { |
| $extract_dir = tempdir(CLEANUP => 0); |
| print "Temporarily extract to $extract_dir\n"; |
| }; |
| |
| # Add some random extra to avoid clashes with multiple archives |
| $extract_dir = catdir($extract_dir, random_string('cccccc')); |
| |
| # Extract to temporary directory |
| if ($archive->extract_all($extract_dir, $sequential_extraction ? 1: $jobs)) { |
| @input = ($extract_dir); |
| } |
| else { |
| $log->error('Unable to extract from primary archive ' . $input[0] . |
| ' to ' . $extract_dir); |
| exit 1; |
| }; |
| } |
| |
| # Can't create archive object |
| else { |
| $log->error('Unable to extract from primary archive ' . $input[0]); |
| exit 1; |
| }; |
| }; |
| |
| # Zero means: everything runs in the parent process |
| my $pool = Parallel::ForkManager->new($jobs); |
| |
| my $count = 0; # Texts to process |
| my $iter = 1; # Current text in process |
| |
| my $tar_archive; |
| my $output_dir = $output; |
| my $tar_fh; |
| |
| # Initialize tar archive |
| if ($to_tar) { |
| $tar_archive = Archive::Tar::Builder->new( |
| ignore_errors => 1 |
| ); |
| |
| # Set output name |
| my $tar_file = $output; |
| unless ($tar_file =~ /\.tar$/) { |
| $tar_file .= '.tar'; |
| }; |
| |
| # Initiate the tar file |
| print "Writing to file $tar_file\n"; |
| $tar_fh = IO::File->new($tar_file, 'w'); |
| $tar_fh->binmode(1); |
| |
| # Set handle |
| $tar_archive->set_handle($tar_fh); |
| |
| # Output to temporary directory |
| $output_dir = File::Temp->newdir; |
| }; |
| |
| # Report on fork message |
| $pool->run_on_finish ( |
| sub { |
| my ($pid, $code) = @_; |
| my $data = pop; |
| |
| print 'Convert ['. ($jobs > 0 ? "\$$pid:" : '') . |
| ($iter++) . "/$count]" . |
| ($code ? " $code" : '') . |
| ' ' . $data->[0] . "\n"; |
| |
| if (!$code && $to_tar && $data->[2]) { |
| my $filename = $data->[2]; |
| |
| # Lock filehandle |
| if (flock($tar_fh, LOCK_EX)) { |
| |
| my $clean_file = fileparse($filename); |
| |
| # Archive and remove file |
| $tar_archive->archive_as($filename => $clean_file); |
| unlink $filename; |
| |
| # Unlock filehandle |
| flock($tar_fh, LOCK_UN); |
| } |
| else { |
| $log->warn("Unable to add $filename to archive"); |
| }; |
| }; |
| |
| $data->[1] = undef if $data->[1]; |
| } |
| ); |
| |
| my $t; |
| my $temp; |
| print "Reading data ...\n"; |
| |
| # unless (Cache::FastMmap->new( |
| # share_file => $cache_file, |
| # cache_size => $cache_size, |
| # init_file => $cache_init |
| # )) { |
| # print "Unable to intialize cache '$cache_file'\n\n"; |
| # exit(1); |
| # }; |
| |
| |
| # Input is a directory |
| if (-d $input[0]) { |
| # TODO: |
| # Replace with Mojo::File |
| my $it = Directory::Iterator->new($input[0]); |
| my @dirs; |
| my $dir; |
| |
| # Todo: Make a DO WHILE |
| while (1) { |
| if (!$it->is_directory && ($dir = $it->get) && $dir =~ s{/data\.xml$}{}) { |
| push @dirs, $dir; |
| $it->prune; |
| }; |
| last unless $it->next; |
| }; |
| |
| print "Start processing ...\n"; |
| $t = Benchmark->new; |
| $count = scalar @dirs; |
| |
| DIRECTORY_LOOP: |
| for (my $i = 0; $i < $count; $i++) { |
| |
| my $filename = catfile( |
| $output_dir, |
| get_file_name($input[0], $dirs[$i]) . '.json' . ($gzip ? '.gz' : '') |
| ); |
| |
| # Get the next fork |
| $pool->start and next DIRECTORY_LOOP; |
| |
| if (my $return = $batch_file->process($dirs[$i] => $filename)) { |
| $pool->finish( |
| 0, |
| [ |
| "Processed " . $filename . ($return == -1 ? " - already existing" : ''), |
| undef, |
| $filename |
| ] |
| ); |
| } |
| else { |
| $pool->finish(1, ["Unable to process " . $dirs[$i]]); |
| }; |
| }; |
| } |
| |
| # Input is a file |
| elsif (-f($input[0]) && (my $archive = KorAP::XML::Archive->new($input[0]))) { |
| |
| unless ($archive->test_unzip) { |
| $log->error("Unzip is not installed or incompatible."); |
| exit 1; |
| }; |
| |
| # Add further annotation archived |
| $archive->attach($_) foreach @input[1..$#input]; |
| |
| # Get sigles to extract |
| my $prefix = set_sigle($archive); |
| |
| print "Start processing ...\n"; |
| $t = Benchmark->new; |
| my @dirs = $archive->list_texts; |
| $count = scalar @dirs; |
| |
| ARCHIVE_LOOP: |
| for (my $i = 0; $i < $count; $i++) { |
| |
| # Split path information |
| my ($prefix, $corpus, $doc, $text) = $archive->split_path($dirs[$i]); |
| |
| my $filename = catfile( |
| $output_dir, |
| get_file_name( |
| $input[0], |
| catfile($corpus, $doc, $text) |
| . '.json' . ($gzip ? '.gz' : '') |
| ) |
| ); |
| |
| # Get the next fork |
| $pool->start and next ARCHIVE_LOOP; |
| |
| # Create temporary file |
| $temp = File::Temp->newdir; |
| |
| # TODO: Check if $filename exist at the beginning, |
| # because extraction can be horrible slow! |
| |
| # Extract from archive |
| if ($archive->extract_sigle([join('/', $corpus, $doc, $text)], $temp, $sequential_extraction ? 1 : $jobs)) { |
| |
| # Create corpus directory |
| my $input = catdir("$temp", $corpus); |
| |
| # Temporary directory |
| my $dir = catdir($input, $doc, $text); |
| |
| # Write file |
| if (my $return = $batch_file->process($dir => $filename)) { |
| |
| # Delete temporary file |
| $pool->finish( |
| 0, |
| [ |
| "Processed " . $filename . ($return == -1 ? " - already existing" : ''), |
| $temp, |
| $filename |
| ] |
| ); |
| } |
| else { |
| # Delete temporary file |
| $pool->finish(1, ["Unable to process " . $dir, $temp]); |
| }; |
| } |
| |
| # Unable to extract |
| else { |
| $pool->finish(1, ["Unable to extract " . $dirs[$i], $temp]); |
| }; |
| }; |
| } |
| |
| else { |
| print "Input is neither a directory nor an archive.\n\n"; |
| }; |
| |
| $pool->wait_all_children; |
| |
| # Delete cache file |
| unlink($cache_file) if $cache_delete; |
| |
| # Close tar filehandle |
| if ($to_tar && $tar_fh) { |
| $tar_archive->finish; |
| $tar_fh->close; |
| print "Wrote to tar archive.\n"; |
| }; |
| |
| print timestr(timediff(Benchmark->new, $t))."\n"; |
| print "Done.\n"; |
| }; |
| |
| |
| # For an archive, this will create the list |
| # of all sigles to process |
| sub set_sigle { |
| my $archive = shift; |
| |
| my $prefix = 1; |
| my @dirs = (); |
| |
| # No sigles given |
| unless (@sigle) { |
| |
| # Get files |
| foreach ($archive->list_texts) { |
| |
| push @dirs, $_; |
| |
| # Split path information |
| ($prefix, my ($corpus, $doc, $text)) = $archive->split_path($_); |
| |
| # TODO: Make this OS independent |
| push @sigle, join '/', $corpus, $doc, $text; |
| }; |
| } |
| |
| # Check sigle for doc sigles |
| else { |
| my @new_sigle; |
| |
| my $prefix_check = 0; |
| |
| # Iterate over all sigle |
| foreach (@sigle) { |
| |
| # Sigle is a doc sigle |
| if ($_ =~ m!^(?:\.[/\\])?[^/\\]+?[/\\][^/\\]+?$!) { |
| |
| print "$_ ..."; |
| # Check if a prefix is needed |
| unless ($prefix_check) { |
| |
| if ($prefix = $archive->check_prefix) { |
| print " with prefix ..."; |
| }; |
| $prefix_check = 1; |
| }; |
| |
| print "\n"; |
| |
| print '... ' . ( |
| $archive->extract_sigle([$_], $output, $sequential_extraction ? 1 : $jobs) |
| ? '' : 'not ' |
| ); |
| print "extracted.\n"; |
| } |
| |
| # Sigle is a text sigle |
| else { |
| push @new_sigle, $_; |
| |
| unless ($prefix_check) { |
| |
| if ($prefix = $archive->check_prefix) { |
| print " with prefix ..."; |
| }; |
| $prefix_check = 1; |
| }; |
| }; |
| }; |
| @sigle = @new_sigle; |
| }; |
| |
| return $prefix; |
| }; |
| |
| |
| # Cleanup temporary extraction directory |
| if ($extract_dir) { |
| my $objects = remove_tree($extract_dir, { safe => 1 }); |
| $log->info("Removed directory $extract_dir with $objects objects"); |
| }; |
| |
| |
| print "\n"; |
| |
| __END__ |
| |
| =pod |
| |
| =encoding utf8 |
| |
| =head1 NAME |
| |
| korapxml2krill - Merge KorAP-XML data and create Krill documents |
| |
| |
| =head1 SYNOPSIS |
| |
| korapxml2krill [archive|extract] --input <directory|archive> [options] |
| |
| |
| =head1 DESCRIPTION |
| |
| L<KorAP::XML::Krill> is a library to convert KorAP-XML documents to files |
| compatible with the L<Krill|https://github.com/KorAP/Krill> indexer. |
| The C<korapxml2krill> command line tool is a simple wrapper of this library. |
| |
| |
| =head1 INSTALLATION |
| |
| The preferred way to install L<KorAP::XML::Krill> is to use L<cpanm|App::cpanminus>. |
| |
| $ cpanm https://github.com/KorAP/KorAP-XML-Krill.git |
| |
| In case everything went well, the C<korapxml2krill> tool will |
| be available on your command line immediately. |
| Minimum requirement for L<KorAP::XML::Krill> is Perl 5.16. |
| Optional support for L<Sys::Info> to calculate available cores. |
| In addition to work with zip archives, the C<unzip> tool needs to be present. |
| |
| =head1 ARGUMENTS |
| |
| $ korapxml2krill -z --input <directory> --output <filename> |
| |
| Without arguments, C<korapxml2krill> converts a directory of a single KorAP-XML document. |
| It expects the input to point to the text level folder. |
| |
| =over 2 |
| |
| =item B<archive> |
| |
| $ korapxml2krill archive -z --input <directory|archive> --output <directory|tar> |
| |
| Converts an archive of KorAP-XML documents. It expects a directory |
| (pointing to the corpus level folder) or one or more zip files as input. |
| |
| =item B<extract> |
| |
| $ korapxml2krill extract --input <archive> --output <directory> --sigle <SIGLE> |
| |
| Extracts KorAP-XML documents from a zip file. |
| |
| =item B<serial> |
| |
| $ korapxml2krill serial -i <archive1> -i <archive2> -o <directory> -cfg <config-file> |
| |
| Convert archives sequentially. The inputs are not merged but treated |
| as they are (so they may be premerged or globs). |
| the C<--out> directory is treated as the base directory where subdirectories |
| are created based on the archive name. In case the C<--to-tar> flag is given, |
| the output will be a tar file. |
| |
| |
| =back |
| |
| |
| =head1 OPTIONS |
| |
| =over 2 |
| |
| =item B<--input|-i> <directory|zip file> |
| |
| Directory or zip file(s) of documents to convert. |
| |
| Without arguments, C<korapxml2krill> expects a folder of a single KorAP-XML |
| document, while C<archive> expects a KorAP-XML corpus folder or a zip |
| file to batch process multiple files. |
| C<extract> expects zip files only. |
| |
| C<archive> supports multiple input zip files with the constraint, |
| that the first archive listed contains all primary data files |
| and all meta data files. |
| |
| -i file/news.zip -i file/news.malt.zip -i "#file/news.tt.zip" |
| |
| Input may also be defined using BSD glob wildcards. |
| |
| -i 'file/news*.zip' |
| |
| The extended input array will be sorted in length order, so the shortest |
| path needs to contain all primary data files and all meta data files. |
| |
| (The directory structure follows the base directory format, |
| that may include a C<.> root folder. |
| In this case further archives lacking a C<.> root folder |
| need to be passed with a hash sign in front of the archive's name. |
| This may require to quote the parameter.) |
| |
| To support zip files, a version of C<unzip> needs to be installed that is |
| compatible with the archive file. |
| |
| B<The root folder switch using the hash sign is experimental and |
| may vanish in future versions.> |
| |
| |
| =item B<--input-base|-ib> <directory> |
| |
| The base directory for inputs. |
| |
| |
| =item B<--output|-o> <directory|file> |
| |
| Output folder for archive processing or |
| document name for single output (optional), |
| writes to C<STDOUT> by default |
| (in case C<output> is not mandatory due to further options). |
| |
| =item B<--overwrite|-w> |
| |
| Overwrite files that already exist. |
| |
| |
| =item B<--token|-t> <foundry>#<file> |
| |
| Define the default tokenization by specifying |
| the name of the foundry and optionally the name |
| of the layer-file. Defaults to C<OpenNLP#tokens>. |
| This will directly take the file instead of running |
| the layer implementation! |
| |
| |
| =item B<--base-sentences|-bs> <foundry>#<layer> |
| |
| Define the layer for base sentences. |
| If given, this will be used instead of using C<Base#Sentences>. |
| Currently C<DeReKo#Structure> and C<DGD#Structure> are the only additional |
| layers supported. |
| |
| Defaults to unset. |
| |
| |
| =item B<--base-paragraphs|-bp> <foundry>#<layer> |
| |
| Define the layer for base paragraphs. |
| If given, this will be used instead of using C<Base#Paragraphs>. |
| Currently C<DeReKo#Structure> is the only additional layer supported. |
| |
| Defaults to unset. |
| |
| |
| =item B<--base-pagebreaks|-bpb> <foundry>#<layer> |
| |
| Define the layer for base pagebreaks. |
| Currently C<DeReKo#Structure> is the only layer supported. |
| |
| Defaults to unset. |
| |
| |
| =item B<--skip|-s> <foundry>[#<layer>] |
| |
| Skip specific annotations by specifying the foundry |
| (and optionally the layer with a C<#>-prefix), |
| e.g. C<Mate> or C<Mate#Morpho>. Alternatively you can skip C<#ALL>. |
| Can be set multiple times. |
| |
| |
| =item B<--anno|-a> <foundry>#<layer> |
| |
| Convert specific annotations by specifying the foundry |
| (and optionally the layer with a C<#>-prefix), |
| e.g. C<Mate> or C<Mate#Morpho>. |
| Can be set multiple times. |
| |
| |
| =item B<--non-word-tokens|-nwt> |
| |
| Tokenize non-word tokens like word tokens (defined as matching |
| C</[\d\w]/>). Useful to treat punctuations as tokens. |
| |
| Defaults to unset. |
| |
| |
| =item B<--non-verbal-tokens|-nvt> |
| |
| Tokenize non-verbal tokens marked as in the primary data as |
| the unicode symbol 'Black Vertical Rectangle' aka \x25ae. |
| |
| Defaults to unset. |
| |
| |
| =item B<--jobs|-j> |
| |
| Define the number of concurrent jobs in seperated forks |
| for archive processing. |
| Defaults to C<0> (everything runs in a single process). |
| |
| If C<sequential-extraction> is not set to false, this will |
| also apply to extraction. |
| |
| Pass -1, and the value will be set automatically to 5 |
| times the number of available cores, in case L<Sys::Info> |
| is available. |
| This is I<experimental>. |
| |
| |
| =item B<--koral|-k> |
| |
| Version of the output format. Supported versions are: |
| C<0> for legacy serialization, C<0.03> for serialization |
| with metadata fields as key-values on the root object, |
| C<0.4> for serialization with metadata fields as a list |
| of C<"@type":"koral:field"> objects. |
| |
| Currently defaults to C<0.03>. |
| |
| |
| =item B<--sequential-extraction|-se> |
| |
| Flag to indicate, if the C<jobs> value also applies to extraction. |
| Some systems may have problems with extracting multiple archives |
| to the same folder at the same time. |
| Can be flagged using C<--no-sequential-extraction> as well. |
| Defaults to C<false>. |
| |
| |
| =item B<--meta|-m> |
| |
| Define the metadata parser to use. Defaults to C<I5>. |
| Metadata parsers can be defined in the C<KorAP::XML::Meta> namespace. |
| This is I<experimental>. |
| |
| |
| =item B<--gzip|-z> |
| |
| Compress the output. |
| Expects a defined C<output> file in single processing. |
| |
| |
| =item B<--cache|-c> |
| |
| File to mmap a cache (using L<Cache::FastMmap>). |
| Defaults to C<korapxml2krill.cache> in the calling directory. |
| |
| |
| =item B<--cache-size|-cs> |
| |
| Size of the cache. Defaults to C<50m>. |
| |
| |
| =item B<--cache-init|-ci> |
| |
| Initialize cache file. |
| Can be flagged using C<--no-cache-init> as well. |
| Defaults to C<true>. |
| |
| |
| =item B<--cache-delete|-cd> |
| |
| Delete cache file after processing. |
| Can be flagged using C<--no-cache-delete> as well. |
| Defaults to C<true>. |
| |
| |
| =item B<--config|-cfg> |
| |
| Configure the parameters of your call in a file |
| of key-value pairs with whitespace separator |
| |
| overwrite 1 |
| token DeReKo#Structure |
| ... |
| |
| Supported parameters are: |
| C<overwrite>, C<gzip>, C<jobs>, C<input-base>, |
| C<token>, C<log>, C<cache>, C<cache-size>, C<cache-delete>, C<meta>, |
| C<output>, C<koral>, |
| C<tempary-extract>, C<sequential-extraction>, |
| C<base-sentences>, C<base-paragraphs>, |
| C<base-pagebreaks>, |
| C<skip> (semicolon separated), C<sigle> |
| (semicolon separated), C<anno> (semicolon separated). |
| |
| Configuration parameters will always be overwritten by |
| passed parameters. |
| |
| |
| =item B<--temporary-extract|-te> |
| |
| Only valid for the C<archive> command. |
| |
| This will first extract all files into a |
| directory and then will archive. |
| If the directory is given as C<:temp:>, |
| a temporary directory is used. |
| This is especially useful to avoid |
| massive unzipping and potential |
| network latency. |
| |
| |
| =item B<--to-tar> |
| |
| Only valid for the C<archive> command. |
| |
| Writes the output into a tar archive. |
| |
| |
| =item B<--sigle|-sg> |
| |
| Extract the given texts. |
| Can be set multiple times. |
| I<Currently only supported on C<extract>.> |
| Sigles have the structure C<Corpus>/C<Document>/C<Text>. |
| In case the C<Text> path is omitted, the whole document will be extracted. |
| On the document level, the postfix wildcard C<*> is supported. |
| |
| |
| =item B<--log|-l> |
| |
| The L<Log::Any> log level, defaults to C<ERROR>. |
| |
| |
| =item B<--help|-h> |
| |
| Print help information. |
| |
| |
| =item B<--version|-v> |
| |
| Print version information. |
| |
| =back |
| |
| |
| =head1 ANNOTATION SUPPORT |
| |
| L<KorAP::XML::Krill> has built-in importer for some annotation foundries and layers |
| developed in the KorAP project that are part of the KorAP preprocessing pipeline. |
| The base foundry with paragraphs, sentences, and the text element are mandatory for |
| L<Krill|https://github.com/KorAP/Krill>. |
| |
| Base |
| #Paragraphs |
| #Sentences |
| |
| Connexor |
| #Morpho |
| #Phrase |
| #Sentences |
| #Syntax |
| |
| CoreNLP |
| #Constituency |
| #Morpho |
| #NamedEntities |
| #Sentences |
| |
| CMC |
| #Morpho |
| |
| DeReKo |
| #Structure |
| |
| DGD |
| #Morpho |
| #Structure |
| |
| DRuKoLa |
| #Morpho |
| |
| Glemm |
| #Morpho |
| |
| HNC |
| #Morpho |
| |
| LWC |
| #Dependency |
| |
| Malt |
| #Dependency |
| |
| MarMoT |
| #Morpho |
| |
| Mate |
| #Dependency |
| #Morpho |
| |
| MDParser |
| #Dependency |
| |
| OpenNLP |
| #Morpho |
| #Sentences |
| |
| RWK |
| #Morpho |
| #Structure |
| |
| Sgbr |
| #Lemma |
| #Morpho |
| |
| Talismane |
| #Dependency |
| #Morpho |
| |
| TreeTagger |
| #Morpho |
| #Sentences |
| |
| XIP |
| #Constituency |
| #Morpho |
| #Sentences |
| |
| |
| More importers are in preparation. |
| New annotation importers can be defined in the C<KorAP::XML::Annotation> namespace. |
| See the built-in annotation importers as examples. |
| |
| |
| =head1 About KorAP-XML |
| |
| KorAP-XML (Bański et al. 2012) is an implementation of the KorAP |
| data model (Bański et al. 2013), where text data are stored physically |
| separated from their interpretations (i.e. annotations). |
| A text document in KorAP-XML therefore consists of several files |
| containing primary data, metadata and annotations. |
| |
| The structure of a single KorAP-XML document can be as follows: |
| |
| - data.xml |
| - header.xml |
| + base |
| - tokens.xml |
| - ... |
| + struct |
| - structure.xml |
| - ... |
| + corenlp |
| - morpho.xml |
| - constituency.xml |
| - ... |
| + tree_tagger |
| - morpho.xml |
| - ... |
| - ... |
| |
| The C<data.xml> contains the primary data, the C<header.xml> contains |
| the metadata, and the annotation layers are stored in subfolders |
| like C<base>, C<struct> or C<corenlp> |
| (so-called "foundries"; Bański et al. 2013). |
| |
| Metadata is available in the TEI-P5 variant I5 |
| (Lüngen and Sperberg-McQueen 2012). See the documentation in |
| L<KorAP::XML::Meta::I5> for translatable fields. |
| |
| Annotations correspond to a variant of the TEI-P5 feature structures |
| (TEI Consortium; Lee et al. 2004). |
| Annotation feature structures refer to character sequences of the primary text |
| inside the C<text> element of the C<data.xml>. |
| A single annotation containing the lemma of a token can have the following structure: |
| |
| <span from="0" to="3"> |
| <fs type="lex" xmlns="http://www.tei-c.org/ns/1.0"> |
| <f name="lex"> |
| <fs> |
| <f name="lemma">zum</f> |
| </fs> |
| </f> |
| </fs> |
| </span> |
| |
| The C<from> and C<to> attributes are refering to the character span |
| in the primary text. |
| Depending on the kind of annotation (e.g. token-based, span-based, relation-based), |
| the structure may vary. See L<KorAP::XML::Annotation::*> for various |
| annotation preprocessors. |
| |
| Multiple KorAP-XML documents are organized on three levels following |
| the "IDS Textmodell" (Lüngen and Sperberg-McQueen 2012): |
| corpus E<gt> document E<gt> text. On each level metadata information |
| can be stored, that C<korapxml2krill> will merge to a single metadata |
| object per text. A corpus is therefore structured as follows: |
| |
| + <corpus> |
| - header.xml |
| + <document> |
| - header.xml |
| + <text> |
| - data.xml |
| - header.xml |
| - ... |
| - ... |
| |
| A single text can be identified by the concatenation of |
| the corpus identifier, the document identifier and the text identifier. |
| This identifier is called the text sigle |
| (e.g. a text with the identifier C<18486> in the document C<060> in the |
| corpus C<WPD17> has the text sigle C<WPD17/060/18486>, see C<--sigle>). |
| |
| These corpora are often stored in zip files, with which C<korapxml2krill> |
| can deal with. Corpora may also be split in multiple zip archives |
| (e.g. one zip file per foundry), which is also supported (see C<--input>). |
| |
| Examples for KorAP-XML files are included in L<KorAP::XML::Krill> |
| in form of a test suite. |
| The resulting JSON format merges all annotation layers |
| based on a single token stream. |
| |
| =head2 References |
| |
| Piotr Bański, Cyril Belica, Helge Krause, Marc Kupietz, Carsten Schnober, Oliver Schonefeld, and Andreas Witt (2011): |
| KorAP data model: first approximation, December. |
| |
| Piotr Bański, Peter M. Fischer, Elena Frick, Erik Ketzan, Marc Kupietz, Carsten Schnober, Oliver Schonefeld and Andreas Witt (2012): |
| "The New IDS Corpus Analysis Platform: Challenges and Prospects", |
| Proceedings of the Eighth International Conference on Language Resources and Evaluation (LREC 2012). |
| L<PDF|http://www.lrec-conf.org/proceedings/lrec2012/pdf/789_Paper.pdf> |
| |
| Piotr Bański, Elena Frick, Michael Hanl, Marc Kupietz, Carsten Schnober and Andreas Witt (2013): |
| "Robust corpus architecture: a new look at virtual collections and data access", |
| Corpus Linguistics 2013. Abstract Book. Lancaster: UCREL, pp. 23-25. |
| L<PDF|https://ids-pub.bsz-bw.de/frontdoor/deliver/index/docId/4485/file/Ba%c5%84ski_Frick_Hanl_Robust_corpus_architecture_2013.pdf> |
| |
| Kiyong Lee, Lou Burnard, Laurent Romary, Eric de la Clergerie, Thierry Declerck, |
| Syd Bauman, Harry Bunt, Lionel Clément, Tomaz Erjavec, Azim Roussanaly and Claude Roux (2004): |
| "Towards an international standard on featurestructure representation", |
| Proceedings of the fourth International Conference on Language Resources and Evaluation (LREC 2004), |
| pp. 373-376. |
| L<PDF|http://www.lrec-conf.org/proceedings/lrec2004/pdf/687.pdf> |
| |
| Harald Lüngen and C. M. Sperberg-McQueen (2012): |
| "A TEI P5 Document Grammar for the IDS Text Model", |
| Journal of the Text Encoding Initiative, Issue 3 | November 2012. |
| L<PDF|https://journals.openedition.org/jtei/pdf/508> |
| |
| TEI Consortium, eds: |
| "Feature Structures", |
| Guidelines for Electronic Text Encoding and Interchange. |
| L<html|https://www.tei-c.org/release/doc/tei-p5-doc/en/html/FS.html> |
| |
| =head1 AVAILABILITY |
| |
| https://github.com/KorAP/KorAP-XML-Krill |
| |
| |
| =head1 COPYRIGHT AND LICENSE |
| |
| Copyright (C) 2015-2020, L<IDS Mannheim|https://www.ids-mannheim.de/> |
| |
| Author: L<Nils Diewald|https://nils-diewald.de/> |
| |
| Contributor: Eliza Margaretha |
| |
| L<KorAP::XML::Krill> is developed as part of the L<KorAP|http://korap.ids-mannheim.de/> |
| Corpus Analysis Platform at the |
| L<Leibniz Institute for the German Language (IDS)|http://ids-mannheim.de/>, |
| member of the |
| L<Leibniz-Gemeinschaft|http://www.leibniz-gemeinschaft.de/>. |
| |
| This program is free software published under the |
| L<BSD-2 License|https://raw.githubusercontent.com/KorAP/KorAP-XML-Krill/master/LICENSE>. |
| |
| =cut |