| #!/usr/bin/env perl | 
 | use strict; | 
 | use warnings; | 
 | use v5.10; | 
 | use FindBin; | 
 | BEGIN { unshift @INC, "$FindBin::Bin/../lib" }; | 
 | use File::Spec::Functions qw/catfile catdir/; | 
 | use Getopt::Long qw/GetOptions :config no_auto_abbrev/; | 
 | use Benchmark qw/:hireswallclock/; | 
 | use IO::Compress::Gzip qw/$GzipError/; | 
 | use POSIX qw/ceil/; | 
 | use Log::Any qw($log); | 
 | use Log::Any::Adapter; | 
 | use Pod::Usage; | 
 | use Cache::FastMmap; | 
 | use Directory::Iterator; | 
 | use KorAP::XML::Krill qw!get_file_name get_file_name_from_glob!; | 
 | use KorAP::XML::Archive; | 
 | use KorAP::XML::TarBuilder; | 
 | use KorAP::XML::Tokenizer; | 
 | use KorAP::XML::Batch::File; | 
 | use Config::Simple; | 
 | use Parallel::ForkManager; | 
 | use File::Glob ':bsd_glob'; | 
 | use File::Temp qw/tempdir/; | 
 | use File::Path qw(remove_tree make_path); | 
 | use File::Basename; | 
 | use Mojo::Collection 'c'; | 
 | use String::Random qw(random_string); | 
 | use IO::File; | 
 | use Fcntl qw(:flock SEEK_END); | 
 |  | 
 | # use KorAP::XML::ForkPool; | 
 | # TODO: use Parallel::Loops | 
 | # TODO: make output files | 
 |  | 
 | # TODO: Use KorAP::XML::ForkPool! | 
 |  | 
 | # CHANGES: | 
 | # ---------------------------------------------------------- | 
 | # 2013/11/25 | 
 | # - Initial release | 
 | # | 
 | # 2014/10/29 | 
 | # - Merges foundry data to create indexer friendly documents | 
 | # | 
 | # 2016/02/04 | 
 | # - renamed to korapxml2krill | 
 | # - added Schreibgebrauch support | 
 | # | 
 | # 2016/02/12 | 
 | # - fixed foundry skipping | 
 | # - Support overwrite in archive processing | 
 | # | 
 | # 2016/02/14 | 
 | # - Added version information | 
 | # - Added support for archive files | 
 | # | 
 | # 2016/02/15 | 
 | # - Fixed temporary directory bug | 
 | # - Improved skipping before unzipping | 
 | # - Added EXPERIMENTAL concurrency support | 
 | # | 
 | # 2016/02/23 | 
 | # - Merge korapxml2krill and korapxml2krill_dir | 
 | # | 
 | # 2016/02/27 | 
 | # - Added extract function | 
 | # | 
 | # 2016/03/17 | 
 | # - Added meta switch | 
 | # | 
 | # 2016/03/18 | 
 | # - Added meta data caching | 
 | # | 
 | # 2016/06/27 | 
 | # - Added multi archive support | 
 | # - Added prefix negation support | 
 | # - Added Malt#Dependency support | 
 | # | 
 | # 2016/07/06 | 
 | # - Added MDParser#Dependency | 
 | # | 
 | # 2016/10/15 | 
 | # - Fixed temporary path issue in script | 
 | # | 
 | # 2016/10/24 | 
 | # - Improved Windows support | 
 | # | 
 | # 2016/10/24 | 
 | # - Added support for document extraction | 
 | # | 
 | # 2016/10/27 | 
 | # - Added wildcard support for document extraction | 
 | # | 
 | # 2016/12/21 | 
 | # - added support for base-sentences and base-tokenizations | 
 | # | 
 | # 2017/01/20 | 
 | # - added support for DRuKoLa annotations | 
 | # | 
 | # 2017/02/08 | 
 | # - added support for pagebreak annotations | 
 | # | 
 | # 2017/04/06 | 
 | # - added support for wildcards in input | 
 | # | 
 | # 2017/04/07 | 
 | # - support configuration option | 
 | # - support for temporary extraction | 
 | # | 
 | # 2017/04/12 | 
 | # - support serial processing | 
 | # - support input root | 
 | # - introduced --sequential-extraction flag | 
 | # | 
 | # 2017/06/19 | 
 | # - added support for DCK | 
 | # | 
 | # 2017/06/29 | 
 | # - Fixed exit codes | 
 | # | 
 | # 2017/07/04 | 
 | # - Fixed tar building process | 
 | # | 
 | # 2018/01/16 | 
 | # - Added LWC support | 
 | # | 
 | # 2018/07/19 | 
 | # - Preliminary support for HNC. | 
 | # | 
 | # 2019/01/22 | 
 | # - Preliminary support for DGD. | 
 | # - Support for non-word tokens. | 
 | # | 
 | # 2019/02/13 | 
 | # - Support for 'koral:field' array. | 
 | # - Support for Koral versioning. | 
 | # - Ignore temporary extract parameter on | 
 | #   directory archiving. | 
 | # | 
 | # 2019/08/08 | 
 | # - Support for Talismane. | 
 | # | 
 | # 2019/12/17 | 
 | # - Added support for DGD pseudo-sentences | 
 | #   based on anchor milestones. | 
 | # - Support for non-verbal annotations. | 
 | # | 
 | # 2020/04/23 | 
 | # - Added support for Redewiedergabe-Korpus structure | 
 | #   annotations, based on sentence and paragraph milestones | 
 | # - Added support for Redewiedergabe-Korpus morphology | 
 | # | 
 | # 2021/10/11 | 
 | # - Introduced support for Gingko | 
 | # | 
 | # 2022/01/17 | 
 | # - Support for temporary extraction in config | 
 | # - Introduced support for Gingko | 
 | # | 
 | # 2022/07/21 | 
 | # - Support for NKJP | 
 | # | 
 | # 2022/07/27 | 
 | # - Support for preferred language transformation | 
 | # | 
 | # 2023/02/05 | 
 | # - Support for UD | 
 | # | 
 | # 2023/02/13 | 
 | # - Fix temporary-extract handling from configuration file. | 
 | # | 
 | # 2024/03/20 | 
 | # - Added Spacy support. | 
 | # | 
 | # 2024/03/22 | 
 | # - Improve core count logging. | 
 | # ---------------------------------------------------------- | 
 |  | 
 | our $LAST_CHANGE = '2024/06/05'; | 
 | our $LOCAL = $FindBin::Bin; | 
 | our $KORAL_VERSION = 0.03; | 
 | our $VERSION_MSG = <<"VERSION"; | 
 | Version $KorAP::XML::Krill::VERSION - diewald\@ids-mannheim.de - $LAST_CHANGE | 
 | VERSION | 
 |  | 
 | # Parse comand | 
 | my $cmd; | 
 | our @ARGV; | 
 | if ($ARGV[0] && index($ARGV[0], '-') != 0) { | 
 |   $cmd = shift @ARGV; | 
 | }; | 
 | my @keep_argv = @ARGV; | 
 |  | 
 | my (@skip, @sigle, @anno, @input); | 
 |  | 
 | # Configuration hash | 
 | my %cfg = (); | 
 |  | 
 | # Count jobs/cores if not set | 
 | sub count_jobs { | 
 |   my ($cores, $jobs); | 
 |   my $msg = 'Unable to determine number of cores - set to 1'; | 
 |   if (eval("use Sys::Info; 1;") && eval("use Sys::Info::Constants qw( :device_cpu ); 1;")) { | 
 |     $cores = Sys::Info->new->device('CPU')->count; | 
 |     if ($cores <= 0) { | 
 |       $log->error($msg); | 
 |       $cores = 1; | 
 |     } | 
 |   } | 
 |   else { | 
 |     $log->error($msg); | 
 |     $cores = 1; | 
 |   }; | 
 |  | 
 |   $jobs = ceil(5 * $cores); | 
 |   return $jobs, "Run using $jobs jobs on $cores cores"; | 
 | } | 
 |  | 
 | # Parse options from the command line | 
 | GetOptions( | 
 |   'input|i=s'   => \@input, | 
 |   'input-base|ib=s' => \($cfg{input_base}), | 
 |   'output|o=s'  => \($cfg{output}), | 
 |   'overwrite|w' => \($cfg{overwrite}), | 
 |   'meta|m=s'    => \($cfg{meta}), | 
 |   'token|t=s'   => \($cfg{token}), | 
 |   'base-sentences|bs=s'   => \($cfg{base_sentences}), | 
 |   'base-paragraphs|bp=s'  => \($cfg{base_paragraphs}), | 
 |   'base-pagebreaks|bpb=s' => \($cfg{base_pagebreaks}), | 
 |   'gzip|z'      => \($cfg{gzip}), | 
 |   'temporary-extract|te=s' => \($cfg{temporary_extract}), | 
 |   'skip|s=s'    => \@skip, | 
 |   'sigle|sg=s'  => \@sigle, | 
 |   'cache|c=s'   => \($cfg{cache}), | 
 |   'config|cfg=s' => \(my $cfg_file), | 
 |   'lang=s'        => \($cfg{lang}), | 
 |   'log|l=s'     => \($cfg{log}), | 
 |   'anno|a=s'    => \@anno, | 
 |   'primary|p!'  => sub { | 
 |     warn 'Primary flag no longer supported!'; | 
 |   }, | 
 |   'quiet'       => \($cfg{quiet}), | 
 |   'pretty|y'    => sub { | 
 |     warn 'Pretty flag no longer supported!'; | 
 |   }, | 
 |   'jobs|j=i'    => \($cfg{jobs}), | 
 |   'koral|k=f'    => \($cfg{koral}), | 
 |   'to-tar'      => \($cfg{to_tar}), | 
 |   'non-word-tokens|nwt' => \($cfg{non_word_tokens}), | 
 |   'non-verbal-tokens|nvt' => \($cfg{non_verbal_tokens}), | 
 |   'sequential-extraction|se' => \($cfg{sequential_extraction}), | 
 |   'cache-size|cs=s'  => \($cfg{cache_size}), | 
 |   'cache-delete|cd!' => \($cfg{cache_delete}), | 
 |   'cache-init|ci!'   => \($cfg{cache_init}), | 
 |   'help|h'      => sub { | 
 |     pod2usage( | 
 |       -sections => 'NAME|SYNOPSIS|ARGUMENTS|OPTIONS', | 
 |       -verbose  => 99, | 
 |       -msg      => $VERSION_MSG, | 
 |       -output   => '-' | 
 |     ); | 
 |   }, | 
 |   'version|v'   => sub { | 
 |     pod2usage( | 
 |       -verbose  => 0, | 
 |       -msg      => $VERSION_MSG, | 
 |       -output   => '-' | 
 |     ), | 
 |   }, | 
 |   'job-count|jc' => sub { | 
 |     my ($j, $msg) = count_jobs(); | 
 |     pod2usage( | 
 |       -verbose  => 0, | 
 |       -msg      => $msg, | 
 |       -output   => '-' | 
 |     ) | 
 |   } | 
 | ); | 
 |  | 
 | my %ERROR_HASH = ( | 
 |   -sections => 'NAME|SYNOPSIS|ARGUMENTS|OPTIONS', | 
 |   -verbose  => 99, | 
 |   -msg      => $VERSION_MSG, | 
 |   -output   => '-', | 
 |   -exit     => 1 | 
 | ); | 
 |  | 
 | # Load from configuration and fill non-given data | 
 | if ($cfg_file && -e $cfg_file) { | 
 |   my %config; | 
 |  | 
 |   print "Reading config from $cfg_file\n"; | 
 |  | 
 |   Config::Simple->import_from($cfg_file, \%config); | 
 |  | 
 |   foreach (qw!output cache-size input-base token overwrite | 
 |               meta base-sentences base-paragraphs base-pagebreaks | 
 |               gzip to-tar log lang cache non-word-tokens | 
 |               non-verbal-tokens sequential-extraction | 
 |               temporary-extract cache-init cache-delete | 
 |               koral extract-dir jobs quiet!) { | 
 |     my $underlined = $_ =~ tr/-/_/r; | 
 |     if (!defined($cfg{$underlined}) && defined $config{$_}) { | 
 |       $cfg{$underlined} = $config{$_}; | 
 |     }; | 
 |   }; | 
 |  | 
 |   # Skip | 
 |   if (!scalar(@skip) && defined $config{'skip'}) { | 
 |     @skip = split /\s*;\s*/, $config{'skip'} ; | 
 |   }; | 
 |  | 
 |   # Sigle | 
 |   if (!scalar(@sigle) && defined $config{'sigle'}) { | 
 |     @sigle = split /\s*;\s*/, $config{'sigle'} ; | 
 |   }; | 
 |  | 
 |   # Anno | 
 |   if (!scalar(@anno) && defined $config{'anno'}) { | 
 |     @anno = split /\s*;\s*/, $config{'anno'} ; | 
 |   }; | 
 | }; | 
 |  | 
 | # Init variables and set default values | 
 | my $output           = $cfg{output}; | 
 | my $input_base       = $cfg{input_base}; | 
 | my $gzip             = $cfg{gzip}; | 
 | my $to_tar           = $cfg{to_tar}; | 
 | my $extract_dir      = $cfg{temporary_extract}; | 
 | my $token_base       = $cfg{token}               // 'OpenNLP#tokens'; | 
 | my $cache_file       = $cfg{cache}               // 'korapxml2krill.cache'; | 
 | my $jobs             = $cfg{jobs}                // 0; | 
 | my $cache_delete     = $cfg{cache_delete}        // 1; | 
 | my $base_sentences   = lc($cfg{base_sentences}   // ''); | 
 | my $base_paragraphs  = lc($cfg{base_paragraphs}  // ''); | 
 | my $base_pagebreaks  = lc($cfg{base_pagebreaks}  // ''); | 
 | my $sequential_extraction = $cfg{sequential_extraction} // 0; | 
 | my $q                = !!($cfg{quiet}) // 0; | 
 |  | 
 | # Get tokenization basis | 
 | my ($token_base_foundry, $token_base_layer) = split(/#/, $token_base) if $token_base; | 
 |  | 
 | # Remove file extension | 
 | $token_base_layer =~ s/\.xml$//i; | 
 |  | 
 | # Convert sigle to path construct | 
 | s!^\s*([^_]+?)_([^\.]+?)\.(.+?)\s*$!$1/$2/$3! foreach @sigle; | 
 |  | 
 | my %skip; | 
 | $skip{lc($_)} = 1 foreach @skip; | 
 |  | 
 | Log::Any::Adapter->set( | 
 |   'Stderr', log_level => uc($cfg{log} // 'ERROR') | 
 | ); | 
 |  | 
 | # Start log slimming | 
 | if ($cmd && $cmd eq 'slimlog') { | 
 |   require KorAP::XML::Log::Slim; | 
 |  | 
 |   my $log_file = shift @ARGV; | 
 |  | 
 |   if (-e $log_file) { | 
 |  | 
 |     my $slimmer = KorAP::XML::Log::Slim->new($log_file); | 
 |  | 
 |     # Run log filter | 
 |     $slimmer->slim_to; | 
 |   } | 
 |  | 
 |   else { | 
 |     warn "Log file can't be found"; | 
 |     exit(1); | 
 |   }; | 
 |  | 
 |   exit; | 
 | }; | 
 |  | 
 |  | 
 | if ($cmd && $output && (!defined($to_tar)) && (!-e $output || !-d $output)) { | 
 |   $log->error("Directory '$output' does not exist."); | 
 |   exit 1; | 
 | }; | 
 |  | 
 | # Input has to be defined | 
 | pod2usage(%ERROR_HASH) unless @input; | 
 |  | 
 | # Gzip has no effect, if no output is given | 
 | pod2usage(%ERROR_HASH) if $gzip && !$output; | 
 |  | 
 | # Start serial processing | 
 | if ($cmd && $cmd eq 'serial') { | 
 |  | 
 |   # Remove all inputs | 
 |   my $remove_next = 0; | 
 |   @keep_argv = @{c(@keep_argv)->grep( | 
 |     sub { | 
 |       # Input flag | 
 |       if ($_ eq '-i' || $_ eq '--input' || $_ eq '--output' || $_ eq '-o') { | 
 |         $remove_next = 1; | 
 |         return 0; | 
 |       } | 
 |  | 
 |       # input value | 
 |       elsif ($remove_next) { | 
 |         $remove_next = 0; | 
 |         return 0; | 
 |       }; | 
 |  | 
 |       # Pass parameter | 
 |       return 1; | 
 |     } | 
 |   )->to_array}; | 
 |  | 
 |  | 
 |   # Iterate over all inputs | 
 |   foreach (@input) { | 
 |  | 
 |     # This will create a directory | 
 |     my $new_out = catdir($output, get_file_name_from_glob($_)); | 
 |  | 
 |     # Create new path, in case the output is not meant to be tarred | 
 |     unless ($to_tar) { | 
 |       if (make_path($new_out) == 0 && !-d $new_out) { | 
 |         $log->error("Can\'t create path $new_out"); | 
 |         exit 1; | 
 |       }; | 
 |     }; | 
 |  | 
 |     # Create archive command | 
 |     my @archive_cmd = ($^X, $0, 'archive', @keep_argv, '-i', $_, '-o', $new_out); | 
 |     unless ($q) { | 
 |       print "Start serial processing of $_ to $new_out\n"; | 
 |       print 'Command: ', join(' ', @archive_cmd), "\n"; | 
 |     }; | 
 |  | 
 |     # Start archiving | 
 |     system @archive_cmd; | 
 |   }; | 
 |  | 
 |   exit; | 
 | }; | 
 |  | 
 | # Define supported (and preinstalled) transformation modules | 
 | my @layers = (); | 
 | push(@layers, ['Base', 'Sentences']) unless $base_sentences; | 
 | push(@layers, ['Base', 'Paragraphs']) unless $base_paragraphs; | 
 |  | 
 | # Connexor | 
 | push(@layers, ['Connexor', 'Morpho'], | 
 |      ['Connexor', 'Syntax'], | 
 |      ['Connexor', 'Phrase'], | 
 |      ['Connexor', 'Sentences']); | 
 |  | 
 | # CoreNLP | 
 | push(@layers, | 
 |      ['CoreNLP', 'NamedEntities'], | 
 |      ['CoreNLP', 'Sentences'], | 
 |      ['CoreNLP', 'Morpho'], | 
 |      ['CoreNLP', 'Constituency']); | 
 |  | 
 | # CorpusExplorer | 
 | push(@layers, | 
 |      ['CorpusExplorer', 'Morpho']); | 
 |  | 
 | # CMC | 
 | push(@layers, ['CMC', 'Morpho']); | 
 |  | 
 | # DeReKo | 
 | my @dereko_attr = (); | 
 | if ($base_sentences eq 'dereko#structure') { | 
 |   push @dereko_attr, 'sentences'; | 
 | }; | 
 | if ($base_paragraphs eq 'dereko#structure') { | 
 |   push @dereko_attr, 'paragraphs'; | 
 | }; | 
 |  | 
 | if ($base_pagebreaks eq 'dereko#structure') { | 
 |   push @dereko_attr, 'pagebreaks'; | 
 | }; | 
 |  | 
 | if ($dereko_attr[0]) { | 
 |   push(@layers, ['DeReKo', 'Structure', 'base-' . join('-', @dereko_attr)]); | 
 | } | 
 | else { | 
 |   push(@layers, ['DeReKo', 'Structure']); | 
 | }; | 
 |  | 
 | # DGD | 
 | push(@layers, ['DGD', 'Morpho']); | 
 | if ($base_sentences eq 'dgd#structure') { | 
 |   push(@layers, ['DGD', 'Structure', 'base-sentence']); | 
 | } | 
 |  | 
 | # DRuKoLa | 
 | push(@layers, | 
 |      ['DRuKoLa', 'Morpho']); | 
 |  | 
 | # Gingko | 
 | push(@layers, | 
 |      ['Gingko', 'Morpho']); | 
 |  | 
 | # Glemm | 
 | push(@layers, | 
 |      ['Glemm', 'Morpho']); | 
 |  | 
 | # HNC | 
 | push(@layers, | 
 |      ['HNC', 'Morpho']); | 
 |  | 
 | # LWC | 
 | push(@layers, | 
 |      ['LWC', 'Dependency']); | 
 |  | 
 | # Malt | 
 | push(@layers, | 
 |      ['Malt', 'Dependency']); | 
 |  | 
 | # Marmot | 
 | push(@layers, | 
 |      ['MarMoT', 'Morpho']); | 
 |  | 
 | # Mate | 
 | push(@layers, | 
 |      ['Mate', 'Morpho'], | 
 |      ['Mate', 'Dependency']); | 
 |  | 
 | # MDParser | 
 | push(@layers, | 
 |      ['MDParser', 'Dependency']); | 
 |  | 
 | # NKJP | 
 | push(@layers, | 
 |      ['NKJP', 'Morpho'], | 
 |      ['NKJP', 'NamedEntities']); | 
 |  | 
 | # OpenNLP | 
 | push(@layers, | 
 |      ['OpenNLP', 'Morpho'], | 
 |      ['OpenNLP', 'Sentences']); | 
 |  | 
 | # Redewiedergabe | 
 | push(@layers, ['RWK', 'Morpho']); | 
 | if ($base_sentences eq 'rwk#structure') { | 
 |   push(@layers, ['RWK', 'Structure']); | 
 | }; | 
 |  | 
 | # Schreibgebrauch | 
 | push(@layers, | 
 |      ['Sgbr', 'Lemma'], | 
 |      ['Sgbr', 'Morpho']); | 
 |  | 
 | # Spacy | 
 | push(@layers, | 
 |      ['Spacy', 'Morpho']); | 
 |  | 
 | # Talismane | 
 | push(@layers, | 
 |      ['Talismane', 'Dependency'], | 
 |      ['Talismane', 'Morpho']); | 
 |  | 
 | # TreeTagger | 
 | push(@layers, | 
 |      ['TreeTagger', 'Morpho'], | 
 |      ['TreeTagger', 'Sentences']); | 
 |  | 
 | # UDPipe | 
 | push(@layers, | 
 |     ['UDPipe', 'Morpho'], | 
 |     ['UDPipe', 'Dependency']); | 
 |  | 
 | # XIP | 
 | push(@layers, | 
 |      ['XIP', 'Morpho'], | 
 |      ['XIP', 'Constituency'], | 
 |      ['XIP', 'Sentences'], | 
 |      ['XIP', 'Dependency']); | 
 |  | 
 |  | 
 | # Check filters | 
 | my @filtered_anno; | 
 | if ($skip{'#all'}) { | 
 |   foreach (@anno) { | 
 |     push @filtered_anno, [ split('#', $_) ]; | 
 |   }; | 
 | } | 
 |  | 
 | # Add all annotations that are not skipped | 
 | else { | 
 |   # Add to index file - respect skipping | 
 |   foreach my $info (@layers) { | 
 |     # Skip if Foundry or Foundry#Layer should be skipped | 
 |     unless ($skip{lc($info->[0])} || $skip{lc($info->[0]) . '#' . lc($info->[1])}) { | 
 |       push @filtered_anno, $info; | 
 |     }; | 
 |   }; | 
 | }; | 
 |  | 
 |  | 
 | # TODO: This should not be initialized for batch | 
 | my $cache = Cache::FastMmap->new( | 
 |   share_file => $cache_file, | 
 |   cache_size => ($cfg{cache_size} // '50m'), | 
 |   init_file  => ($cfg{cache_init} // 1), | 
 |   unlink_on_exit => $cache_delete | 
 | ); | 
 |  | 
 | # Create batch object | 
 | my $batch_file = KorAP::XML::Batch::File->new( | 
 |   cache     => $cache, | 
 |   meta_type => $cfg{meta}, | 
 |   overwrite => $cfg{overwrite}, | 
 |   foundry   => $token_base_foundry, | 
 |   layer     => $token_base_layer, | 
 |   gzip      => $gzip, | 
 |   log       => $log, | 
 |   koral     => ($cfg{koral} // $KORAL_VERSION), | 
 |   anno      => \@filtered_anno, | 
 |   non_word_tokens   => ($cfg{non_word_tokens}   // 0), | 
 |   non_verbal_tokens => ($cfg{non_verbal_tokens} // 0), | 
 |   lang      => $cfg{lang}, | 
 | ); | 
 |  | 
 | # Auto adjust jobs | 
 | if ($jobs eq '-1') { | 
 |   ($jobs, my $msg) = count_jobs(); | 
 |   print $msg . "\n" unless $q; | 
 | }; | 
 |  | 
 | # Glob and prefix files | 
 | if (@input > 0) { | 
 |  | 
 |   my @new_input = (); | 
 |  | 
 |   # Iterate over all inputs | 
 |   foreach my $wild_card (@input) { | 
 |  | 
 |     # Prefix with input root | 
 |     $wild_card = $input_base ? catfile($input_base, $wild_card) : $wild_card; | 
 |  | 
 |     push (@new_input, bsd_glob($wild_card)); | 
 |   }; | 
 |  | 
 |   # Sort files by length | 
 |   @input = sort { length($a) <=> length($b) } @new_input; | 
 |  | 
 |   print 'Input is ' . join(', ', @input)."\n" unless $q; | 
 | }; | 
 |  | 
 | # Process a single file | 
 | unless ($cmd) { | 
 |   my $input = $input[0]; | 
 |  | 
 |   BEGIN { | 
 |     $main::TIME = Benchmark->new; | 
 |     $main::LAST_STOP = Benchmark->new; | 
 |   }; | 
 |  | 
 |   sub stop_time { | 
 |     my $new = Benchmark->new; | 
 |     $log->info( | 
 |       'The code took: '. | 
 |         timestr(timediff($new, $main::LAST_STOP)) . | 
 |         ' (overall: ' . timestr(timediff($new, $main::TIME)) . ')' | 
 |       ); | 
 |     $main::LAST_STOP = $new; | 
 |   }; | 
 |  | 
 |   # Create and parse new document | 
 |   $input =~ s{([^/])$}{$1/}o; | 
 |  | 
 |   # Process file | 
 |   $batch_file->process($input, $output); | 
 |  | 
 |   stop_time; | 
 |  | 
 |   exit; | 
 | }; | 
 |  | 
 |  | 
 | # Extract XML files | 
 | if ($cmd eq 'extract') { | 
 |  | 
 |   # Output is required | 
 |   pod2usage(%ERROR_HASH) unless $output; | 
 |  | 
 |   # Create new archive object | 
 |   if (-f($input[0]) && (my $archive = KorAP::XML::Archive->new($input[0]))) { | 
 |  | 
 |     # Check zip capabilities | 
 |     unless ($archive->test_unzip) { | 
 |       $log->error("Unzip is not installed or incompatible."); | 
 |       exit 1; | 
 |     }; | 
 |  | 
 |     # Add further annotation archived | 
 |     $archive->attach($_) foreach @input[1..$#input]; | 
 |  | 
 |     # Will set @sigle | 
 |     my $prefix = set_sigle($archive); | 
 |  | 
 |     # Iterate over all given sigles and extract | 
 |     foreach (@sigle) { | 
 |  | 
 |       unless ($q) { | 
 |         print "$_ ...\n"; | 
 |  | 
 |         # TODO: Make this OS independent | 
 |         print '... ' . ( | 
 |  | 
 |           # TODO: | 
 |           #   - prefix??? | 
 |           $archive->extract_sigle(0, [$_], $output, $jobs) | 
 |           ? '' : 'not ' | 
 |         ); | 
 |         print "extracted.\n"; | 
 |       } else { | 
 |         $archive->extract_sigle(1, [$_], $output, $jobs); | 
 |       } | 
 |     }; | 
 |   } | 
 |  | 
 |   # Can't create archive object | 
 |   else { | 
 |     $log->error('Unable to extract from primary archive ' . $input[0]); | 
 |     exit 1; | 
 |   }; | 
 | } | 
 |  | 
 |  | 
 | # Process an archive | 
 | elsif ($cmd eq 'archive') { | 
 |  | 
 |   my $archive_output; | 
 |  | 
 |   # First extract, then archive | 
 |   if (defined $extract_dir && !-d $input[0]) { | 
 |  | 
 |     # Create new archive object | 
 |     if (-f($input[0]) && (my $archive = KorAP::XML::Archive->new($input[0]))) { | 
 |  | 
 |       # Check zip capabilities | 
 |       unless ($archive->test_unzip) { | 
 |         $log->error("Unzip is not installed or incompatible."); | 
 |         exit 1; | 
 |       }; | 
 |  | 
 |       # Add further annotation archived | 
 |       $archive->attach($_) foreach @input[1..$#input]; | 
 |  | 
 |       # Create a temporary directory | 
 |       if ($extract_dir eq ':temp:') { | 
 |         $extract_dir = tempdir(CLEANUP => 0); | 
 |         print "Temporarily extract to $extract_dir\n" unless $q; | 
 |       }; | 
 |  | 
 |       # Add some random extra to avoid clashes with multiple archives | 
 |       $extract_dir = catdir($extract_dir, random_string('cccccc')); | 
 |  | 
 |       # Extract to temporary directory | 
 |       if ($archive->extract_all($q, $extract_dir, $sequential_extraction ? 1: $jobs)) { | 
 |         print "Extract sequentially to $extract_dir\n" unless $q; | 
 |         @input = ($extract_dir); | 
 |       } | 
 |       else { | 
 |         $log->error('Unable to extract from primary archive ' . $input[0] . | 
 |                       ' to ' . $extract_dir); | 
 |         exit 1; | 
 |       }; | 
 |     } | 
 |  | 
 |     # Can't create archive object | 
 |     else { | 
 |       $log->error('Unable to extract from primary archive ' . $input[0]); | 
 |       exit 1; | 
 |     }; | 
 |   }; | 
 |  | 
 |   # Zero means: everything runs in the parent process | 
 |   my $pool = Parallel::ForkManager->new($jobs); | 
 |  | 
 |   my $count = 0;  # Texts to process | 
 |   my $iter  = 1;  # Current text in process | 
 |  | 
 |   my $tar_archive; | 
 |   my $output_dir = $output; | 
 |   my $tar_fh; | 
 |  | 
 |   # Initialize tar archive | 
 |   if ($to_tar) { | 
 |  | 
 |     # Set output name | 
 |     my $tar_file = $output; | 
 |     unless ($tar_file =~ /\.tar$/) { | 
 |       $tar_file .= '.tar'; | 
 |     }; | 
 |  | 
 |     # Initiate the tar file | 
 |     print "Writing to file $tar_file\n" unless $q; | 
 |     $tar_fh = IO::File->new($tar_file, 'w'); | 
 |     $tar_fh->binmode(1); | 
 |  | 
 |     # Use tar builder for archiving | 
 |     if (eval("use Archive::Tar::Builder; 1;")) { | 
 |       $tar_archive = Archive::Tar::Builder->new( | 
 |         ignore_errors => 1 | 
 |       ); | 
 |  | 
 |       # Set handle | 
 |       $tar_archive->set_handle($tar_fh); | 
 |     } | 
 |  | 
 |     # Fallback solution | 
 |     else { | 
 |       $tar_archive = KorAP::XML::TarBuilder->new( | 
 |         $tar_fh | 
 |       ); | 
 |     }; | 
 |  | 
 |     # Output to temporary directory | 
 |     $output_dir = File::Temp->newdir; | 
 |   }; | 
 |  | 
 |   # Report on fork message | 
 |   $pool->run_on_finish ( | 
 |     sub { | 
 |       my ($pid, $code) = @_; | 
 |       my $data = pop; | 
 |  | 
 |       unless ($q) { | 
 |         print 'Convert ['. ($jobs > 0 ? "\$$pid:" : '') . | 
 |           $iter . "/$count]" . | 
 |           ($code ? " $code" : '') . | 
 |           ' ' . $data->[0] . "\n"; | 
 |       }; | 
 |       $iter++; | 
 |  | 
 |       if (!$code && $to_tar && $data->[2]) { | 
 |         my $filename = $data->[2]; | 
 |  | 
 |         # Lock filehandle | 
 |         if (flock($tar_fh, LOCK_EX)) { | 
 |  | 
 |           my $clean_file = fileparse($filename); | 
 |  | 
 |           # Archive and remove file | 
 |           $tar_archive->archive_as($filename => $clean_file); | 
 |           unlink $filename; | 
 |  | 
 |           # Unlock filehandle | 
 |           flock($tar_fh, LOCK_UN); | 
 |         } | 
 |         else { | 
 |           $log->warn("Unable to add $filename to archive"); | 
 |         }; | 
 |       }; | 
 |  | 
 |       $data->[1] = undef if $data->[1]; | 
 |     } | 
 |   ); | 
 |  | 
 |   my $t; | 
 |   my $temp; | 
 |   print "Reading data ...\n" unless $q; | 
 |  | 
 |   #  unless (Cache::FastMmap->new( | 
 |   #    share_file => $cache_file, | 
 |   #    cache_size => $cache_size, | 
 |   #    init_file => $cache_init | 
 |   #  )) { | 
 |   #    print "Unable to intialize cache '$cache_file'\n\n"; | 
 |   #    exit(1); | 
 |   #  }; | 
 |  | 
 |  | 
 |   # Input is a directory | 
 |   if (-d $input[0]) { | 
 |     # TODO: | 
 |     #   Replace with Mojo::File | 
 |     my $it = Directory::Iterator->new($input[0]); | 
 |     my @dirs; | 
 |     my $dir; | 
 |  | 
 |     # Todo: Make a DO WHILE | 
 |     while (1) { | 
 |       if (!$it->is_directory && ($dir = $it->get) && $dir =~ s{/data\.xml$}{}) { | 
 |         push @dirs, $dir; | 
 |         $it->prune; | 
 |       }; | 
 |       last unless $it->next; | 
 |     }; | 
 |  | 
 |     print "Start processing ...\n" unless $q; | 
 |     $t = Benchmark->new; | 
 |     $count = scalar @dirs; | 
 |  | 
 |   DIRECTORY_LOOP: | 
 |     for (my $i = 0; $i < $count; $i++) { | 
 |  | 
 |       my $filename = catfile( | 
 |         $output_dir, | 
 |         get_file_name($input[0], $dirs[$i]) . '.json' . ($gzip ? '.gz' : '') | 
 |       ); | 
 |  | 
 |       # Get the next fork | 
 |       $pool->start and next DIRECTORY_LOOP; | 
 |  | 
 |       if (my $return = $batch_file->process($dirs[$i] => $filename)) { | 
 |         $pool->finish( | 
 |           0, | 
 |           [ | 
 |             "Processed " . $filename . ($return == -1 ? " - already existing" : ''), | 
 |             undef, | 
 |             $filename | 
 |           ] | 
 |         ); | 
 |       } | 
 |       else { | 
 |         $pool->finish(1, ["Unable to process " . $dirs[$i]]); | 
 |       }; | 
 |     }; | 
 |   } | 
 |  | 
 |   # Input is a file | 
 |   elsif (-f($input[0]) && (my $archive = KorAP::XML::Archive->new($input[0]))) { | 
 |  | 
 |     unless ($archive->test_unzip) { | 
 |       $log->error("Unzip is not installed or incompatible."); | 
 |       exit 1; | 
 |     }; | 
 |  | 
 |     # Add further annotation archived | 
 |     $archive->attach($_) foreach @input[1..$#input]; | 
 |  | 
 |     # Get sigles to extract | 
 |     my $prefix = set_sigle($archive); | 
 |  | 
 |     print "Start processing ...\n" unless $q; | 
 |     $t = Benchmark->new; | 
 |     my @dirs = $archive->list_texts; | 
 |     $count = scalar @dirs; | 
 |  | 
 |   ARCHIVE_LOOP: | 
 |     for (my $i = 0; $i < $count; $i++) { | 
 |  | 
 |       # Split path information | 
 |       my ($prefix, $corpus, $doc, $text) = $archive->split_path($dirs[$i]); | 
 |  | 
 |       my $filename = catfile( | 
 |         $output_dir, | 
 |         get_file_name( | 
 |           $input[0], | 
 |           catfile($corpus, $doc, $text) | 
 |             . '.json' . ($gzip ? '.gz' : '') | 
 |           ) | 
 |       ); | 
 |  | 
 |       # Get the next fork | 
 |       $pool->start and next ARCHIVE_LOOP; | 
 |  | 
 |       # Create temporary file | 
 |       $temp = File::Temp->newdir; | 
 |  | 
 |       # TODO: Check if $filename exist at the beginning, | 
 |       # because extraction can be horrible slow! | 
 |  | 
 |       # Extract from archive | 
 |       if ($archive->extract_sigle($q, [join('/', $corpus, $doc, $text)], $temp, $sequential_extraction ? 1 : $jobs)) { | 
 |  | 
 |         # Create corpus directory | 
 |         my $input = catdir("$temp", $corpus); | 
 |  | 
 |         # Temporary directory | 
 |         my $dir = catdir($input, $doc, $text); | 
 |  | 
 |         # Write file | 
 |         if (my $return = $batch_file->process($dir => $filename)) { | 
 |  | 
 |           # Delete temporary file | 
 |           $pool->finish( | 
 |             0, | 
 |             [ | 
 |               "Processed " . $filename . ($return == -1 ? " - already existing" : ''), | 
 |               $temp, | 
 |               $filename | 
 |             ] | 
 |           ); | 
 |         } | 
 |         else { | 
 |           # Delete temporary file | 
 |           $pool->finish(1, ["Unable to process " . $dir, $temp]); | 
 |         }; | 
 |       } | 
 |  | 
 |       # Unable to extract | 
 |       else { | 
 |         $pool->finish(1, ["Unable to extract " . $dirs[$i], $temp]); | 
 |       }; | 
 |     }; | 
 |   } | 
 |  | 
 |   else { | 
 |     print "Input is neither a directory nor an archive.\n\n" unless $q; | 
 |   }; | 
 |  | 
 |   $pool->wait_all_children; | 
 |  | 
 |   # Close tar filehandle | 
 |   if ($to_tar && $tar_fh) { | 
 |     $tar_archive->finish; | 
 |     $tar_fh->close; | 
 |     print "Wrote to tar archive.\n" unless $q; | 
 |   }; | 
 |   unless ($q) { | 
 |     print timestr(timediff(Benchmark->new, $t))."\n"; | 
 |     print "Done.\n"; | 
 |   }; | 
 | }; | 
 |  | 
 |  | 
 | # For an archive, this will create the list | 
 | # of all sigles to process | 
 | sub set_sigle { | 
 |   my $archive = shift; | 
 |  | 
 |   my $prefix = 1; | 
 |   my @dirs = (); | 
 |  | 
 |   # No sigles given | 
 |   unless (@sigle) { | 
 |  | 
 |     # Get files | 
 |     foreach ($archive->list_texts) { | 
 |  | 
 |       push @dirs, $_; | 
 |  | 
 |       # Split path information | 
 |       ($prefix, my ($corpus, $doc, $text)) = $archive->split_path($_); | 
 |  | 
 |       # TODO: Make this OS independent | 
 |       push @sigle, join '/', $corpus, $doc, $text; | 
 |     }; | 
 |   } | 
 |  | 
 |   # Check sigle for doc sigles | 
 |   else { | 
 |     my @new_sigle; | 
 |  | 
 |     my $prefix_check = 0; | 
 |  | 
 |     # Iterate over all sigle | 
 |     foreach (@sigle) { | 
 |  | 
 |       # Sigle is a doc sigle | 
 |       if ($_ =~ m!^(?:\.[/\\])?[^/\\]+?[/\\][^/\\]+?$!) { | 
 |  | 
 |         print "$_ ..." unless $q; | 
 |         # Check if a prefix is needed | 
 |         unless ($prefix_check) { | 
 |  | 
 |           if ($prefix = $archive->check_prefix && !$q) { | 
 |             print " with prefix ..."; | 
 |           }; | 
 |           $prefix_check = 1; | 
 |         }; | 
 |  | 
 |         unless ($q) { | 
 |           print "\n"; | 
 |  | 
 |           print '... ' . ( | 
 |             $archive->extract_sigle($q, [$_], $output, $sequential_extraction ? 1 : $jobs) | 
 |             ? '' : 'not ' | 
 |           ); | 
 |           print "extracted.\n"; | 
 |         } | 
 |         else { | 
 |           $archive->extract_sigle($q, [$_], $output, $sequential_extraction ? 1 : $jobs) | 
 |         }; | 
 |       } | 
 |  | 
 |       # Sigle is a text sigle | 
 |       else { | 
 |         push @new_sigle, $_; | 
 |  | 
 |         unless ($prefix_check) { | 
 |  | 
 |           if ($prefix = $archive->check_prefix && !$q) { | 
 |             print " with prefix ..."; | 
 |           }; | 
 |           $prefix_check = 1; | 
 |         }; | 
 |       }; | 
 |     }; | 
 |     @sigle = @new_sigle; | 
 |   }; | 
 |  | 
 |   return $prefix; | 
 | }; | 
 |  | 
 |  | 
 | # Cleanup temporary extraction directory | 
 | if ($extract_dir) { | 
 |   my $objects = remove_tree($extract_dir, { safe => 1 }); | 
 |   $log->info("Removed directory $extract_dir with $objects objects"); | 
 | }; | 
 |  | 
 |  | 
 | print "\n"; | 
 |  | 
 | __END__ | 
 |  | 
 | =pod | 
 |  | 
 | =encoding utf8 | 
 |  | 
 | =head1 NAME | 
 |  | 
 | korapxml2krill - Merge KorAP-XML data and create Krill documents | 
 |  | 
 |  | 
 | =head1 SYNOPSIS | 
 |  | 
 |   $ korapxml2krill [archive|extract] --input <directory|archive> [options] | 
 |  | 
 |  | 
 | =head1 DESCRIPTION | 
 |  | 
 | L<KorAP::XML::Krill> is a library to convert KorAP-XML documents to files | 
 | compatible with the L<Krill|https://github.com/KorAP/Krill> indexer. | 
 | The C<korapxml2krill> command line tool is a simple wrapper of this library. | 
 |  | 
 |  | 
 | =head1 INSTALLATION | 
 |  | 
 | The preferred way to install L<KorAP::XML::Krill> is to use L<cpanm|App::cpanminus>. | 
 |  | 
 |   $ cpanm https://github.com/KorAP/KorAP-XML-Krill.git | 
 |  | 
 | In case everything went well, the C<korapxml2krill> tool will | 
 | be available on your command line immediately. | 
 | Minimum requirement for L<KorAP::XML::Krill> is Perl 5.32. | 
 | Optionally installing L<Archive::Tar::Builder> speeds up archive building. | 
 | Optional support for L<Sys::Info> to calculate available cores is available. | 
 | In addition to work with zip archives, the C<unzip> tool needs to be present. | 
 |  | 
 | =head1 ARGUMENTS | 
 |  | 
 |   $ korapxml2krill -z --input <directory> --output <filename> | 
 |  | 
 | Without arguments, C<korapxml2krill> converts a directory of a single KorAP-XML document. | 
 | It expects the input to point to the text level folder. | 
 |  | 
 | =over 2 | 
 |  | 
 | =item B<archive> | 
 |  | 
 |   $ korapxml2krill archive -z --input <directory|archive> --output <directory|tar> | 
 |  | 
 | Converts an archive of KorAP-XML documents. It expects a directory | 
 | (pointing to the corpus level folder) or one or more zip files as input. | 
 |  | 
 | =item B<extract> | 
 |  | 
 |   $ korapxml2krill extract --input <archive> --output <directory> --sigle <SIGLE> | 
 |  | 
 | Extracts KorAP-XML documents from a zip file. | 
 |  | 
 | =item B<serial> | 
 |  | 
 |   $ korapxml2krill serial -i <archive1> -i <archive2> -o <directory> -cfg <config-file> | 
 |  | 
 | Convert archives sequentially. The inputs are not merged but treated | 
 | as they are (so they may be premerged or globs). | 
 | the C<--out> directory is treated as the base directory where subdirectories | 
 | are created based on the archive name. In case the C<--to-tar> flag is given, | 
 | the output will be a tar file. | 
 |  | 
 |  | 
 | =item B<slimlog> | 
 |  | 
 |   $ korapxml2krill slimlog <logfile> > <logfile-slim> | 
 |  | 
 | Filters out all useless aka succesfull information from logs, to simplify | 
 | log checks. Expects no further options. | 
 |  | 
 |  | 
 | =back | 
 |  | 
 |  | 
 | =head1 OPTIONS | 
 |  | 
 | =over 2 | 
 |  | 
 | =item B<--input|-i> <directory|zip file> | 
 |  | 
 | Directory or zip file(s) of documents to convert. | 
 |  | 
 | Without arguments, C<korapxml2krill> expects a folder of a single KorAP-XML | 
 | document, while C<archive> expects a KorAP-XML corpus folder or a zip | 
 | file to batch process multiple files. | 
 | C<extract> expects zip files only. | 
 |  | 
 | C<archive> supports multiple input zip files with the constraint, | 
 | that the first archive listed contains all primary data files | 
 | and all meta data files. | 
 |  | 
 |   -i file/news.zip -i file/news.malt.zip -i "#file/news.tt.zip" | 
 |  | 
 | Input may also be defined using BSD glob wildcards. | 
 |  | 
 |   -i 'file/news*.zip' | 
 |  | 
 | The extended input array will be sorted in length order, so the shortest | 
 | path needs to contain all primary data files and all meta data files. | 
 |  | 
 | (The directory structure follows the base directory format, | 
 | that may include a C<.> root folder. | 
 | In this case further archives lacking a C<.> root folder | 
 | need to be passed with a hash sign in front of the archive's name. | 
 | This may require to quote the parameter.) | 
 |  | 
 | To support zip files, a version of C<unzip> needs to be installed that is | 
 | compatible with the archive file. | 
 |  | 
 | B<The root folder switch using the hash sign is experimental and | 
 | may vanish in future versions.> | 
 |  | 
 |  | 
 | =item B<--input-base|-ib> <directory> | 
 |  | 
 | The base directory for inputs. | 
 |  | 
 |  | 
 | =item B<--output|-o> <directory|file> | 
 |  | 
 | Output folder for archive processing or | 
 | document name for single output (optional), | 
 | writes to C<STDOUT> by default | 
 | (in case C<output> is not mandatory due to further options). | 
 |  | 
 | =item B<--overwrite|-w> | 
 |  | 
 | Overwrite files that already exist. | 
 |  | 
 |  | 
 | =item B<--token|-t> <foundry>#<file> | 
 |  | 
 | Define the default tokenization by specifying | 
 | the name of the foundry and optionally the name | 
 | of the layer-file. Defaults to C<OpenNLP#tokens>. | 
 | This will directly take the file instead of running | 
 | the layer implementation! | 
 |  | 
 |  | 
 | =item B<--base-sentences|-bs> <foundry>#<layer> | 
 |  | 
 | Define the layer for base sentences. | 
 | If given, this will be used instead of using C<Base#Sentences>. | 
 | Currently C<DeReKo#Structure> and C<DGD#Structure> are the only additional | 
 | layers supported. | 
 |  | 
 |  Defaults to unset. | 
 |  | 
 |  | 
 | =item B<--base-paragraphs|-bp> <foundry>#<layer> | 
 |  | 
 | Define the layer for base paragraphs. | 
 | If given, this will be used instead of using C<Base#Paragraphs>. | 
 | Currently C<DeReKo#Structure> and C<DGD#Structure> are the only additional | 
 | layer supported. | 
 |  | 
 |  Defaults to unset. | 
 |  | 
 |  | 
 | =item B<--base-pagebreaks|-bpb> <foundry>#<layer> | 
 |  | 
 | Define the layer for base pagebreaks. | 
 | Currently C<DeReKo#Structure> is the only layer supported. | 
 |  | 
 |  Defaults to unset. | 
 |  | 
 |  | 
 | =item B<--skip|-s> <foundry>[#<layer>] | 
 |  | 
 | Skip specific annotations by specifying the foundry | 
 | (and optionally the layer with a C<#>-prefix), | 
 | e.g. C<Mate> or C<Mate#Morpho>. Alternatively you can skip C<#ALL>. | 
 | Can be set multiple times. | 
 |  | 
 |  | 
 | =item B<--anno|-a> <foundry>#<layer> | 
 |  | 
 | Convert specific annotations by specifying the foundry | 
 | (and optionally the layer with a C<#>-prefix), | 
 | e.g. C<Mate> or C<Mate#Morpho>. | 
 | Can be set multiple times. | 
 |  | 
 |  | 
 | =item B<--non-word-tokens|-nwt> | 
 |  | 
 | Tokenize non-word tokens like word tokens (defined as matching | 
 | C</[\d\w]/>). Useful to treat punctuations as tokens. | 
 |  | 
 |  Defaults to unset. | 
 |  | 
 |  | 
 | =item B<--non-verbal-tokens|-nvt> | 
 |  | 
 | Tokenize non-verbal tokens marked as in the primary data as  | 
 | the unicode symbol 'Black Vertical Rectangle' aka \x25ae. | 
 |  | 
 |  Defaults to unset. | 
 |  | 
 |  | 
 | =item B<--jobs|-j> | 
 |  | 
 | Define the number of spawned forks for concurrent jobs | 
 | of archive processing. | 
 | Defaults to C<0> (everything runs in a single process). | 
 |  | 
 | If C<sequential-extraction> is not set to true, this will | 
 | also apply to extraction. | 
 |  | 
 | Pass C<-1>, and the value will be set automatically to 5 | 
 | times the number of available cores, in case L<Sys::Info> | 
 | is available and can read CPU count (see C<--job-count>). | 
 | Be aware, that the report of available cores | 
 | may not work in certain conditions. Benchmarking the processing | 
 | speed based on the number of jobs may be valuable. | 
 |  | 
 | This is I<experimental>. | 
 |  | 
 |  | 
 | =item B<--job-count|-jc> | 
 |  | 
 | Print job and core information that would be used if | 
 | C<-1> was passed to C<--jobs>. | 
 |  | 
 |  | 
 | =item B<--koral|-k> | 
 |  | 
 | Version of the output format. Supported versions are: | 
 | C<0> for legacy serialization, C<0.03> for serialization | 
 | with metadata fields as key-values on the root object, | 
 | C<0.4> for serialization with metadata fields as a list | 
 | of C<"@type":"koral:field"> objects. | 
 |  | 
 | Currently defaults to C<0.03>. | 
 |  | 
 |  | 
 | =item B<--sequential-extraction|-se> | 
 |  | 
 | Flag to indicate, if the C<jobs> value also applies to extraction. | 
 | Some systems may have problems with extracting multiple archives | 
 | to the same folder at the same time. | 
 | Can be flagged using C<--no-sequential-extraction> as well. | 
 | Defaults to C<false>. | 
 |  | 
 |  | 
 | =item B<--meta|-m> | 
 |  | 
 | Define the metadata parser to use. Defaults to C<I5>. | 
 | Metadata parsers can be defined in the C<KorAP::XML::Meta> namespace. | 
 | This is I<experimental>. | 
 |  | 
 |  | 
 | =item B<--gzip|-z> | 
 |  | 
 | Compress the output. | 
 | Expects a defined C<output> file in single processing. | 
 |  | 
 |  | 
 | =item B<--cache|-c> | 
 |  | 
 | File to mmap a cache (using L<Cache::FastMmap>). | 
 | Defaults to C<korapxml2krill.cache> in the calling directory. | 
 |  | 
 |  | 
 | =item B<--cache-size|-cs> | 
 |  | 
 | Size of the cache. Defaults to C<50m>. | 
 |  | 
 |  | 
 | =item B<--cache-init|-ci> | 
 |  | 
 | Initialize cache file. | 
 | Can be flagged using C<--no-cache-init> as well. | 
 | Defaults to C<true>. | 
 |  | 
 |  | 
 | =item B<--cache-delete|-cd> | 
 |  | 
 | Delete cache file after processing. | 
 | Can be flagged using C<--no-cache-delete> as well. | 
 | Defaults to C<true>. | 
 |  | 
 |  | 
 | =item B<--config|-cfg> | 
 |  | 
 | Configure the parameters of your call in a file | 
 | of key-value pairs with whitespace separator | 
 |  | 
 |   overwrite 1 | 
 |   token     DeReKo#Structure | 
 |   ... | 
 |  | 
 | Supported parameters are: | 
 | C<overwrite>, C<gzip>, C<jobs>, C<input-base>, | 
 | C<token>, C<log>, | 
 | C<cache>, C<cache-size>, C<cache-init>, C<cache-delete>, C<meta>, | 
 | C<output>, C<koral>, | 
 | C<temporary-extract>, C<sequential-extraction>, | 
 | C<base-sentences>, C<base-paragraphs>, | 
 | C<base-pagebreaks>, | 
 | C<skip> (semicolon separated), C<sigle> | 
 | (semicolon separated), C<anno> (semicolon separated). | 
 |  | 
 | Configuration parameters will always be overwritten by | 
 | passed parameters. | 
 |  | 
 |  | 
 | =item B<--temporary-extract|-te> | 
 |  | 
 | Only valid for the C<archive> and C<serial> | 
 | commands. | 
 |  | 
 | This will first extract all files into a | 
 | directory and then will archive. | 
 | If the directory is given as C<:temp:>, | 
 | a temporary directory is used. | 
 | This is especially useful to avoid | 
 | massive unzipping and potential | 
 | network latency. | 
 |  | 
 |  | 
 | =item B<--to-tar> | 
 |  | 
 | Only valid for the C<archive> command. | 
 |  | 
 | Writes the output into a tar archive. | 
 |  | 
 |  | 
 | =item B<--sigle|-sg> | 
 |  | 
 | Extract the given texts. | 
 | Can be set multiple times. | 
 | I<Currently only supported on C<extract>.> | 
 | Sigles have the structure C<Corpus>/C<Document>/C<Text>. | 
 | In case the C<Text> path is omitted, the whole document will be extracted. | 
 | On the document level, the postfix wildcard C<*> is supported. | 
 |  | 
 | =item B<--lang> | 
 |  | 
 | Preferred language for metadata fields. In case multiple titles are | 
 | given (on any level) with different C<xml:lang> attributes, | 
 | the language given is preferred. | 
 | Because titles may have different sources and different priorities, | 
 | non-specific language titles may still be preferred in case the title | 
 | source has a higher priority. | 
 |  | 
 |  | 
 | =item B<--log|-l> | 
 |  | 
 | The L<Log::Any> log level, defaults to C<ERROR>. | 
 |  | 
 |  | 
 | =item B<--quiet> | 
 |  | 
 | Silence all information (non-log) outputs. | 
 |  | 
 |  | 
 | =item B<--help|-h> | 
 |  | 
 | Print help information. | 
 |  | 
 |  | 
 | =item B<--version|-v> | 
 |  | 
 | Print version information. | 
 |  | 
 | =back | 
 |  | 
 |  | 
 | =head1 ANNOTATION SUPPORT | 
 |  | 
 | L<KorAP::XML::Krill> has built-in importer for some annotation foundries and layers | 
 | developed in the KorAP project that are part of the KorAP preprocessing pipeline. | 
 | The base foundry with paragraphs, sentences, and the text element are mandatory for | 
 | L<Krill|https://github.com/KorAP/Krill>. | 
 |  | 
 |   Base | 
 |     #Paragraphs | 
 |     #Sentences | 
 |  | 
 |   Connexor | 
 |     #Morpho | 
 |     #Phrase | 
 |     #Sentences | 
 |     #Syntax | 
 |  | 
 |   CoreNLP | 
 |     #Constituency | 
 |     #Morpho | 
 |     #NamedEntities | 
 |     #Sentences | 
 |  | 
 |   CorpusExplorer | 
 |     #Morpho | 
 |  | 
 |   CMC | 
 |     #Morpho | 
 |  | 
 |   DeReKo | 
 |     #Structure | 
 |  | 
 |   DGD | 
 |     #Morpho | 
 |     #Structure | 
 |  | 
 |   DRuKoLa | 
 |     #Morpho | 
 |  | 
 |   Glemm | 
 |     #Morpho | 
 |  | 
 |   Gingko | 
 |     #Morpho | 
 |  | 
 |   HNC | 
 |     #Morpho | 
 |  | 
 |   LWC | 
 |     #Dependency | 
 |  | 
 |   Malt | 
 |     #Dependency | 
 |  | 
 |   MarMoT | 
 |     #Morpho | 
 |  | 
 |   Mate | 
 |     #Dependency | 
 |     #Morpho | 
 |  | 
 |   MDParser | 
 |     #Dependency | 
 |  | 
 |   NKJP | 
 |     #Morpho | 
 |     #NamedEntities | 
 |  | 
 |   OpenNLP | 
 |     #Morpho | 
 |     #Sentences | 
 |  | 
 |   RWK | 
 |     #Morpho | 
 |     #Structure | 
 |  | 
 |   Sgbr | 
 |     #Lemma | 
 |     #Morpho | 
 |  | 
 |   Spacy | 
 |     #Morpho | 
 |  | 
 |   Talismane | 
 |     #Dependency | 
 |     #Morpho | 
 |  | 
 |   TreeTagger | 
 |     #Morpho | 
 |     #Sentences | 
 |  | 
 |   UDPipe | 
 |     #Dependency | 
 |     #Morpho | 
 |  | 
 |   XIP | 
 |     #Constituency | 
 |     #Morpho | 
 |     #Sentences | 
 |  | 
 |  | 
 | More importers are in preparation. | 
 | New annotation importers can be defined in the C<KorAP::XML::Annotation> namespace. | 
 | See the built-in annotation importers as examples. | 
 |  | 
 |  | 
 | =head1 METADATA SUPPORT | 
 |  | 
 | L<KorAP::XML::Krill> has built-in importer for some meta data variants | 
 | that are part of the KorAP preprocessing pipeline. | 
 |  | 
 | =over 2 | 
 |  | 
 | =item B<I5> | 
 |  | 
 | Meta data for all I5 files | 
 |  | 
 | =item B<Sgbr> | 
 |  | 
 | Meta data from the Schreibgebrauch project | 
 |  | 
 | =item B<Gingko> | 
 |  | 
 | Meta data from the Gingko project in addition to I5 | 
 |  | 
 | =item B<ICC> | 
 |  | 
 | Meta data for the ICC in addition to I5 | 
 |  | 
 | =item B<NKJP> | 
 |  | 
 | Meta data for the NKJP corpora | 
 |  | 
 | =back | 
 |  | 
 | New meta data importers can be defined in the C<KorAP::XML::Meta> namespace. | 
 | See the built-in meta data importers as examples. | 
 |  | 
 | The I5 metadata definition is based on TEI-P5 and supports C<E<lt>xenoDataE<gt>> | 
 | with C<E<lt>meta<gt>> elements like | 
 |  | 
 |   <meta type="..." name="..." project="..." desc="...">...</meta> | 
 |  | 
 | that are directly translated to Krill objects. The supported values are: | 
 |  | 
 | =over 2 | 
 |  | 
 | =item C<type> | 
 |  | 
 | =over 4 | 
 |  | 
 | =item C<string> | 
 |  | 
 | String meta data value | 
 |  | 
 | =item C<keyword> | 
 |  | 
 | String meta data value, that can be given multiple times | 
 |  | 
 | =item C<text> | 
 |  | 
 | String meta data value, that is tokenized and can be searched as token sequences | 
 |  | 
 | =item C<date> | 
 |  | 
 | Date meta data value (as "yyyy/mm/dd" with optional granularity) | 
 |  | 
 | =item C<integer> | 
 |  | 
 | Numerical meta data value | 
 |  | 
 | =item C<attachement> | 
 |  | 
 | Non-indexed meta data value (only retrievable) | 
 |  | 
 | =item C<uri> | 
 |  | 
 | Non-indexed attached URI, takes the desc as the title for links | 
 |  | 
 | =back | 
 |  | 
 | =item C<name> | 
 |  | 
 | The key of the meta object, that may be prefixed by C<corpus> or C<doc>, in case the | 
 | C<E<lt>xenoDataE<gt>> information is located on these levels. The text level introduces | 
 | no prefixes. | 
 |  | 
 | =item C<project> (optional) | 
 |  | 
 | A prefixed namespace of the key | 
 |  | 
 | =item C<desc> (optional) | 
 |  | 
 | A description of the key | 
 |  | 
 | =item text content | 
 |  | 
 | The value of the meta object | 
 |  | 
 | =back | 
 |  | 
 |  | 
 | =head1 About KorAP-XML | 
 |  | 
 | KorAP-XML (Bański et al. 2012) is an implementation of the KorAP | 
 | data model (Bański et al. 2013), where text data are stored physically | 
 | separated from their interpretations (i.e. annotations). | 
 | A text document in KorAP-XML therefore consists of several files | 
 | containing primary data, metadata and annotations. | 
 |  | 
 | The structure of a single KorAP-XML document can be as follows: | 
 |  | 
 |   - data.xml | 
 |   - header.xml | 
 |     + base | 
 |       - tokens.xml | 
 |       - ... | 
 |     + struct | 
 |       - structure.xml | 
 |       - ... | 
 |     + corenlp | 
 |       - morpho.xml | 
 |       - constituency.xml | 
 |       - ... | 
 |     + tree_tagger | 
 |       - morpho.xml | 
 |       - ... | 
 |     - ... | 
 |  | 
 | The C<data.xml> contains the primary data, the C<header.xml> contains | 
 | the metadata, and the annotation layers are stored in subfolders | 
 | like C<base>, C<struct> or C<corenlp> | 
 | (so-called "foundries"; Bański et al. 2013). | 
 |  | 
 | Metadata is available in the TEI-P5 variant I5 | 
 | (Lüngen and Sperberg-McQueen 2012). See the documentation in | 
 | L<KorAP::XML::Meta::I5> for translatable fields. | 
 |  | 
 | Annotations correspond to a variant of the TEI-P5 feature structures | 
 | (TEI Consortium; Lee et al. 2004). | 
 | Annotation feature structures refer to character sequences of the primary text | 
 | inside the C<text> element of the C<data.xml>. | 
 | A single annotation containing the lemma of a token can have the following structure: | 
 |  | 
 |   <span from="0" to="3"> | 
 |     <fs type="lex" xmlns="http://www.tei-c.org/ns/1.0"> | 
 |       <f name="lex"> | 
 |         <fs> | 
 |           <f name="lemma">zum</f> | 
 |         </fs> | 
 |       </f> | 
 |     </fs> | 
 |   </span> | 
 |  | 
 | The C<from> and C<to> attributes are refering to the character span | 
 | in the primary text. | 
 | Depending on the kind of annotation (e.g. token-based, span-based, relation-based), | 
 | the structure may vary. See L<KorAP::XML::Annotation::*> for various | 
 | annotation preprocessors. | 
 |  | 
 | Multiple KorAP-XML documents are organized on three levels following | 
 | the "IDS Textmodell" (Lüngen and Sperberg-McQueen 2012): | 
 | corpus E<gt> document E<gt> text. On each level metadata information | 
 | can be stored, that C<korapxml2krill> will merge to a single metadata | 
 | object per text. A corpus is therefore structured as follows: | 
 |  | 
 |   + <corpus> | 
 |     - header.xml | 
 |     + <document> | 
 |       - header.xml | 
 |       + <text> | 
 |         - data.xml | 
 |         - header.xml | 
 |         - ... | 
 |     - ... | 
 |  | 
 | A single text can be identified by the concatenation of | 
 | the corpus identifier, the document identifier and the text identifier. | 
 | This identifier is called the text sigle | 
 | (e.g. a text with the identifier C<18486> in the document C<060> in the | 
 | corpus C<WPD17> has the text sigle C<WPD17/060/18486>, see C<--sigle>). | 
 |  | 
 | These corpora are often stored in zip files, with which C<korapxml2krill> | 
 | can deal with. Corpora may also be split in multiple zip archives | 
 | (e.g. one zip file per foundry), which is also supported (see C<--input>). | 
 |  | 
 | Examples for KorAP-XML files are included in L<KorAP::XML::Krill> | 
 | in form of a test suite. | 
 | The resulting JSON format merges all annotation layers | 
 | based on a single token stream. | 
 |  | 
 | =head2 References | 
 |  | 
 | Piotr Bański, Cyril Belica, Helge Krause, Marc Kupietz, Carsten Schnober, Oliver Schonefeld, and Andreas Witt (2011): | 
 | KorAP data model: first approximation, December. | 
 |  | 
 | Piotr Bański, Peter M. Fischer, Elena Frick, Erik Ketzan, Marc Kupietz, Carsten Schnober, Oliver Schonefeld and Andreas Witt (2012): | 
 | "The New IDS Corpus Analysis Platform: Challenges and Prospects", | 
 | Proceedings of the Eighth International Conference on Language Resources and Evaluation (LREC 2012). | 
 | L<PDF|http://www.lrec-conf.org/proceedings/lrec2012/pdf/789_Paper.pdf> | 
 |  | 
 | Piotr Bański, Elena Frick, Michael Hanl, Marc Kupietz, Carsten Schnober and Andreas Witt (2013): | 
 | "Robust corpus architecture: a new look at virtual collections and data access", | 
 | Corpus Linguistics 2013. Abstract Book. Lancaster: UCREL, pp. 23-25. | 
 | L<PDF|https://ids-pub.bsz-bw.de/frontdoor/deliver/index/docId/4485/file/Ba%c5%84ski_Frick_Hanl_Robust_corpus_architecture_2013.pdf> | 
 |  | 
 | Kiyong Lee, Lou Burnard, Laurent Romary, Eric de la Clergerie, Thierry Declerck, | 
 | Syd Bauman, Harry Bunt, Lionel Clément, Tomaz Erjavec, Azim Roussanaly and Claude Roux (2004): | 
 | "Towards an international standard on featurestructure representation", | 
 | Proceedings of the fourth International Conference on Language Resources and Evaluation (LREC 2004), | 
 | pp. 373-376. | 
 | L<PDF|http://www.lrec-conf.org/proceedings/lrec2004/pdf/687.pdf> | 
 |  | 
 | Harald Lüngen and C. M. Sperberg-McQueen (2012): | 
 | "A TEI P5 Document Grammar for the IDS Text Model", | 
 | Journal of the Text Encoding Initiative, Issue 3 | November 2012. | 
 | L<PDF|https://journals.openedition.org/jtei/pdf/508> | 
 |  | 
 | TEI Consortium, eds: | 
 | "Feature Structures", | 
 | Guidelines for Electronic Text Encoding and Interchange. | 
 | L<html|https://www.tei-c.org/release/doc/tei-p5-doc/en/html/FS.html> | 
 |  | 
 | =head1 AVAILABILITY | 
 |  | 
 |   https://github.com/KorAP/KorAP-XML-Krill | 
 |  | 
 |  | 
 | =head1 COPYRIGHT AND LICENSE | 
 |  | 
 | Copyright (C) 2015-2024, L<IDS Mannheim|https://www.ids-mannheim.de/> | 
 |  | 
 | Author: L<Nils Diewald|https://www.nils-diewald.de/> | 
 |  | 
 | Contributor: Eliza Margaretha, Marc Kupietz | 
 |  | 
 | L<KorAP::XML::Krill> is developed as part of the L<KorAP|https://korap.ids-mannheim.de/> | 
 | Corpus Analysis Platform at the | 
 | L<Leibniz Institute for the German Language (IDS)|https://www.ids-mannheim.de/>, | 
 | member of the | 
 | L<Leibniz-Gemeinschaft|http://www.leibniz-gemeinschaft.de/>. | 
 |  | 
 | This program is free software published under the | 
 | L<BSD-2 License|https://opensource.org/licenses/BSD-2-Clause>. | 
 |  | 
 | =cut |