| Nils Diewald | 2db9ad0 | 2013-10-29 19:26:43 +0000 | [diff] [blame] | 1 | #!/usr/bin/env perl | 
 | 2 | use strict; | 
 | 3 | use warnings; | 
| Akron | 941c1a6 | 2016-02-23 17:41:41 +0100 | [diff] [blame] | 4 | use FindBin; | 
 | 5 | BEGIN { unshift @INC, "$FindBin::Bin/../lib" }; | 
 | 6 | use File::Spec::Functions qw/catfile catdir/; | 
 | 7 | use Getopt::Long qw/GetOptions :config no_auto_abbrev/; | 
| Nils Diewald | 7364d1f | 2013-11-05 19:26:35 +0000 | [diff] [blame] | 8 | use Benchmark qw/:hireswallclock/; | 
 | 9 | use IO::Compress::Gzip qw/$GzipError/; | 
| Nils Diewald | 2db9ad0 | 2013-10-29 19:26:43 +0000 | [diff] [blame] | 10 | use Log::Log4perl; | 
| Akron | 941c1a6 | 2016-02-23 17:41:41 +0100 | [diff] [blame] | 11 | use Pod::Usage; | 
 | 12 | use Directory::Iterator; | 
| Akron | 93d620e | 2016-02-05 19:40:05 +0100 | [diff] [blame] | 13 | use KorAP::XML::Krill; | 
| Akron | 941c1a6 | 2016-02-23 17:41:41 +0100 | [diff] [blame] | 14 | use KorAP::XML::Archive; | 
| Akron | 93d620e | 2016-02-05 19:40:05 +0100 | [diff] [blame] | 15 | use KorAP::XML::Tokenizer; | 
| Akron | 941c1a6 | 2016-02-23 17:41:41 +0100 | [diff] [blame] | 16 | use Parallel::ForkManager; | 
| Akron | 75ba57d | 2016-03-07 23:36:27 +0100 | [diff] [blame] | 17 | # TODO: use Parallel::Loops | 
| Akron | 93d620e | 2016-02-05 19:40:05 +0100 | [diff] [blame] | 18 |  | 
| Akron | 941c1a6 | 2016-02-23 17:41:41 +0100 | [diff] [blame] | 19 | # CHANGES: | 
 | 20 | # ---------------------------------------------------------- | 
 | 21 | # 2013/11/25 | 
 | 22 | # - Initial release | 
 | 23 | # | 
 | 24 | # 2014/10/29 | 
 | 25 | # - Merges foundry data to create indexer friendly documents | 
 | 26 | # | 
| Akron | 93d620e | 2016-02-05 19:40:05 +0100 | [diff] [blame] | 27 | # 2016/02/04 | 
 | 28 | # - renamed to korapxml2krill | 
 | 29 | # - added Schreibgebrauch support | 
| Akron | 069bd71 | 2016-02-12 19:09:06 +0100 | [diff] [blame] | 30 | # | 
 | 31 | # 2016/02/12 | 
 | 32 | # - fixed foundry skipping | 
| Akron | 941c1a6 | 2016-02-23 17:41:41 +0100 | [diff] [blame] | 33 | # - Support overwrite in archive processing | 
| Akron | 150b29e | 2016-02-14 23:06:48 +0100 | [diff] [blame] | 34 | # | 
 | 35 | # 2016/02/14 | 
 | 36 | # - Added version information | 
| Akron | 941c1a6 | 2016-02-23 17:41:41 +0100 | [diff] [blame] | 37 | # - Added support for archive files | 
 | 38 | # | 
 | 39 | # 2016/02/15 | 
 | 40 | # - Fixed temporary directory bug | 
 | 41 | # - Improved skipping before unzipping | 
 | 42 | # - Added EXPERIMENTAL concurrency support | 
 | 43 | # | 
 | 44 | # 2016/02/23 | 
 | 45 | # - Merge korapxml2krill and korapxml2krill_dir | 
| Akron | e10ad32 | 2016-02-27 10:54:26 +0100 | [diff] [blame] | 46 | # | 
 | 47 | # 2016/02/27 | 
 | 48 | # - Added extract function | 
| Akron | 941c1a6 | 2016-02-23 17:41:41 +0100 | [diff] [blame] | 49 | # ---------------------------------------------------------- | 
| Akron | 069bd71 | 2016-02-12 19:09:06 +0100 | [diff] [blame] | 50 |  | 
| Akron | ee13019 | 2016-03-03 18:24:53 +0100 | [diff] [blame] | 51 | our $LAST_CHANGE = '2016/03/02'; | 
| Akron | 941c1a6 | 2016-02-23 17:41:41 +0100 | [diff] [blame] | 52 | our $LOCAL = $FindBin::Bin; | 
 | 53 | our $VERSION_MSG = <<"VERSION"; | 
 | 54 | Version $KorAP::XML::Krill::VERSION - diewald\@ids-mannheim.de - $LAST_CHANGE | 
 | 55 | VERSION | 
 | 56 |  | 
 | 57 |  | 
 | 58 | # Parse comand | 
 | 59 | my $cmd; | 
 | 60 | our @ARGV; | 
 | 61 | if ($ARGV[0] && index($ARGV[0], '-') != 0) { | 
 | 62 |   $cmd = shift @ARGV; | 
| Akron | 150b29e | 2016-02-14 23:06:48 +0100 | [diff] [blame] | 63 | }; | 
| Akron | 93d620e | 2016-02-05 19:40:05 +0100 | [diff] [blame] | 64 |  | 
| Akron | e10ad32 | 2016-02-27 10:54:26 +0100 | [diff] [blame] | 65 | my (@skip, @sigle); | 
 | 66 |  | 
| Akron | 941c1a6 | 2016-02-23 17:41:41 +0100 | [diff] [blame] | 67 | # Parse options from the command line | 
| Nils Diewald | 7364d1f | 2013-11-05 19:26:35 +0000 | [diff] [blame] | 68 | GetOptions( | 
| Akron | 941c1a6 | 2016-02-23 17:41:41 +0100 | [diff] [blame] | 69 |   'input|i=s'   => \(my $input), | 
 | 70 |   'output|o=s'  => \(my $output), | 
 | 71 |   'overwrite|w' => \(my $overwrite), | 
 | 72 |   'human|m'     => \(my $text), | 
 | 73 |   'token|t=s'   => \(my $token_base), | 
 | 74 |   'gzip|z'      => \(my $gzip), | 
| Akron | e10ad32 | 2016-02-27 10:54:26 +0100 | [diff] [blame] | 75 |   'skip|s=s'    => \@skip, | 
 | 76 |   'sigle|sg=s'  => \@sigle, | 
| Akron | 941c1a6 | 2016-02-23 17:41:41 +0100 | [diff] [blame] | 77 |   'log|l=s'     => \(my $log_level = 'ERROR'), | 
 | 78 |   'allow|a=s'   => \(my @allow), | 
 | 79 |   'primary|p!'  => \(my $primary), | 
 | 80 |   'pretty|y'    => \(my $pretty), | 
 | 81 |   'jobs|j=i'    => \(my $jobs = 0), | 
 | 82 |   'help|h'      => sub { | 
 | 83 |     pod2usage( | 
 | 84 |       -sections => 'NAME|SYNOPSIS|ARGUMENTS|OPTIONS', | 
 | 85 |       -verbose => 99, | 
 | 86 |       -msg => $VERSION_MSG, | 
 | 87 |     ); | 
 | 88 |   }, | 
 | 89 |   'version|v'   => sub { | 
 | 90 |     pod2usage( | 
 | 91 |       -verbose => 0, | 
 | 92 |       -msg => $VERSION_MSG | 
 | 93 |     ) | 
 | 94 |   } | 
| Nils Diewald | 7364d1f | 2013-11-05 19:26:35 +0000 | [diff] [blame] | 95 | ); | 
 | 96 |  | 
| Akron | 941c1a6 | 2016-02-23 17:41:41 +0100 | [diff] [blame] | 97 | my %ERROR_HASH = ( | 
 | 98 |   -sections => 'NAME|SYNOPSIS|ARGUMENTS|OPTIONS', | 
 | 99 |   -verbose => 99, | 
 | 100 |   -msg => $VERSION_MSG, | 
 | 101 |   -exit => 1 | 
 | 102 | ); | 
| Nils Diewald | 7364d1f | 2013-11-05 19:26:35 +0000 | [diff] [blame] | 103 |  | 
| Akron | 941c1a6 | 2016-02-23 17:41:41 +0100 | [diff] [blame] | 104 | # Input has to be defined | 
 | 105 | pod2usage(%ERROR_HASH) unless $input; | 
| Nils Diewald | 7364d1f | 2013-11-05 19:26:35 +0000 | [diff] [blame] | 106 |  | 
| Nils Diewald | 7364d1f | 2013-11-05 19:26:35 +0000 | [diff] [blame] | 107 |  | 
| Akron | 941c1a6 | 2016-02-23 17:41:41 +0100 | [diff] [blame] | 108 | # Initialize log4perl object | 
| Nils Diewald | 7364d1f | 2013-11-05 19:26:35 +0000 | [diff] [blame] | 109 | Log::Log4perl->init({ | 
 | 110 |   'log4perl.rootLogger' => uc($log_level) . ', STDERR', | 
 | 111 |   'log4perl.appender.STDERR' => 'Log::Log4perl::Appender::ScreenColoredLevels', | 
 | 112 |   'log4perl.appender.STDERR.layout' => 'PatternLayout', | 
 | 113 |   'log4perl.appender.STDERR.layout.ConversionPattern' => '[%r] %F %L %c - %m%n' | 
 | 114 | }); | 
 | 115 |  | 
 | 116 | my $log = Log::Log4perl->get_logger('main'); | 
 | 117 |  | 
| Akron | 941c1a6 | 2016-02-23 17:41:41 +0100 | [diff] [blame] | 118 |  | 
 | 119 | # Get file name based on path information | 
 | 120 | sub get_file_name ($) { | 
 | 121 |   my $file = shift; | 
 | 122 |   $file =~ s/^?\/?$input//; | 
 | 123 |   $file =~ tr/\//-/; | 
 | 124 |   $file =~ s{^-+}{}; | 
 | 125 |   return $file; | 
| Nils Diewald | 59094f2 | 2014-11-05 18:20:50 +0000 | [diff] [blame] | 126 | }; | 
 | 127 |  | 
| Akron | 941c1a6 | 2016-02-23 17:41:41 +0100 | [diff] [blame] | 128 |  | 
 | 129 | # Write file | 
 | 130 | sub write_file { | 
 | 131 |   my $anno = shift; | 
 | 132 |   my $file = get_file_name $anno; | 
 | 133 |  | 
 | 134 |   # TODO: This should be done directly with a data structure! KorAP::XML::Wrap | 
 | 135 |  | 
 | 136 |   my $call = 'perl ' . $LOCAL . '/korapxml2krill -i ' . | 
 | 137 |     $anno . ' -o ' . $output . '/' . $file . '.json'; | 
 | 138 |   $call .= '.gz -z' if $gzip; | 
 | 139 |   $call .= ' -m' if $text; | 
 | 140 |   $call .= ' -w' if $overwrite; | 
 | 141 |   $call .= ' -t ' . $token_base if $token_base; | 
 | 142 |   $call .= ' -l ' . $log_level if $log_level; | 
 | 143 |   $call .= ' --no-primary ' if $primary; | 
 | 144 |   $call .= ' -y ' . $pretty if $pretty; | 
 | 145 |   $call .= ' -a ' . $_ foreach @allow; | 
 | 146 |   $call .= ' -s ' . $_ foreach @skip; | 
 | 147 |   system($call); | 
 | 148 |   return "$file"; | 
| Nils Diewald | 7364d1f | 2013-11-05 19:26:35 +0000 | [diff] [blame] | 149 | }; | 
 | 150 |  | 
| Nils Diewald | 2db9ad0 | 2013-10-29 19:26:43 +0000 | [diff] [blame] | 151 |  | 
| Akron | e10ad32 | 2016-02-27 10:54:26 +0100 | [diff] [blame] | 152 | # Convert sigle to path construct | 
 | 153 | s!^\s*([^_]+?)_([^\.]+?)\.(.+?)\s*$!$1/$2/$3! foreach @sigle; | 
 | 154 |  | 
| Akron | 941c1a6 | 2016-02-23 17:41:41 +0100 | [diff] [blame] | 155 | # Process a single file | 
 | 156 | unless ($cmd) { | 
| Nils Diewald | 59094f2 | 2014-11-05 18:20:50 +0000 | [diff] [blame] | 157 |  | 
| Akron | 941c1a6 | 2016-02-23 17:41:41 +0100 | [diff] [blame] | 158 |   # Can't print gzip to STDOUT | 
 | 159 |   pod2usage(%ERROR_HASH) if $gzip && !$output; | 
| Nils Diewald | 2db9ad0 | 2013-10-29 19:26:43 +0000 | [diff] [blame] | 160 |  | 
| Akron | 941c1a6 | 2016-02-23 17:41:41 +0100 | [diff] [blame] | 161 |   my %skip; | 
 | 162 |   $skip{lc($_)} = 1 foreach @skip; | 
| Nils Diewald | 2db9ad0 | 2013-10-29 19:26:43 +0000 | [diff] [blame] | 163 |  | 
 | 164 |  | 
| Akron | 941c1a6 | 2016-02-23 17:41:41 +0100 | [diff] [blame] | 165 |   # Ignore processing | 
 | 166 |   if (!$overwrite && $output && -e $output) { | 
 | 167 |     $log->trace($output . ' already exists'); | 
 | 168 |     exit(0); | 
| Nils Diewald | 7364d1f | 2013-11-05 19:26:35 +0000 | [diff] [blame] | 169 |   }; | 
| Akron | 941c1a6 | 2016-02-23 17:41:41 +0100 | [diff] [blame] | 170 |  | 
 | 171 |   BEGIN { | 
 | 172 |     $main::TIME = Benchmark->new; | 
 | 173 |     $main::LAST_STOP = Benchmark->new; | 
 | 174 |   }; | 
 | 175 |  | 
 | 176 |   sub stop_time { | 
 | 177 |     my $new = Benchmark->new; | 
 | 178 |     $log->trace( | 
 | 179 |       'The code took: '. | 
 | 180 | 	timestr(timediff($new, $main::LAST_STOP)) . | 
 | 181 | 	  ' (overall: ' . timestr(timediff($new, $main::TIME)) . ')' | 
 | 182 | 	); | 
 | 183 |     $main::LAST_STOP = $new; | 
 | 184 |   }; | 
 | 185 |  | 
 | 186 |   # Create and parse new document | 
 | 187 |   $input =~ s{([^/])$}{$1/}; | 
 | 188 |   my $doc = KorAP::XML::Krill->new( path => $input ); | 
 | 189 |  | 
 | 190 |   unless ($doc->parse) { | 
 | 191 |     $log->warn($output . " can't be processed - no document data"); | 
 | 192 |     exit(0); | 
 | 193 |   }; | 
 | 194 |  | 
 | 195 |   my ($token_base_foundry, $token_base_layer) = (qw/OpenNLP Tokens/); | 
 | 196 |   if ($token_base) { | 
 | 197 |     ($token_base_foundry, $token_base_layer) = split /#/, $token_base; | 
 | 198 |   }; | 
 | 199 |  | 
 | 200 |   # Get tokenization | 
 | 201 |   my $tokens = KorAP::XML::Tokenizer->new( | 
 | 202 |     path => $doc->path, | 
 | 203 |     doc => $doc, | 
 | 204 |     foundry => $token_base_foundry, | 
 | 205 |     layer => $token_base_layer, | 
 | 206 |     name => 'tokens' | 
 | 207 |   ); | 
 | 208 |  | 
 | 209 |   # Unable to process base tokenization | 
 | 210 |   unless ($tokens->parse) { | 
 | 211 |     $log->error($output . " can't be processed - no base tokenization"); | 
 | 212 |     exit(0); | 
 | 213 |   }; | 
 | 214 |  | 
 | 215 |   my @layers; | 
 | 216 |   push(@layers, ['Base', 'Sentences']); | 
 | 217 |   push(@layers, ['Base', 'Paragraphs']); | 
 | 218 |  | 
 | 219 |   # Connexor | 
 | 220 |   push(@layers, ['Connexor', 'Morpho']); | 
 | 221 |   push(@layers, ['Connexor', 'Syntax']); | 
 | 222 |   push(@layers, ['Connexor', 'Phrase']); | 
 | 223 |   push(@layers, ['Connexor', 'Sentences']); | 
 | 224 |  | 
 | 225 |   # CoreNLP | 
 | 226 |   push(@layers, ['CoreNLP', 'NamedEntities']); | 
 | 227 |   push(@layers, ['CoreNLP', 'Sentences']); | 
 | 228 |   push(@layers, ['CoreNLP', 'Morpho']); | 
 | 229 |   push(@layers, ['CoreNLP', 'Constituency']); | 
 | 230 |  | 
 | 231 |   # DeReKo | 
 | 232 |   push(@layers, ['DeReKo', 'Structure']); | 
 | 233 |  | 
 | 234 |   # Glemm | 
 | 235 |   push(@layers, ['Glemm', 'Morpho']); | 
 | 236 |  | 
 | 237 |   # Malt | 
 | 238 |   # push(@layers, ['Malt', 'Dependency']); | 
 | 239 |  | 
 | 240 |   # Mate | 
 | 241 |   push(@layers, ['Mate', 'Morpho']); | 
 | 242 |   push(@layers, ['Mate', 'Dependency']); | 
 | 243 |  | 
 | 244 |   # OpenNLP | 
 | 245 |   push(@layers, ['OpenNLP', 'Morpho']); | 
 | 246 |   push(@layers, ['OpenNLP', 'Sentences']); | 
 | 247 |  | 
 | 248 |   # Schreibgebrauch | 
 | 249 |   push(@layers, ['Sgbr', 'Lemma']); | 
 | 250 |   push(@layers, ['Sgbr', 'Morpho']); | 
 | 251 |  | 
 | 252 |   # TreeTagger | 
 | 253 |   push(@layers, ['TreeTagger', 'Morpho']); | 
 | 254 |   push(@layers, ['TreeTagger', 'Sentences']); | 
 | 255 |  | 
 | 256 |   # XIP | 
 | 257 |   push(@layers, ['XIP', 'Morpho']); | 
 | 258 |   push(@layers, ['XIP', 'Constituency']); | 
 | 259 |   push(@layers, ['XIP', 'Sentences']); | 
 | 260 |   push(@layers, ['XIP', 'Dependency']); | 
 | 261 |  | 
 | 262 |  | 
 | 263 |   if ($skip{'#all'}) { | 
 | 264 |     foreach (@allow) { | 
 | 265 |       $tokens->add(split('#', $_)); | 
| Nils Diewald | 7364d1f | 2013-11-05 19:26:35 +0000 | [diff] [blame] | 266 |       stop_time; | 
| Nils Diewald | 2db9ad0 | 2013-10-29 19:26:43 +0000 | [diff] [blame] | 267 |     }; | 
| Nils Diewald | 7364d1f | 2013-11-05 19:26:35 +0000 | [diff] [blame] | 268 |   } | 
 | 269 |   else { | 
| Akron | 941c1a6 | 2016-02-23 17:41:41 +0100 | [diff] [blame] | 270 |     # Add to index file - respect skipping | 
 | 271 |     foreach my $info (@layers) { | 
 | 272 |       # Skip if Foundry or Foundry#Layer should be skipped | 
 | 273 |       unless ($skip{lc($info->[0])} || $skip{lc($info->[0]) . '#' . lc($info->[1])}) { | 
 | 274 | 	$tokens->add(@$info); | 
 | 275 | 	stop_time; | 
 | 276 |       }; | 
 | 277 |     }; | 
| Nils Diewald | 2db9ad0 | 2013-10-29 19:26:43 +0000 | [diff] [blame] | 278 |   }; | 
 | 279 |  | 
| Akron | 941c1a6 | 2016-02-23 17:41:41 +0100 | [diff] [blame] | 280 |   my $file; | 
 | 281 |  | 
 | 282 |   my $print_text = $text ? $tokens->to_string($primary) : | 
 | 283 |     ($pretty ? $tokens->to_pretty_json($primary) : $tokens->to_json($primary)); | 
 | 284 |  | 
 | 285 |   if ($output) { | 
 | 286 |  | 
 | 287 |     if ($gzip) { | 
 | 288 |       $file = IO::Compress::Gzip->new($output, Minimal => 1); | 
 | 289 |     } | 
 | 290 |     else { | 
 | 291 |       $file = IO::File->new($output, "w"); | 
 | 292 |     }; | 
 | 293 |  | 
 | 294 |     $file->print($print_text); | 
 | 295 |     $file->close; | 
 | 296 |   } | 
 | 297 |  | 
 | 298 |   else { | 
 | 299 |     print $print_text . "\n"; | 
 | 300 |   }; | 
 | 301 |  | 
 | 302 |   stop_time; | 
| Nils Diewald | 7364d1f | 2013-11-05 19:26:35 +0000 | [diff] [blame] | 303 | } | 
| Nils Diewald | 59094f2 | 2014-11-05 18:20:50 +0000 | [diff] [blame] | 304 |  | 
| Akron | e10ad32 | 2016-02-27 10:54:26 +0100 | [diff] [blame] | 305 | # Extract XML files | 
 | 306 | elsif ($cmd eq 'extract') { | 
 | 307 |  | 
 | 308 |   pod2usage(%ERROR_HASH) unless $output; | 
 | 309 |  | 
 | 310 |   # TODO: Support sigles and full archives | 
 | 311 |  | 
 | 312 |   if ($output && (!-e $output || !-d $output)) { | 
 | 313 |     print "Directory '$output' does not exist.\n\n"; | 
 | 314 |     exit(0); | 
 | 315 |   }; | 
 | 316 |  | 
 | 317 |   if (-f($input) && (my $archive = KorAP::XML::Archive->new($input))) { | 
 | 318 |  | 
 | 319 |     unless ($archive->test_unzip) { | 
 | 320 |       print "Unzip is not installed or incompatible.\n\n"; | 
 | 321 |       exit(1); | 
 | 322 |     }; | 
 | 323 |  | 
 | 324 |     # Test will be skipped | 
 | 325 |  | 
 | 326 |     # Iterate over all given sigles and extract | 
 | 327 |     foreach (@sigle) { | 
 | 328 |       print "$_ "; | 
 | 329 |       print '' . ($archive->extract('./'. $_, $output) ? '' : 'not '); | 
 | 330 |       print "extracted.\n"; | 
 | 331 |     }; | 
 | 332 |  | 
 | 333 |     print "\n"; | 
 | 334 |     exit(1); | 
 | 335 |   }; | 
 | 336 | } | 
 | 337 |  | 
| Akron | 941c1a6 | 2016-02-23 17:41:41 +0100 | [diff] [blame] | 338 | # Process an archive | 
 | 339 | elsif ($cmd eq 'archive') { | 
| Nils Diewald | 2db9ad0 | 2013-10-29 19:26:43 +0000 | [diff] [blame] | 340 |  | 
| Akron | e10ad32 | 2016-02-27 10:54:26 +0100 | [diff] [blame] | 341 |   # TODO: Support sigles | 
 | 342 |  | 
| Akron | 941c1a6 | 2016-02-23 17:41:41 +0100 | [diff] [blame] | 343 |   pod2usage(%ERROR_HASH) unless $output; | 
 | 344 |  | 
 | 345 |   if ($output && (!-e $output || !-d $output)) { | 
 | 346 |     print "Directory '$output' does not exist.\n\n"; | 
 | 347 |     exit(0); | 
 | 348 |   }; | 
 | 349 |  | 
 | 350 |   # Zero means: everything runs in the parent process | 
 | 351 |   my $pool = Parallel::ForkManager->new($jobs); | 
 | 352 |  | 
 | 353 |   my $count = 0; # Texts to process | 
 | 354 |   my $iter  = 1;  # Current text in process | 
 | 355 |  | 
 | 356 |   # Report on fork message | 
 | 357 |   $pool->run_on_finish ( | 
 | 358 |     sub { | 
 | 359 |       my ($pid, $code) = shift; | 
 | 360 |       my $data = pop; | 
 | 361 |       print 'Convert ['. ($jobs > 0 ? "$pid:" : '') . | 
 | 362 | 	($iter++) . "/$count]" . | 
 | 363 | 	  ($code ? " $code" : '') . | 
 | 364 | 	    " $$data\n"; | 
 | 365 |     } | 
 | 366 |   ); | 
 | 367 |  | 
 | 368 |   my $t; | 
 | 369 |   print "Reading data ...\n"; | 
 | 370 |  | 
 | 371 |   # Input is a directory | 
 | 372 |   if (-d $input) { | 
 | 373 |     my $it = Directory::Iterator->new($input); | 
 | 374 |     my @dirs; | 
 | 375 |     my $dir; | 
 | 376 |  | 
 | 377 |     while (1) { | 
 | 378 |       if (!$it->is_directory && ($dir = $it->get) && $dir =~ s{/data\.xml$}{}) { | 
 | 379 | 	push @dirs, $dir; | 
 | 380 | 	$it->prune; | 
 | 381 |       }; | 
 | 382 |       last unless $it->next; | 
 | 383 |     }; | 
 | 384 |  | 
 | 385 |     print "Start processing ...\n"; | 
 | 386 |     $t = Benchmark->new; | 
 | 387 |     $count = scalar @dirs; | 
 | 388 |  | 
 | 389 |   DIRECTORY_LOOP: | 
 | 390 |     for (my $i = 0; $i < $count; $i++) { | 
 | 391 |  | 
 | 392 |       unless ($overwrite) { | 
 | 393 | 	my $filename = catfile( | 
 | 394 | 	  $output, | 
 | 395 | 	  get_file_name($dirs[$i]) . '.json' . ($gzip ? '.gz' : '') | 
 | 396 | 	); | 
 | 397 |  | 
 | 398 | 	if (-e $filename) { | 
 | 399 | 	  $iter++; | 
 | 400 | 	  print "Skip $filename\n"; | 
 | 401 | 	  next; | 
 | 402 | 	}; | 
 | 403 |       }; | 
 | 404 |  | 
 | 405 |       # Get the next fork | 
 | 406 |       my $pid = $pool->start and next DIRECTORY_LOOP; | 
 | 407 |       my $msg; | 
 | 408 |  | 
 | 409 |       $msg = write_file($dirs[$i]); | 
 | 410 |       $pool->finish(0, \$msg); | 
 | 411 |     }; | 
 | 412 |   } | 
 | 413 |  | 
 | 414 |   # Input is a file | 
 | 415 |   elsif (-f($input) && (my $archive = KorAP::XML::Archive->new($input))) { | 
 | 416 |     unless ($archive->test_unzip) { | 
 | 417 |       print "Unzip is not installed or incompatible.\n\n"; | 
 | 418 |       exit(1); | 
 | 419 |     }; | 
 | 420 |  | 
 | 421 |     unless ($archive->test) { | 
 | 422 |       print "Zip archive not compatible.\n\n"; | 
 | 423 |       exit(1); | 
 | 424 |     }; | 
 | 425 |  | 
 | 426 |     print "Start processing ...\n"; | 
 | 427 |     $t = Benchmark->new; | 
 | 428 |     my @dirs = $archive->list_texts; | 
 | 429 |     $count = scalar @dirs; | 
 | 430 |  | 
 | 431 |   ARCHIVE_LOOP: | 
 | 432 |     for (my $i = 0; $i < $count; $i++) { | 
 | 433 |  | 
 | 434 |       # Split path information | 
 | 435 |       my ($prefix, $corpus, $doc, $text) = $archive->split_path($dirs[$i]); | 
 | 436 |  | 
 | 437 |       unless ($overwrite) { | 
 | 438 | 	my $filename = catfile( | 
 | 439 | 	  $output, | 
 | 440 | 	  get_file_name(catdir($doc, $text)) . '.json' . ($gzip ? '.gz' : '') | 
 | 441 | 	); | 
 | 442 |  | 
 | 443 | 	if (-e $filename) { | 
 | 444 | 	  $iter++; | 
 | 445 | 	  print "Skip $filename\n"; | 
 | 446 | 	  next; | 
 | 447 | 	}; | 
 | 448 |       }; | 
 | 449 |  | 
 | 450 |       # Get the next fork | 
 | 451 |       my $pid = $pool->start and next ARCHIVE_LOOP; | 
 | 452 |  | 
 | 453 |       # Create temporary file | 
 | 454 |       my $temp = File::Temp->newdir; | 
 | 455 |  | 
 | 456 |       my $msg; | 
 | 457 |  | 
 | 458 |       # Extract from archive | 
 | 459 |       if ($archive->extract($dirs[$i], $temp)) { | 
 | 460 |  | 
 | 461 | 	# Create corpus directory | 
 | 462 | 	$input = catdir("$temp", $corpus); | 
 | 463 |  | 
 | 464 | 	# Temporary directory | 
 | 465 | 	my $dir = catdir($input, $doc, $text); | 
 | 466 |  | 
 | 467 | 	# Write file | 
 | 468 | 	$msg = write_file($dir); | 
 | 469 |  | 
 | 470 | 	$temp = undef; | 
 | 471 | 	$pool->finish(0, \$msg); | 
 | 472 |       } | 
 | 473 |       else { | 
 | 474 |  | 
 | 475 | 	$temp = undef; | 
 | 476 | 	$msg = "Unable to extract " . $dirs[$i] . "\n"; | 
 | 477 | 	$pool->finish(1, \$msg); | 
 | 478 |       }; | 
 | 479 |     }; | 
 | 480 |   } | 
 | 481 |  | 
 | 482 |   else { | 
 | 483 |     print "Input is neither a directory nor an archive.\n\n"; | 
 | 484 |   }; | 
 | 485 |  | 
 | 486 |   $pool->wait_all_children; | 
 | 487 |  | 
 | 488 |   print "Done.\n"; | 
 | 489 |   print timestr(timediff(Benchmark->new, $t))."\n\n"; | 
 | 490 | } | 
 | 491 |  | 
 | 492 | # Unknown command | 
 | 493 | else { | 
 | 494 |   warn "Unknown command '$cmd'.\n\n"; | 
 | 495 |   pod2usage(%ERROR_HASH); | 
 | 496 | } | 
| Nils Diewald | 2db9ad0 | 2013-10-29 19:26:43 +0000 | [diff] [blame] | 497 |  | 
 | 498 | __END__ | 
| Akron | 941c1a6 | 2016-02-23 17:41:41 +0100 | [diff] [blame] | 499 |  | 
 | 500 | =pod | 
 | 501 |  | 
 | 502 | =encoding utf8 | 
 | 503 |  | 
 | 504 | =head1 NAME | 
 | 505 |  | 
 | 506 | korapxml2krill - Merge KorapXML data and create Krill friendly documents | 
 | 507 |  | 
 | 508 |  | 
 | 509 | =head1 SYNOPSIS | 
 | 510 |  | 
 | 511 |   $ korapxml2krill [archive] -z --input <directory> --output <filename> | 
 | 512 |  | 
 | 513 |  | 
 | 514 | =head1 DESCRIPTION | 
 | 515 |  | 
 | 516 | L<KorAP::XML::Krill> is a library to convert KorAP-XML documents to files | 
 | 517 | compatible with the L<Krill|https://github.com/KorAP/Krill> indexer. | 
 | 518 |  | 
 | 519 |  | 
 | 520 | =head1 INSTALLATION | 
 | 521 |  | 
 | 522 | The preferred way to install L<KorAP::XML::Krill> is to use L<cpanm|App::cpanminus>. | 
 | 523 |  | 
 | 524 |   $ cpanm https://github.com/KorAP/KorAP-XML-Krill | 
 | 525 |  | 
 | 526 | In case everything went well, the C<korapxml2krill> command line tool will | 
 | 527 | be available. | 
 | 528 |  | 
 | 529 |  | 
 | 530 | =head1 ARGUMENTS | 
 | 531 |  | 
 | 532 | =over 2 | 
 | 533 |  | 
 | 534 | =item B<archive> | 
 | 535 |  | 
| Akron | e10ad32 | 2016-02-27 10:54:26 +0100 | [diff] [blame] | 536 | Process an archive as a Zip-file or a folder of KorAP-XML documents. | 
 | 537 |  | 
 | 538 | =item B<extract> | 
 | 539 |  | 
 | 540 | Extract KorAP-XML files from a Zip-file. | 
| Akron | 941c1a6 | 2016-02-23 17:41:41 +0100 | [diff] [blame] | 541 |  | 
 | 542 | =back | 
 | 543 |  | 
 | 544 |  | 
 | 545 | =head1 OPTIONS | 
 | 546 |  | 
 | 547 | =over 2 | 
 | 548 |  | 
 | 549 | =item B<--input|-i> <directory|file> | 
 | 550 |  | 
 | 551 | Directory or archive file of documents to index. | 
 | 552 |  | 
 | 553 | =item B<--output|-o> <directory|file> | 
 | 554 |  | 
 | 555 | Output folder for archive processing or | 
 | 556 | document name for single output (optional), | 
 | 557 | writes to <STDOUT> by default. | 
 | 558 |  | 
 | 559 | =item B<--overwrite|-w> | 
 | 560 |  | 
 | 561 | Overwrite files that already exist. | 
 | 562 |  | 
 | 563 | =item B<--token|-t> <foundry>[#<file>] | 
 | 564 |  | 
 | 565 | Define the default tokenization by specifying | 
 | 566 | the name of the foundry and optionally the name | 
 | 567 | of the layer-file. Defaults to OpenNLP#tokens. | 
 | 568 |  | 
 | 569 | =item B<--skip|-s> <foundry>[#<layer>] | 
 | 570 |  | 
 | 571 | Skip specific foundries by specifying the name | 
 | 572 | or specific layers by defining the name | 
 | 573 | with a # in front of the foundry, | 
 | 574 | e.g. Mate#Morpho. Alternatively you can skip #ALL. | 
 | 575 | Can be set multiple times. | 
 | 576 |  | 
 | 577 | =item B<--allow|-a> <foundry>#<layer> | 
 | 578 |  | 
 | 579 | Allow specific foundries and layers by defining them | 
 | 580 | combining the foundry name with a # and the layer name. | 
 | 581 |  | 
 | 582 | =item B<--primary|-p> | 
 | 583 |  | 
 | 584 | Output primary data or not. Defaults to true. | 
 | 585 | Can be flagged using --no-primary as well. | 
 | 586 |  | 
 | 587 | =item B<--jobs|-j> | 
 | 588 |  | 
 | 589 | Define the number of concurrent jobs in seperated forks | 
 | 590 | for archive processing, defaults to 0. This is B<EXPERIMENTAL>! | 
 | 591 |  | 
 | 592 | =item B<--human|-m> | 
 | 593 |  | 
 | 594 | Represent the data human friendly, while the output defaults to JSON. | 
 | 595 |  | 
 | 596 | =item B<--pretty|-y> | 
 | 597 |  | 
 | 598 | Pretty print JSON output. | 
 | 599 |  | 
 | 600 | =item B<--gzip|-z> | 
 | 601 |  | 
 | 602 | Compress the output (expects a defined output file in single processing). | 
 | 603 |  | 
| Akron | e10ad32 | 2016-02-27 10:54:26 +0100 | [diff] [blame] | 604 | =item B<--sigle|-sg> | 
 | 605 |  | 
 | 606 | Extract the given text sigles. | 
 | 607 | Currently only supported on C<extract>. | 
 | 608 | Can be set multiple times. | 
 | 609 |  | 
| Akron | 941c1a6 | 2016-02-23 17:41:41 +0100 | [diff] [blame] | 610 | =item B<--log|-l> | 
 | 611 |  | 
 | 612 | The L<Log4perl> log level, defaults to C<ERROR>. | 
 | 613 |  | 
 | 614 | =item B<--help|-h> | 
 | 615 |  | 
 | 616 | Print this document. | 
 | 617 |  | 
 | 618 | =item B<--version|-v> | 
 | 619 |  | 
 | 620 | Print version information. | 
 | 621 |  | 
 | 622 | =back | 
 | 623 |  | 
 | 624 | =head1 AVAILABILITY | 
 | 625 |  | 
 | 626 |   https://github.com/KorAP/KorAP-XML-Krill | 
 | 627 |  | 
 | 628 |  | 
 | 629 | =head1 COPYRIGHT AND LICENSE | 
 | 630 |  | 
 | 631 | Copyright (C) 2015-2016, L<IDS Mannheim|http://www.ids-mannheim.de/> | 
 | 632 | Author: L<Nils Diewald|http://nils-diewald.de/> | 
 | 633 |  | 
 | 634 | L<KorAP::XML::Krill> is developed as part of the L<KorAP|http://korap.ids-mannheim.de/> | 
 | 635 | Corpus Analysis Platform at the | 
 | 636 | L<Institute for the German Language (IDS)|http://ids-mannheim.de/>, | 
 | 637 | member of the | 
 | 638 | L<Leibniz-Gemeinschaft|http://www.leibniz-gemeinschaft.de/en/about-us/leibniz-competition/projekte-2011/2011-funding-line-2/>. | 
 | 639 |  | 
 | 640 | This program is free software published under the | 
 | 641 | L<BSD-2 License|https://raw.githubusercontent.com/KorAP/KorAP-XML-Krill/master/LICENSE>. | 
 | 642 |  | 
 | 643 | =cut |