blob: f50a736a4ab339b2705c910d6876e59e80cd5eee [file] [log] [blame]
#!/usr/bin/env perl
use strict;
use warnings;
use POSIX;
use Log::Any '$log';
use Log::Any::Adapter;
use Pod::Usage;
use Getopt::Long qw(GetOptions :config no_auto_abbrev);
use Encode;
my $MAX_SENTENCE_LENGTH=10000;
my $COMMENT_START="#";
my $COMMENT_END="";
my $test=0;
my $text_no=0;
my %opts;
my %plain_texts;
my %sentence_ends;
my %metadata;
my $offsets = 1;
my $comments = 1;
my $extract_metadata = 0;
my @extract_metadata_regex;
my $lm_training_data = 0;
our $VERSION = '0.6.0';
our $VERSION_MSG = "\nkorapxml2conllu - v$VERSION\n";
use constant {
# Set to 1 for minimal more debug output (no need to be parametrized)
DEBUG => $ENV{KORAPXMLCONLLU_DEBUG} // 0
};
GetOptions(
'sigle-pattern|p=s' => \(my $sigle_pattern = ''),
'extract-attributes-regex|e=s' => \(my $extract_attributes_regex = ''),
's-bounds-from-morpho' => \(my $s_bounds_from_morpho = 0),
'log|l=s' => \(my $log_level = 'warn'),
'columns|c=n' => \(my $columns = 10),
'word2vec|lm-training-data!' => \$lm_training_data,
'token-separator|s=s' => \(my $token_separator = "\n"),
'offsets!' => \$offsets,
'comments!' => \$comments,
'extract-metadata-regex|m=s@' => \@extract_metadata_regex,
'help|h' => sub {
pod2usage(
-verbose => 99,
-sections => 'NAME|DESCRIPTION|SYNOPSIS|ARGUMENTS|OPTIONS|EXAMPLES',
-msg => $VERSION_MSG,
-output => '-'
)
},
'version|v' => sub {
pod2usage(
-verbose => 0,
-msg => $VERSION_MSG,
-output => '-'
);
}
);
if (@extract_metadata_regex) {
$extract_metadata = 1;
}
# Establish logger
binmode(STDERR, ':encoding(UTF-8)');
Log::Any::Adapter->set('Stderr', log_level => $log_level);
$log->notice('Debugging is activated') if DEBUG;
if ($lm_training_data) {
$columns = 1;
$comments = 0;
$offsets = 0;
$token_separator = " ";
}
my $docid="";
my ($current_id, $current_from, $current_to, $token);
my $current;
my ($unknown, $known) = (0, 0);
my @current_lines;
my %processedFilenames;
my $zipsiglepattern = (defined($ENV{ZIPSIGLEPATTERN})? $ENV{ZIPSIGLEPATTERN} : "");
my $baseOnly;
my %extras;
my ($ID_idx, $FORM_idx, $LEMMA_idx, $UPOS_idx, $XPOS_idx, $FEATS_idx, $HEAD_idx, $DEPREC_idx, $DEPS_idx, $MISC_idx) = (0..9);
my $UNZIP = `sh -c 'command -v unzip'`;
chomp $UNZIP;
if ($UNZIP eq '') {
warn('No unzip executable found in PATH.');
return 0;
};
foreach my $morpho_zip (@ARGV) {
die "cannot open $morpho_zip" if(! -r $morpho_zip);
my $data_zip = $morpho_zip;
if ($data_zip !~ /\.zip/ && $data_zip =~ /\.conllu?/i) {
open(CONLL, "<$data_zip") or die "cannot open $data_zip";
while(<CONLL>) {
print;
}
close(CONLL);
next;
}
$data_zip =~ s/\.([^\/.]+)\.zip$/.zip/;
my $foundry = $1;
die "cannot open data file $data_zip corresponding to $morpho_zip" if(! -r $data_zip);
my $first=1;
my @conll = ("_") x 10;
my $filename;
$baseOnly = $morpho_zip eq $data_zip;
my ($morphoOrTokenCommand, $plaintextAndStructureCommand);
my $zip_content_pattern = "[sd][ta]*";
if ($extract_metadata) {
$zip_content_pattern = "[sdh][tae]*";
}
if (!$baseOnly) {
$morphoOrTokenCommand = "$UNZIP -c $morpho_zip '*${sigle_pattern}*/morpho.xml' $zipsiglepattern |";
if ($extract_attributes_regex || !$s_bounds_from_morpho) {
$plaintextAndStructureCommand = "$UNZIP -c $data_zip '*${sigle_pattern}*/$zip_content_pattern.xml' $zipsiglepattern |";
} else {
$log->debug("Not reading structure information.");
$plaintextAndStructureCommand = "$UNZIP -c $data_zip '*${sigle_pattern}*/data.xml' $zipsiglepattern |";
}
} else {
$foundry = "base";
$morphoOrTokenCommand = "$UNZIP -l $morpho_zip '*${sigle_pattern}*/morpho.xml' $zipsiglepattern";
if (`$morphoOrTokenCommand` !~ /morpho\.xml/) {
$morphoOrTokenCommand =~ s/morpho\.xml/tokens.xml/;
} else {
$baseOnly = 0;
}
$morphoOrTokenCommand =~ s/-l/-c/;
$morphoOrTokenCommand .= ' |';
$plaintextAndStructureCommand = "$UNZIP -c $data_zip " . "'*${sigle_pattern}*/$zip_content_pattern.xml' " . "$zipsiglepattern |";
}
$log->debug("command to extract annotation and/or tokenization: $morphoOrTokenCommand");
$log->debug("command to extract plain text and structure: $plaintextAndStructureCommand");
open (MORPHO_OR_TOKENPIPE, $morphoOrTokenCommand) or die "cannot unzip $morpho_zip";
open (PLAINTEXTPIPE, $plaintextAndStructureCommand) or die "cannot unzip $data_zip";
print "$COMMENT_START foundry = $foundry$COMMENT_END\n" if ($comments);
while (<MORPHO_OR_TOKENPIPE>) {
if (/^ inflating: (.*)/) {
$filename=$1;
while($processedFilenames{$filename} && !eof(MORPHO_OR_TOKENPIPE)) {
$log->warn("$filename already processed");
while (<MORPHO_OR_TOKENPIPE>) {
last if(/\s+inflating:\s+(.*)/);
}
$filename=$1 if(!eof(MORPHO_OR_TOKENPIPE) && /\s+inflating:\s+(.*)/);
}
} elsif(m@(?:^|\s)docid="([^"]+)"@) {
last if($test && $text_no++ > 3);
if(!$first) {
closeDoc(0);
}
$processedFilenames{$filename}=1;
$docid=$1;
@current_lines=();
$known=$unknown=0;
$current="";
if ($first) {
$first = 0;
}
if(!fetch_plaintext($docid)) { # skip this text
while (<MORPHO_OR_TOKENPIPE>) {
last if(m@</layer>@);
}
}
print STDOUT "$COMMENT_START filename = $filename$COMMENT_END\n$COMMENT_START text_id = $docid$COMMENT_END\n" if($comments);
$log->debug("Analyzing $docid");
} elsif (m@^\s*<f\s+.*name="([^"]+)">([^<]+)</f>@) {
if ($1 eq "lemma") {
$conll[$LEMMA_idx] = $2;
$conll[$LEMMA_idx] =~ s/[\t\n\r]//g; # make sure that lemmas never contain tabs or newlines
if($conll[$LEMMA_idx] eq 'UNKNOWN') {
$conll[$LEMMA_idx] = "--";
$unknown++;
} else {
$known++;
}
} elsif ($1 eq 'pos' || $1 eq "ctag") {
$unknown++;
$conll[$XPOS_idx] = $conll[$UPOS_idx] = $2;
} elsif ($1 eq 'msd') {
$conll[$FEATS_idx] = $2;
} elsif ($1 eq 'certainty') {
$conll[$MISC_idx] = $2;
}
} elsif (/<span /) {
my $last_from = $current_from // -1;
($current_id) = /id="[^0-9]*([^\"]*)"/;
($current_from) = /from="([^\"]*)"/;
($current_to) = /to="([^\"]*)"/;
if($extract_attributes_regex) {
for (my $i = $last_from + 1; $i <= $current_from; $i++) {
if ($extras{$docid}{$i}) {
$current .= $extras{$docid}{$i};
undef $extras{$docid}{$i};
}
}
}
# $log->debug("found span: $current_id $current_from $current_to");
$token = substr($plain_texts{$docid}, $current_from, $current_to - $current_from);
if (!defined $token || length($token) == 0) {
$log->warn("$docid: could not retrieve token at $current_from-$current_to/", length($plain_texts{$docid}), " - ending with: ", substr($plain_texts{$docid},length($plain_texts{$docid})-10));
$token = "_";
}
$token=~s/[\t\n\r]//g; # make sure that tokens never contain tabs or newlines
@conll = ("_") x 10;
$conll[$FORM_idx] = encode("utf-8", $token);
if($baseOnly) {
my @vals = ($current_from, $current_to);
# $log->debug("joining : ", join(" ", @vals));
push @current_lines, \@vals;
$known++;
$conll[$ID_idx] = $#current_lines+1;
if ($columns == 1) {
$current .= "$conll[1]$token_separator" ;
} else {
$current .= join("\t", @conll[0..$columns-1]) . "\n"; # conll columns
}
fetch_plaintext($docid);
if ($sentence_ends{$docid}{$current_to}) {
$log->debug("Using sentence end for $docid \@$current_to");
$current .= "\n";
printTokenRanges();
if ($extract_metadata) {
for (my $i = 0; $i < @extract_metadata_regex; $i++) {
if(!defined($metadata{$docid}[$i])) {
$log->warn("$docid: metadata matching /$extract_metadata_regex[$i]/ was not found, using empty string instead");
$metadata{$docid}[$i]="";
}
print "$metadata{$docid}[$i]\t";
}
}
$current =~ s/ $//;
print STDOUT $current;
$current = "";
$known = 0;
$unknown = 0;
@current_lines = ();
}
}
} elsif (m@^\s*</fs>@) {
my @vals = ($current_from, $current_to);
# $log->debug("joining : ", join(" ", @vals));
push @current_lines, \@vals;
# convert gathered information to CONLL
$conll[$ID_idx] = $#current_lines+1;
if ($columns == 1) {
$current .= "$conll[1]\n";
} else {
$current .= join("\t", @conll[0..$columns-1]) . "\n"; # conll columns
}
if($sentence_ends{$docid}{$current_to} || ($s_bounds_from_morpho && $conll[$XPOS_idx] eq '$.' || ($conll[$XPOS_idx] eq 'SENT' && $token eq '.')) || $known + $unknown >= $MAX_SENTENCE_LENGTH) {
$log->debug("Using sentence end for $docid \@$current_to");
$current .= "\n";
if($known + $unknown > 0) { # only print sentence if it contains some words
printTokenRanges();
print STDOUT $current;
}
$current=""; $known=0; $unknown=0;
@current_lines = ();
}
while (<MORPHO_OR_TOKENPIPE>) {
last if (m@</span>@); # only consider first interpretation
}
}
}
$current .= "\n";
closeDoc(1);
close(MORPHO_OR_TOKENPIPE);
close(PLAINTEXTPIPE);
}
exit;
sub printTokenRanges {
return if(!$offsets);
print "$COMMENT_START start_offsets = ", $current_lines[0]->[0];
foreach my $t (@current_lines) {
print STDOUT " $t->[0]";
}
print "$COMMENT_END\n$COMMENT_START end_offsets = ", $current_lines[$#current_lines]->[1] if($comments);
foreach my $t (@current_lines) {
print STDOUT " $t->[1]";
}
print "$COMMENT_END\n";
}
sub closeDoc {
$log->debug("closing doc");
if($known + $unknown > 0) { # only parse a sentence if it has some words
chomp $current;
chomp $current;
chomp $current;
$current .= "\n\n";
printTokenRanges();
print STDOUT $current;
}
}
# read data.xml to figure out the tokens
# (ideally tokens should also be in in morpho.xml, but they are not)
sub fetch_plaintext {
my ($target_id) = @_;
my $docid;
my $text_started=0;
my $text_count = 0;
my ($current_id, $current_from, $current_to);
if($plain_texts{$target_id} && ($s_bounds_from_morpho || $sentence_ends{$target_id}) && (!$extract_metadata || $metadata{$target_id})) {
$log->debug("Already got $target_id");
return 1;
}
while(<PLAINTEXTPIPE>) {
if ($extract_metadata) {
if (/<textSigle>([^<]+)/) {
$docid = $1;
$docid =~ s@/@_@;
$log->debug("textsigle=$docid");
}
for (my $i=0; $i < @extract_metadata_regex; $i++) {
if ($_ =~ /$extract_metadata_regex[$i]/) {
$metadata{$docid}[$i]=$1;
}
}
}
if(/<raw_text[^>]+docid="([^"]*)/) {
$docid=$1;
$log->debug("Getting plain text for $docid");
$text_started=0;
} elsif(/<raw_text\b/) {
$text_started=0;
} elsif(/(?:^|\s)docid="([^"]*)/) {
$docid=$1;
$log->debug("Getting plain text for $docid");
} elsif(m@<span @) {
($current_id) = /id="[^0-9]*([^\"]*)"/;
($current_from) = /from="([^\"]*)"/;
($current_to) = /to="([^\"]*)"/;
} elsif(m@<f\s[^>]*>s</f>@) {
if ($s_bounds_from_morpho) {
$log->debug("Ignoring sentence end for $docid \@$current_to because of --s-bounds-from-morpho");
} else {
$log->debug("Found sentence end for $docid \@$current_to");
$sentence_ends{$docid}{$current_to} = 1;
}
} elsif($extract_attributes_regex && m@<f\sname="name"[^>]*>([^<]+)</f>@) {
my $current_element = $1;
$log->debug("Looking for matching attributes in $docid");
while(<PLAINTEXTPIPE>) {
last if(m@</fs>@);
if(m@<f\sname="([^"]+)"[^>]*>([^<]+)</f>@) {
my $current_node = "$current_element/$1";
my $value = $2;
if ($current_node =~ /$extract_attributes_regex/) {
$log->debug("Found matching attribute: $docid - $current_node = $value");
$extras{$docid}{$current_from} .= "# $current_node = $value\n";
}
}
}
} elsif (m@<text>(.*)</text>@) {
$_= decode("utf-8", $1, Encode::FB_DEFAULT);
s/&lt;/</go;
s/&gt;/>/go;
s/&amp;/&/go;
tr/…•⋅»«ˮ“”„›‹ʼ‘’‚′‐‑‒–—―⁓⁻₋−﹣-/...""""""'''''''-/;
$plain_texts{$docid} = $_;
last if(!$extract_attributes_regex && ($text_count++ > 1 && $plain_texts{$target_id} && (!$extract_metadata || $metadata{$target_id})));
} elsif (m@<text>(.*)@) {
$_= decode("utf-8", $1, Encode::FB_DEFAULT);
s/&lt;/</go;
s/&gt;/>/go;
s/&amp;/&/go;
tr/…•⋅»«ˮ“”„›‹ʼ‘’‚′‐‑‒–—―⁓⁻₋−﹣-/...""""""'''''''-/;
$plain_texts{$docid} = "$_ ";
$text_started=1;
} elsif ($text_started && m@(.*)</text>@) {
$_= decode("utf-8", $1, Encode::FB_DEFAULT);
s/&lt;/</go;
s/&gt;/>/go;
s/&amp;/&/go;
tr/…•⋅»«ˮ“”„›‹ʼ‘’‚′‐‑‒–—―⁓⁻₋−﹣-/...""""""'''''''-/;
$plain_texts{$docid} .= $_;
$text_started=0;
last if(!$extract_attributes_regex && ($text_count++ > 1 && $plain_texts{$target_id} && (!$extract_metadata || $metadata{$target_id})));
} elsif ($text_started) {
chomp;
$_ = decode("utf-8", $_, Encode::FB_DEFAULT) . ' ';
s/&lt;/</go;
s/&gt;/>/go;
s/&amp;/&/go;
tr/…•⋅»«ˮ“”„›‹ʼ‘’‚′‐‑‒–—―⁓⁻₋−﹣-/...""""""'''''''-/;
$plain_texts{$docid} .= $_;
}
}
$log->debug("Got plain text for $docid");
if(defined($ENV{PLAINTEXTFILTER})) {
if ($plain_texts{$docid} !~ $ENV{PLAINTEXTFILTER}) {
$plain_texts{$docid} = undef;
$log->info("Skipping $docid");
return(undef);
} else {
$log->debug("Using $docid");
}
}
return(1);
}
=pod
=encoding utf8
=head1 NAME
korapxml2conllu - Conversion of KorAP-XML zips to CoNLL-U
=head1 SYNOPSIS
korapxml2conllu zca15.tree_tagger.zip > zca15.conllu
=head1 DESCRIPTION
C<korapxml2conllu> is a script to Convert L<KorAP-XML format|https://github.com/KorAP/KorAP-XML-Krill#about-korap-xml> base or morpho zips to CoNLL(-U) format with all information necessary
for reconstruction in comment lines.
=head1 INSTALLATION
$ cpanm https://github.com/KorAP/KorAP-XML-CoNLL-U.git
=head1 OPTIONS
=over 2
=item B<--sigle-pattern|-p>
Convert only texts from the KorAP XML zip files with folder names (i.e. sigles) matching the glob pattern.
=item B<--extract-attribute-pattern|-e>
Extract element/attribute regular expressions to comments.
=item B<--columns>=I<int> | B<-c> I<int>
Print n columns (default: 10). If n=1, only the token itself is printed.
=item B<--s-bounds-from-morpho>
Get sentence boundary information from tagger output rather than from s annotation in structure.xml files.
=item B<--token-separator>=I<string>
Token separator.
=item B<--(no)comments>
Switch comment printing on or off (default: on).
=item B<--(no)offsets>
Switch offsets printing on or off (default: on).
=item B<--word2vec>
Print output in word2vec (tokenized) one sentence per line format.
=item B<--extract-metadata-regex|-m>
Can be used to extract and print strings from the text headers. Currently only works together with --word2vec option.
For example:
korapxml2conllu -m '<textSigle>([^<.]+)' -m '<creatDate>([^<]{7})' --word2vec t/data/wdf19.zip
Will print the document sigle, year and month of the creation date and one sentence per line, separated by tabs.
=item B<--help|-h>
Print help information.
=item B<--version|-v>
Print version information.
=item B<--log|-l>
Loglevel for I<Log::Any>. Defaults to C<warn>.
=back
=head1 EXAMPLES
=head2 Extract to CoNNL-U with posting and div ids:
korapxml2conllu -e '(posting/id|div/id)' t/data/wdf19.zip
=head2 Extract to word2vec input format metadata columns:
korapxml2conllu --word2vec t/data/wdf19.zip
=head1 COPYRIGHT AND LICENSE
Copyright (C) 2021-2022, L<IDS Mannheim|https://www.ids-mannheim.de/>
Author: Marc Kupietz
Contributors: Nils Diewald
L<KorAP::XML::CoNNL-U> is developed as part of the L<KorAP|https://korap.ids-mannheim.de/>
Corpus Analysis Platform at the
L<Leibniz Institute for the German Language (IDS)|http://ids-mannheim.de/>,
member of the
L<Leibniz-Gemeinschaft|http://www.leibniz-gemeinschaft.de/>.
This program is free software published under the
L<BSD-2 License|https://opensource.org/licenses/BSD-2-Clause>.