blob: 0fe71be45a1a32762607bbb056fa5a0457b162d0 [file] [log] [blame]
FROM --platform=linux/amd64 debian:bookworm-slim
WORKDIR /euralex
RUN echo "Dies ist ein Test. Also, nur ein Beispiel." > example.txt
RUN apt-get update && \
apt-get install -y git \
wget \
unzip \
perl \
golang
############
# Check WC #
############
RUN echo "WC\n" && wc -w ./example.txt
##################
# Install SoMaJo #
##################
RUN apt-get install -y python3-dev \
python3 \
python3-pip && \
pip3 install SoMaJo
RUN echo "SOMAJO\n" && somajo-tokenizer --split_sentences ./example.txt
###################
# Install OpenNLP #
###################
RUN apt-get install -y openjdk-11-jre
RUN wget https://dlcdn.apache.org/opennlp/opennlp-1.9.4/apache-opennlp-1.9.4-bin.zip && \
unzip apache-opennlp-1.9.4-bin.zip -x apache-opennlp-1.9.4/docs/* && \
rm apache-opennlp-1.9.4-bin.zip && \
mv apache-opennlp-1.9.4 opennlp && \
mkdir ./opennlp/models && \
wget https://dlcdn.apache.org/opennlp/models/ud-models-1.0/opennlp-de-ud-gsd-sentence-1.0-1.9.3.bin && \
wget https://dlcdn.apache.org/opennlp/models/ud-models-1.0/opennlp-de-ud-gsd-tokens-1.0-1.9.3.bin && \
mv opennlp-de-ud-gsd-sentence-1.0-1.9.3.bin ./opennlp/models/ && \
mv opennlp-de-ud-gsd-tokens-1.0-1.9.3.bin ./opennlp/models/
RUN echo "OpenNLP (1)" && cat example.txt | ./opennlp/bin/opennlp SimpleTokenizer
RUN echo "OpenNLP (2)" && cat example.txt | ./opennlp/bin/opennlp TokenizerME ./opennlp/models/opennlp-de-ud-gsd-tokens-1.0-1.9.3.bin
RUN echo "OpenNLP (3)" && cat example.txt | ./opennlp/bin/opennlp SentenceDetector ./opennlp/models/opennlp-de-ud-gsd-sentence-1.0-1.9.3.bin
######################
# Install TreeTagger #
######################
RUN mkdir ./treetagger && \
cd treetagger && \
wget https://cis.uni-muenchen.de/~schmid/tools/TreeTagger/data/tagger-scripts.tar.gz && \
tar -xvzf tagger-scripts.tar.gz && \
rm tagger-scripts.tar.gz
RUN echo "TreeTagger" && cat example.txt | ./treetagger/cmd/utf8-tokenize.perl -a ./treetagger/lib/german-abbreviations
####################
# Install deep-eos #
####################
RUN wget https://github.com/dbmdz/deep-eos/archive/refs/tags/v0.1.zip && \
unzip v0.1.zip && \
mv deep-eos-0.1 deep-eos && \
cd deep-eos && \
wget https://github.com/dbmdz/deep-eos/releases/download/v0.1/cnn-de.model && \
wget https://github.com/dbmdz/deep-eos/releases/download/v0.1/cnn-de.vocab && \
wget https://github.com/dbmdz/deep-eos/releases/download/v0.1/bi-lstm-de.model && \
wget https://github.com/dbmdz/deep-eos/releases/download/v0.1/bi-lstm-de.vocab && \
wget https://github.com/dbmdz/deep-eos/releases/download/v0.1/lstm-de.model && \
wget https://github.com/dbmdz/deep-eos/releases/download/v0.1/lstm-de.vocab
RUN pip3 install --upgrade pip && \
pip3 install --upgrade tensorflow
RUN pip3 install keras
RUN sed -i 's/from keras.utils import plot_model/from tensorflow.keras.utils import plot_model/' ./deep-eos/eos.py
RUN echo "deep-eos (1)" && python3 ./deep-eos/main.py --input-file example.txt --model-filename ./deep-eos/cnn-de.model --vocab-filename ./deep-eos/cnn-de.vocab --eos-marker "§" tag
RUN echo "deep-eos (2)" && python3 ./deep-eos/main.py --input-file example.txt --model-filename ./deep-eos/bi-lstm-de.model --vocab-filename ./deep-eos/bi-lstm-de.vocab --eos-marker "§" tag
RUN echo "deep-eos (3)" && python3 ./deep-eos/main.py --input-file example.txt --model-filename ./deep-eos/lstm-de.model --vocab-filename ./deep-eos/lstm-de.vocab --eos-marker "§" tag
###################
# Install Datok #
###################
RUN wget https://github.com/KorAP/Datok/archive/refs/tags/v0.1.1.zip && \
unzip v0.1.1.zip && \
rm v0.1.1.zip && \
mv Datok-0.1.1 Datok && \
cd Datok && \
go build ./cmd/datok.go
RUN echo "DATOK\n" && cat example.txt | ./Datok/datok tokenize -t ./Datok/testdata/tokenizer.matok -
ENTRYPOINT [ "sh" ]
LABEL maintainer="korap@ids-mannheim.de"
LABEL description="Tokenizer evaluation for EURALEX"