Remove Euralex from Dockerfile to make it more neutral and reusable

Change-Id: I780db5322f09a41acbd1534cad958daaaf2e46f5
diff --git a/Dockerfile b/Dockerfile
index f63d0f7..8958fe9 100644
--- a/Dockerfile
+++ b/Dockerfile
@@ -1,6 +1,6 @@
 FROM --platform=linux/amd64 debian:bookworm-slim
 
-WORKDIR /euralex
+WORKDIR /tokenbench
 
 RUN echo "Dies ist ein Test. Also, nur ein Beispiel." > example.txt
 
@@ -108,7 +108,7 @@
 
 RUN echo "JTok\n" && \
     cd ./JTok/bin && \
-    sh tokenize /euralex/example.txt de
+    sh tokenize /tokenbench/example.txt de
 
 
 ##################
@@ -136,7 +136,7 @@
     make && \
     make install && \
     ldconfig && \
-    echo "abbrevs /euralex/Waste/de-dstar-dtiger/abbr.lex\nstopwords /euralex/Waste/de-dstar-dtiger/stop.lex\nconjunctions /euralex/Waste/de-dstar-dtiger/conj.lex\nmodel /euralex/Waste/de-dstar-dtiger/model.hmm" > /euralex/Waste/waste.rc
+    echo "abbrevs /tokenbench/Waste/de-dstar-dtiger/abbr.lex\nstopwords /tokenbench/Waste/de-dstar-dtiger/stop.lex\nconjunctions /tokenbench/Waste/de-dstar-dtiger/conj.lex\nmodel /tokenbench/Waste/de-dstar-dtiger/model.hmm" > /tokenbench/Waste/waste.rc
 
 RUN echo "Waste\n" && cat ./example.txt | waste -N --rcfile=./Waste/waste.rc
 
@@ -145,7 +145,7 @@
 # Install nnsplit #
 ###################
 
-COPY nnsplit_bench /euralex/nnsplit_bench/
+COPY nnsplit_bench /tokenbench/nnsplit_bench/
 
 RUN apt-get update && apt-get install -y cargo
 
@@ -189,7 +189,7 @@
 RUN pip3 install click==8.0.4 && \
     pip3 install -U spacy==3.2.3
 
-COPY spacy /euralex/spacy/
+COPY spacy /tokenbench/spacy/
 
 RUN echo "SpaCy" && python3 ./spacy/spacy_tok.py example.txt
 
@@ -214,7 +214,7 @@
 
 # Run with threads!
 RUN echo "StanfordNLP" && \
-    CLASSPATH=/euralex/stanford-corenlp-4.4.0/* java edu.stanford.nlp.pipeline.StanfordCoreNLP \
+    CLASSPATH=/tokenbench/stanford-corenlp-4.4.0/* java edu.stanford.nlp.pipeline.StanfordCoreNLP \
     -annotators tokenize \
     -tokenize.language=german \
     -file example.txt
@@ -226,7 +226,7 @@
 
 RUN pip3 install cutter-ng==2.5
 
-COPY cutter /euralex/cutter/
+COPY cutter /tokenbench/cutter/
 
 RUN echo "Cutter\n" && python3 ./cutter/cutter.py nosent example.txt
 
@@ -237,7 +237,7 @@
 
 RUN pip3 install -U blingfire==0.1.8
 
-COPY blingfire /euralex/blingfire/
+COPY blingfire /tokenbench/blingfire/
 
 RUN echo "BlingFire\n" && python3 ./blingfire/blingfire_tok.py example.txt
 
@@ -268,18 +268,18 @@
 RUN echo "KorAP-Tokenizer\n" && cat example.txt | java -jar KorAP-Tokenizer/KorAP-Tokenizer.jar -l de -s -
 
 
-RUN useradd -ms /bin/bash euralex
+RUN useradd -ms /bin/bash tokenbench
 
 RUN rm -r ./nnsplit_bench && \
-    rm /euralex/v0.1.zip
+    rm /tokenbench/v0.1.zip
 
-RUN chown euralex:euralex -R /euralex
+RUN chown tokenbench:tokenbench -R /tokenbench
 
-USER euralex
+USER tokenbench
 
-WORKDIR /euralex
+WORKDIR /tokenbench
 
 ENTRYPOINT [ "perl" ]
 
 LABEL maintainer="korap@ids-mannheim.de"
-LABEL description="Tokenizer evaluation for EURALEX"
\ No newline at end of file
+LABEL description="Tokenizer evaluation"
\ No newline at end of file
diff --git a/Readme.md b/Readme.md
index 2d0cf6f..008aa49 100644
--- a/Readme.md
+++ b/Readme.md
@@ -1,4 +1,4 @@
-# EURALEX 2022 - Tokenization Benchmark
+# Tokenization Benchmark
 
 This repository contains benchmark scripts for comparing different tokenizers and sentence segmenters of German.  For trouble-free testing, all tools are provided in a Dockerfile.
 
@@ -7,7 +7,7 @@
 To build the Docker image, run
 
 ```shell
-$ docker build -f Dockerfile -t korap/euralex22 .
+$ docker build -f Dockerfile -t korap/tokenbench .
 ```
 This will create and install an image of approximately 12GB.
 
@@ -18,9 +18,9 @@
 
 ```shell
 $ docker run --rm -i \
-  -v ${PWD}/benchmarks:/euralex/benchmarks \
-  -v ${PWD}/corpus:/euralex/corpus \
-  korap/euralex22 benchmarks/[BENCHMARK-SCRIPT]
+  -v ${PWD}/benchmarks:/tokenbench/benchmarks \
+  -v ${PWD}/corpus:/tokenbench/corpus \
+  korap/tokenbench benchmarks/[BENCHMARK-SCRIPT]
 ```
 
 The supported benchmark scripts are:
@@ -61,8 +61,8 @@
 output folders
 
 ```
--v ${PWD}/output_cmc:/euralex/empirist_cmc
--v ${PWD}/output_web:/euralex/empirist_web
+-v ${PWD}/output_cmc:/tokenbench/empirist_cmc
+-v ${PWD}/output_web:/tokenbench/empirist_web
 ```
 
 ### `ud_tokens.pl`
@@ -125,7 +125,6 @@
 
 ## Results
 
-
 In terms of speed, the native output of the tools was measured, while in terms of accuracy, further reshaping was necessary to make it comparable to the gold standard.