)]}'
{
  "commit": "854a11503d3c0e9d22f322c26572d2e032192fc0",
  "tree": "39ea0c4db5401d1097ec1a12ed33d27d217376df",
  "parents": [
    "1d65f9467ab04537821c0d6efd565c49ac3649fb"
  ],
  "author": {
    "name": "Peter Harders",
    "email": "harders@ids-mannheim.de",
    "time": "Wed Jul 22 22:48:02 2020 +0200"
  },
  "committer": {
    "name": "Peter Harders",
    "email": "harders@ids-mannheim.de",
    "time": "Fri Jul 24 20:24:20 2020 +0200"
  },
  "message": "bugfixing Conservative.pm\n\n 1. identified wrong tokenization caused by wrong pattern match ($3)\n    (wrote a test in t/tokenization.t, that shows the wrong tokenization)\n\n 2. removed wrong pattern match ($3) and adjusted test in t/tokenization.t\n\n 3. cleaned up (also changed some comments)\n\n 4. fixed missing tokenization of first punctuation char\n\n 5. exchanged [^A-Za-z0-9] by [\\p{Punct}\\s]\n    (TODO: yet no approp. test found)\n\nChange-Id: Ib494c79c3e6971a57ad874fc62583c625095cf28\n",
  "tree_diff": [
    {
      "type": "modify",
      "old_id": "75b11ebb7c9ea10d8f4edcdd1df953b07e81b336",
      "old_mode": 33188,
      "old_path": "lib/KorAP/XML/TEI/Tokenizer/Aggressive.pm",
      "new_id": "517a90122d1fbcbb47a688e828b3ad28e2813051",
      "new_mode": 33188,
      "new_path": "lib/KorAP/XML/TEI/Tokenizer/Aggressive.pm"
    },
    {
      "type": "modify",
      "old_id": "b3373f5739a09e39a5dabb62b4f3c0d5dea744ef",
      "old_mode": 33188,
      "old_path": "lib/KorAP/XML/TEI/Tokenizer/Conservative.pm",
      "new_id": "36cc14a16c0d891731a2a1cc41403c9c7d534e57",
      "new_mode": 33188,
      "new_path": "lib/KorAP/XML/TEI/Tokenizer/Conservative.pm"
    },
    {
      "type": "add",
      "old_id": "0000000000000000000000000000000000000000",
      "old_mode": 0,
      "old_path": "/dev/null",
      "new_id": "384bdf9148dce5ee15091019332a1ce5cfa9b0f2",
      "new_mode": 33188,
      "new_path": "t/data/wikipedia_small.txt"
    },
    {
      "type": "modify",
      "old_id": "503eeee0216e6213804c542247856a0fb7f653c5",
      "old_mode": 33188,
      "old_path": "t/tokenization.t",
      "new_id": "42f784014e9ac1ba1d984cc2f05ae12b4b9f875f",
      "new_mode": 33188,
      "new_path": "t/tokenization.t"
    }
  ]
}
