A collection of pretrained syntactic models is now available for download at
http://download.tensorflow.org/models/parsey_universal/<language>.zip
After downloading and unzipping a model, you can run it similarly to Parsey McParseface with:
MODEL_DIRECTORY=/where/you/unzipped/the/model/files
cat sentences.txt | syntaxnet/models/parsey_universal/parse.sh \
$MODEL_DIRECTORY > output.conll
These models are trained on Universal Dependencies datasets v1.3. The following table shows their accuracy on Universal Dependencies test sets for different types of annotations.
Language | No. tokens | POS | fPOS | Morph | UAS | LAS -------- | :--: | :--: | :--: | :--: | :--: | :--: | :--: Ancient_Greek-PROIEL | 18502 | 97.14% | 96.97% | 89.77% | 78.74% | 73.15% Ancient_Greek | 25251 | 93.22% | 84.22% | 90.01% | 68.98% | 62.07% Arabic | 28268 | 95.65% | 91.03% | 91.23% | 81.49% | 75.82% Basque | 24374 | 94.88% | - | 87.82% | 78.00% | 73.36% Bulgarian | 15734 | 97.71% | 95.14% | 94.61% | 89.35% | 85.01% Catalan | 59503 | 98.06% | 98.06% | 97.56% | 90.47% | 87.64% Chinese | 12012 | 91.32% | 90.89% | 98.76% | 76.71% | 71.24% Croatian | 4125 | 94.67% | - | 86.69% | 80.65% | 74.06% Czech-CAC | 10862 | 98.11% | 92.43% | 91.43% | 87.28% | 83.44% Czech-CLTT | 4105 | 95.79% | 87.36% | 86.33% | 77.34% | 73.40% Czech | 173920 | 98.12% | 93.76% | 93.13% | 89.47% | 85.93% Danish | 5884 | 95.28% | - | 95.24% | 79.84% | 76.34% Dutch-LassySmall | 4562 | 95.62% | - | 95.44% | 81.63% | 78.08% Dutch | 5843 | 89.89% | 86.03% | 89.12% | 77.70% | 71.21% English-LinES | 8481 | 95.34% | 93.11% | - | 81.50% | 77.37% English | 25096 | 90.48% | 89.71% | 91.30% | 84.79% | 80.38% Estonian | 23670 | 95.92% | 96.76% | 92.73% | 83.10% | 78.83% Finnish-FTB | 16286 | 93.50% | 91.15% | 92.44% | 84.97% | 80.48% Finnish | 9140 | 94.78% | 95.84% | 92.42% | 83.65% | 79.60% French | 7018 | 96.27% | - | 96.05% | 84.68% | 81.05% Galician | 29746 | 96.81% | 96.14% | - | 84.48% | 81.35% German | 16268 | 91.79% | - | - | 79.73% | 74.07% Gothic | 5158 | 95.58% | 96.03% | 87.32% | 79.33% | 71.69% Greek | 5668 | 97.48% | 97.48% | 92.70% | 83.68% | 79.99% Hebrew | 12125 | 95.04% | 95.04% | 92.05% | 84.61% | 78.71% Hindi | 35430 | 96.45% | 95.77% | 90.98% | 93.04% | 89.32% Hungarian | 4235 | 94.00% | - | 75.68% | 78.75% | 71.83% Indonesian | 11780 | 92.62% | - | - | 80.03% | 72.99% Irish | 3821 | 91.34% | 89.95% | 77.07% | 74.51% | 66.29% Italian | 10952 | 97.31% | 97.18% | 97.27% | 89.81% | 87.13% Kazakh | 587 | 75.47% | 75.13% | - | 58.09% | 43.95% Latin-ITTB | 6548 | 97.98% | 92.68% | 93.52% | 84.22% | 81.17% Latin-PROIEL | 14906 | 96.50% | 96.08% | 88.39% | 77.60% | 70.98% Latin | 4832 | 88.04% | 74.07% | 76.03% | 56.00% | 45.80% Latvian | 3985 | 80.95% | 66.60% | 73.60% | 58.92% | 51.47% Norwegian | 30034 | 97.44% | - | 95.58% | 88.61% | 86.22% Old_Church_Slavonic | 5079 | 96.50% | 96.28% | 89.43% | 84.86% | 78.85% Persian | 16022 | 96.20% | 95.72% | 95.90% | 84.42% | 80.28% Polish | 7185 | 95.05% | 85.83% | 86.12% | 88.30% | 82.71% Portuguese-BR | 29438 | 97.07% | 97.07% | 99.91% | 87.91% | 85.44% Portuguese | 6262 | 96.81% | 90.67% | 94.22% | 85.12% | 81.28% Romanian | 18375 | 95.26% | 91.66% | 91.98% | 83.64% | 75.36% Russian-SynTagRus | 107737 | 98.27% | - | 94.91% | 91.68% | 87.44% Russian | 9573 | 95.27% | 95.02% | 87.75% | 81.75% | 77.71% Slovenian-SST | 2951 | 90.00% | 84.48% | 84.38% | 65.06% | 56.96% Slovenian | 14063 | 96.22% | 90.46% | 90.35% | 87.71% | 84.60% Spanish-AnCora | 53594 | 98.28% | 98.28% | 97.82% | 89.26% | 86.50% Spanish | 7953 | 95.27% | - | 95.74% | 85.06% | 81.53% Swedish-LinES | 8228 | 96.00% | 93.77% | - | 81.38% | 77.21% Swedish | 20377 | 96.27% | 94.13% | 94.14% | 83.84% | 80.28% Tamil | 1989 | 79.29% | 71.79% | 75.97% | 64.45% | 55.35% Turkish | 8616 | 93.63% | 92.62% | 86.79% | 82.00% | 71.37% Average | - | 94.27% | 92.93% | 90.38% | 81.12% | 75.85%
These results are obtained using gold text segmentation. Accuracies are measured
over all tokens, including punctuation. POS
, fPOS
are coarse and fine
part-of-speech tagging accuracies. Morph
is full-token accuracy of predicted
morphological attributes. UAS
and LAS
are unlabeled and labeled attachment
scores.
Many of these models also support text segmentation, with:
MODEL_DIRECTORY=/where/you/unzipped/the/model/files
cat sentences.txt | syntaxnet/models/parsey_universal/tokenize.sh \
$MODEL_DIRECTORY > output.conll
Text segmentation is currently available for:
Bulgarian
, Czech
, German
, Greek
, English
, Spanish
, Estonian
,
Basque
, Persian
, Finnish
, Finnish-FTB
, French
, Galician
,
Ancient_Greek
, Ancient_Greek-PROIEL
, Hebrew
, Hindi
, Croatian
,
Hungarian
, Indonesian
, Italian
, Latin
, Latin-PROIEL
, Dutch
,
Norwegian
, Polish
, Portuguese
, Slovenian
, Swedish
, Tamil
.
For Chinese
(traditional) we use a larger text segmentation
model, which can be run with:
MODEL_DIRECTORY=/where/you/unzipped/the/model/files
cat sentences.txt | syntaxnet/models/parsey_universal/tokenize_zh.sh \
$MODEL_DIRECTORY > output.conll