Skip to content

Commit e996b46

Browse files
committed
Dependencies: Upgrade Stanza to 1.11.0; Utils: Add Stanza's Sindhi lemmatizer
1 parent 9c42d4f commit e996b46

File tree

17 files changed

+44
-39
lines changed

17 files changed

+44
-39
lines changed

ACKS.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -50,7 +50,7 @@ As *Wordless* stands on the shoulders of giants, I hereby extend my sincere grat
5050
26|[simplemma](https://github.com/adbar/simplemma)|1.1.2|Adrien Barbaresi|[MIT](https://github.com/adbar/simplemma/blob/main/LICENSE)
5151
27|[spaCy](https://spacy.io/)|3.8.7|Matthew Honnibal, Ines Montani, Sofie Van Landeghem,<br>Adriane Boyd, Paul O'Leary McCann|[MIT](https://github.com/explosion/spaCy/blob/master/LICENSE)
5252
28|[spacy-pkuseg](https://github.com/explosion/spacy-pkuseg)|1.0.1|Ruixuan Luo (罗睿轩), Jingjing Xu (许晶晶), Xuancheng Ren (任宣丞),<br>Yi Zhang (张艺), Zhiyuan Zhang (张之远), Bingzhen Wei (位冰镇),<br>Xu Sun (孙栩), Matthew Honnibal|[MIT](https://github.com/explosion/spacy-pkuseg/blob/master/LICENSE)
53-
29|[Stanza](https://github.com/stanfordnlp/stanza)|1.10.1|Peng Qi (齐鹏), Yuhao Zhang (张宇浩), Yuhui Zhang (张钰晖),<br>Jason Bolton, Tim Dozat, John Bauer|[Apache-2.0](https://github.com/stanfordnlp/stanza/blob/main/LICENSE)
53+
29|[Stanza](https://github.com/stanfordnlp/stanza)|1.11.0|Peng Qi (齐鹏), Yuhao Zhang (张宇浩), Yuhui Zhang (张钰晖),<br>Jason Bolton, Tim Dozat, John Bauer|[Apache-2.0](https://github.com/stanfordnlp/stanza/blob/main/LICENSE)
5454
30|[SudachiPy](https://github.com/WorksApplications/sudachi.rs/tree/develop/python)|0.6.10|WAP Tokushima Laboratory of AI and NLP|[Apache-2.0](https://github.com/WorksApplications/sudachi.rs/blob/develop/LICENSE)
5555
31|[Underthesea](https://undertheseanlp.com/)|6.8.4|Vu Anh|[GPL-3.0-or-later](https://github.com/undertheseanlp/underthesea/blob/main/LICENSE)
5656
32|[VADER](https://github.com/cjhutto/vaderSentiment)|3.3.2|C.J. Hutto|[MIT](https://github.com/cjhutto/vaderSentiment/blob/master/LICENSE.txt)

CHANGELOG.md

Lines changed: 4 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -23,6 +23,7 @@
2323
- File Area: Add support for .srt files
2424
- Menu: Add Edit - Sample
2525
- Settings: Add Settings - Tables - Miscellaneous Settings
26+
- Utils: Add Stanza's Sindhi lemmatizer
2627
- Utils: Add Wordless's character tokenizer
2728
- Work Area: Add Keyword Extractor - Generation Settings - N-gram size
2829
- Work Area: Add Profiler - Export all tables
@@ -41,6 +42,9 @@
4142
### 🔧 Bugfixes
4243
- Work Area: Fix Profiler - Syntactic Complexity
4344

45+
### ⏫ Dependency Changes
46+
- Dependencies: Upgrade Stanza to 1.11.0
47+
4448
## [3.6.0](https://github.com/BLKSerene/Wordless/releases/tag/3.6.0) - 07/27/2025
4549
### 🎉 New Features
4650
- File Area: Add Number of corpora / tokens / types (selected)

doc/doc.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -821,7 +821,7 @@ Sanskrit |✔|✔|✔|✔|✔|✔|✔|✖️
821821
Scottish Gaelic |✔|✔|✖️|✔|✔|✔|✔|✖️
822822
Serbian (Cyrillic script) |✔|✔|✔|✔|✔|✔|✔|✖️
823823
Serbian (Latin script) |✔|✔|✔|✔|✔|✔|✔|✖️
824-
Sindhi |✔|✔|✖️|✔|✖️|✔|✔|✖️
824+
Sindhi |✔|✔|✖️|✔||✔|✔|✖️
825825
Sinhala |⭕️|✔|✖️|✖️|✖️|✔|✖️|✖️
826826
Slovak |✔|✔|✔|✔|✔|✔|✔|✖️
827827
Slovenian |✔|✔|✔|✔|✔|✔|✔|✖️

doc/trs/zho_cn/ACKS.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -50,7 +50,7 @@
5050
26|[simplemma](https://github.com/adbar/simplemma)|1.1.2|Adrien Barbaresi|[MIT](https://github.com/adbar/simplemma/blob/main/LICENSE)
5151
27|[spaCy](https://spacy.io/)|3.8.7|Matthew Honnibal, Ines Montani, Sofie Van Landeghem,<br>Adriane Boyd, Paul O'Leary McCann|[MIT](https://github.com/explosion/spaCy/blob/master/LICENSE)
5252
28|[spacy-pkuseg](https://github.com/explosion/spacy-pkuseg)|1.0.1|罗睿轩, 许晶晶, 任宣丞, 张艺, 张之远, 位冰镇, 孙栩<br>Matthew Honnibal|[MIT](https://github.com/explosion/spacy-pkuseg/blob/master/LICENSE)
53-
29|[Stanza](https://github.com/stanfordnlp/stanza)|1.10.1|齐鹏, 张宇浩, 张钰晖,<br>Jason Bolton, Tim Dozat, John Bauer|[Apache-2.0](https://github.com/stanfordnlp/stanza/blob/main/LICENSE)
53+
29|[Stanza](https://github.com/stanfordnlp/stanza)|1.11.0|齐鹏, 张宇浩, 张钰晖,<br>Jason Bolton, Tim Dozat, John Bauer|[Apache-2.0](https://github.com/stanfordnlp/stanza/blob/main/LICENSE)
5454
30|[SudachiPy](https://github.com/WorksApplications/sudachi.rs/tree/develop/python)|0.6.10|WAP Tokushima Laboratory of AI and NLP|[Apache-2.0](https://github.com/WorksApplications/sudachi.rs/blob/develop/LICENSE)
5555
31|[Underthesea](https://undertheseanlp.com/)|6.8.4|Vu Anh|[GPL-3.0-or-later](https://github.com/undertheseanlp/underthesea/blob/main/LICENSE)
5656
32|[VADER](https://github.com/cjhutto/vaderSentiment)|3.3.2|C.J. Hutto|[MIT](https://github.com/cjhutto/vaderSentiment/blob/master/LICENSE.txt)

doc/trs/zho_tw/ACKS.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -50,7 +50,7 @@
5050
26|[simplemma](https://github.com/adbar/simplemma)|1.1.2|Adrien Barbaresi|[MIT](https://github.com/adbar/simplemma/blob/main/LICENSE)
5151
27|[spaCy](https://spacy.io/)|3.8.7|Matthew Honnibal, Ines Montani, Sofie Van Landeghem,<br>Adriane Boyd, Paul O'Leary McCann|[MIT](https://github.com/explosion/spaCy/blob/master/LICENSE)
5252
28|[spacy-pkuseg](https://github.com/explosion/spacy-pkuseg)|1.0.1|羅睿軒, 許晶晶, 任宣丞, 張藝, 張之遠, 位冰鎮, 孫栩<br>Matthew Honnibal|[MIT](https://github.com/explosion/spacy-pkuseg/blob/master/LICENSE)
53-
29|[Stanza](https://github.com/stanfordnlp/stanza)|1.10.1|齊鵬, 張宇浩, 張鈺暉,<br>Jason Bolton, Tim Dozat, John Bauer|[Apache-2.0](https://github.com/stanfordnlp/stanza/blob/main/LICENSE)
53+
29|[Stanza](https://github.com/stanfordnlp/stanza)|1.11.0|齊鵬, 張宇浩, 張鈺暉,<br>Jason Bolton, Tim Dozat, John Bauer|[Apache-2.0](https://github.com/stanfordnlp/stanza/blob/main/LICENSE)
5454
30|[SudachiPy](https://github.com/WorksApplications/sudachi.rs/tree/develop/python)|0.6.10|WAP Tokushima Laboratory of AI and NLP|[Apache-2.0](https://github.com/WorksApplications/sudachi.rs/blob/develop/LICENSE)
5555
31|[Underthesea](https://undertheseanlp.com/)|6.8.4|Vu Anh|[GPL-3.0-or-later](https://github.com/undertheseanlp/underthesea/blob/main/LICENSE)
5656
32|[VADER](https://github.com/cjhutto/vaderSentiment)|3.3.2|C.J. Hutto|[MIT](https://github.com/cjhutto/vaderSentiment/blob/master/LICENSE.txt)

requirements/requirements_tests.txt

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -28,7 +28,7 @@ pyphen == 0.17.2
2828
pythainlp == 5.1.2
2929
sacremoses == 0.1.1
3030
simplemma == 1.1.2
31-
stanza == 1.10.1
31+
stanza == 1.11.0
3232
underthesea == 6.8.4
3333
vaderSentiment == 3.3.2
3434

requirements/requirements_tests_macos.txt

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -28,7 +28,7 @@ pyphen == 0.17.2
2828
pythainlp == 5.1.2
2929
sacremoses == 0.1.1
3030
simplemma == 1.1.2
31-
stanza == 1.10.1
31+
stanza == 1.11.0
3232
underthesea == 6.8.4
3333
vaderSentiment == 3.3.2
3434

tests/tests_nlp/test_sentence_tokenization.py

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -156,6 +156,8 @@ def wl_test_sentence_tokenize_models(lang, sentence_tokenizer, results):
156156
assert len(sentences) == 3
157157
case _:
158158
assert len(sentences) == 2
159+
case 'grc' | 'snd':
160+
assert len(sentences) == 4
159161
case 'swe':
160162
if sentence_tokenizer == 'spacy_dependency_parser_swe':
161163
assert len(sentences) == 3

tests/tests_nlp/tests_stanza/test_stanza_deu.py

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -21,12 +21,12 @@
2121
def test_stanza_deu():
2222
test_stanza.wl_test_stanza(
2323
lang = 'deu_de',
24-
results_sentence_tokenize = ['Das Deutsche ist eine plurizentrische Sprache, enthält also mehrere Standardvarietäten in verschiedenen Regionen.', 'Ihr Sprachgebiet umfasst Deutschland, Österreich, die Deutschschweiz, Liechtenstein, Luxemburg, Ostbelgien, Südtirol, das Elsass und Lothringen sowie Nordschleswig.', 'Außerdem ist Deutsch eine Minderheitensprache in einigen europäischen und außereuropäischen Ländern, z. B. in Rumänien und Südafrika sowie Nationalsprache im afrikanischen Namibia.', 'Deutsch ist die meistgesprochene Muttersprache in der Europäischen Union (EU).', '[26]'],
25-
results_word_tokenize = ['Das', 'Deutsche', 'ist', 'eine', 'plurizentrische', 'Sprache', ',', 'enthält', 'also', 'mehrere', 'Standardvarietäten', 'in', 'verschiedenen', 'Regionen', '.'],
26-
results_pos_tag = [('Das', 'ART'), ('Deutsche', 'NN'), ('ist', 'VAFIN'), ('eine', 'ART'), ('plurizentrische', 'ADJA'), ('Sprache', 'NN'), (',', '$,'), ('enthält', 'VVFIN'), ('also', 'ADV'), ('mehrere', 'PIAT'), ('Standardvarietäten', 'NN'), ('in', 'APPR'), ('verschiedenen', 'ADJA'), ('Regionen', 'NN'), ('.', '$.')],
27-
results_pos_tag_universal = [('Das', 'DET'), ('Deutsche', 'PROPN'), ('ist', 'AUX'), ('eine', 'DET'), ('plurizentrische', 'ADJ'), ('Sprache', 'NOUN'), (',', 'PUNCT'), ('enthält', 'VERB'), ('also', 'ADV'), ('mehrere', 'DET'), ('Standardvarietäten', 'NOUN'), ('in', 'ADP'), ('verschiedenen', 'ADJ'), ('Regionen', 'NOUN'), ('.', 'PUNCT')],
28-
results_lemmatize = ['der', 'deutsch', 'sein', 'ein', 'plurizentrisch', 'Sprache', ',', 'enthalten', 'also', 'mehr', 'Standardvarietät', 'in', 'verschieden', 'Region', '.'],
29-
results_dependency_parse = [('Das', 'Deutsche', 'det', 1), ('Deutsche', 'Sprache', 'nsubj', 4), ('ist', 'Sprache', 'cop', 3), ('eine', 'Sprache', 'det', 2), ('plurizentrische', 'Sprache', 'amod', 1), ('Sprache', 'Sprache', 'root', 0), (',', 'enthält', 'punct', 1), ('enthält', 'Sprache', 'conj', -2), ('also', 'enthält', 'advmod', -1), ('mehrere', 'Standardvarietäten', 'det', 1), ('Standardvarietäten', 'enthält', 'obj', -3), ('in', 'Regionen', 'case', 2), ('verschiedenen', 'Regionen', 'amod', 1), ('Regionen', 'Standardvarietäten', 'nmod', -3), ('.', 'Sprache', 'punct', -9)],
24+
results_sentence_tokenize = ['Die deutsche Sprache oder Deutsch [dɔɪ̯tʃ][24] ist eine westgermanische Sprache, die weltweit etwa 90 bis 105 Millionen Menschen als Muttersprache und weiteren rund 80 Millionen als Zweit- oder Fremdsprache dient.', 'Das Deutsche ist eine plurizentrische Sprache, enthält also mehrere Standardvarietäten in verschiedenen Regionen.'],
25+
results_word_tokenize = ['Die', 'deutsche', 'Sprache', 'oder', 'Deutsch', '[', 'dɔɪ̯tʃ][24', ']', 'ist', 'eine', 'westgermanische', 'Sprache', ',', 'die', 'weltweit', 'etwa', '90', 'bis', '105', 'Millionen', 'Menschen', 'als', 'Muttersprache', 'und', 'weiteren', 'rund', '80', 'Millionen', 'als', 'Zweit', '-', 'oder', 'Fremdsprache', 'dient', '.'],
26+
results_pos_tag = [('Die', 'ART'), ('deutsche', 'ADJA'), ('Sprache', 'NN'), ('oder', 'KON'), ('Deutsch', 'NN'), ('[', '$('), ('dɔɪ̯tʃ][24', 'NE'), (']', '$('), ('ist', 'VAFIN'), ('eine', 'ART'), ('westgermanische', 'ADJA'), ('Sprache', 'NN'), (',', '$,'), ('die', 'PRELS'), ('weltweit', 'ADJD'), ('etwa', 'ADV'), ('90', 'CARD'), ('bis', 'KON'), ('105', 'CARD'), ('Millionen', 'NN'), ('Menschen', 'NN'), ('als', 'KOKOM'), ('Muttersprache', 'NN'), ('und', 'KON'), ('weiteren', 'ADJA'), ('rund', 'ADV'), ('80', 'CARD'), ('Millionen', 'NN'), ('als', 'KOKOM'), ('Zweit', 'TRUNC'), ('-', '$('), ('oder', 'KON'), ('Fremdsprache', 'NN'), ('dient', 'VVFIN'), ('.', '$.')],
27+
results_pos_tag_universal = [('Die', 'DET'), ('deutsche', 'ADJ'), ('Sprache', 'NOUN'), ('oder', 'CCONJ'), ('Deutsch', 'NOUN'), ('[', 'PUNCT'), ('dɔɪ̯tʃ][24', 'PROPN'), (']', 'PUNCT'), ('ist', 'AUX'), ('eine', 'DET'), ('westgermanische', 'ADJ'), ('Sprache', 'NOUN'), (',', 'PUNCT'), ('die', 'PRON'), ('weltweit', 'ADJ'), ('etwa', 'ADV'), ('90', 'NUM'), ('bis', 'ADP'), ('105', 'NUM'), ('Millionen', 'NOUN'), ('Menschen', 'NOUN'), ('als', 'ADP'), ('Muttersprache', 'NOUN'), ('und', 'CCONJ'), ('weiteren', 'ADJ'), ('rund', 'ADV'), ('80', 'NUM'), ('Millionen', 'NOUN'), ('als', 'ADP'), ('Zweit', 'NOUN'), ('-', 'PUNCT'), ('oder', 'CCONJ'), ('Fremdsprache', 'NOUN'), ('dient', 'VERB'), ('.', 'PUNCT')],
28+
results_lemmatize = ['der', 'deutsch', 'Sprache', 'oder', 'deutsch', '[', 'denfellelhuldellulfelliehueruu', ']', 'sein', 'ein', 'westgermanisch', 'Sprache', ',', 'der', 'weltweit', 'etwa', '90', 'bis', '105', 'Million', 'Mensch', 'als', 'Muttersprache', 'und', 'weit', 'rund', '80', 'Million', 'als', 'Zweit', '-', 'oder', 'Fremdsprache', 'dienen', '.'],
29+
results_dependency_parse = [('Die', 'Sprache', 'det', 2), ('deutsche', 'Sprache', 'amod', 1), ('Sprache', 'Sprache', 'nsubj', 9), ('oder', 'Deutsch', 'cc', 1), ('Deutsch', 'Sprache', 'conj', -2), ('[', 'dɔɪ̯tʃ][24', 'punct', 1), ('dɔɪ̯tʃ][24', 'Sprache', 'appos', -4), (']', 'dɔɪ̯tʃ][24', 'punct', -1), ('ist', 'Sprache', 'cop', 3), ('eine', 'Sprache', 'det', 2), ('westgermanische', 'Sprache', 'amod', 1), ('Sprache', 'Sprache', 'root', 0), (',', 'dient', 'punct', 21), ('die', 'dient', 'nsubj', 20), ('weltweit', 'dient', 'advmod', 19), ('etwa', '90', 'advmod', 1), ('90', 'Millionen', 'nummod', 3), ('bis', '105', 'case', 1), ('105', 'Millionen', 'nmod', 1), ('Millionen', 'Menschen', 'nmod', 1), ('Menschen', 'dient', 'obj', 13), ('als', 'Muttersprache', 'case', 1), ('Muttersprache', 'dient', 'obl', 11), ('und', 'Millionen', 'cc', 4), ('weiteren', 'Millionen', 'amod', 3), ('rund', '80', 'advmod', 1), ('80', 'Millionen', 'nummod', 1), ('Millionen', 'Zweit', 'nmod', 2), ('als', 'Zweit', 'case', 1), ('Zweit', 'dient', 'obl', 4), ('-', 'Fremdsprache', 'punct', 2), ('oder', 'Fremdsprache', 'cc', 1), ('Fremdsprache', 'Zweit', 'conj', -3), ('dient', 'Sprache', 'acl', -22), ('.', 'Sprache', 'punct', -23)],
3030
results_sentiment_analayze = [0]
3131
)
3232

tests/tests_nlp/tests_stanza/test_stanza_fra.py

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -19,16 +19,16 @@
1919
from tests.tests_nlp.tests_stanza import test_stanza
2020

2121
def test_stanza_fra():
22-
results_pos_tag = [('Le', 'DET'), ('français', 'NOUN'), ('est', 'AUX'), ('une', 'DET'), ('langue', 'NOUN'), ('indo-européenne', 'ADJ'), ('de', 'ADP'), ('la', 'DET'), ('famille', 'NOUN'), ('de', 'ADP'), ('les', 'DET'), ('langues', 'NOUN'), ('romanes', 'ADJ'), ('dont', 'PRON'), ('les', 'DET'), ('locuteurs', 'NOUN'), ('sont', 'AUX'), ('appelés', 'VERB'), ('francophones', 'NOUN'), ('.', 'PUNCT')]
22+
results_pos_tag = [('Le', 'DET'), ('français', 'NOUN'), ('est', 'AUX'), ('une', 'DET'), ('langue', 'NOUN'), ('indo-européenne', 'ADJ'), ('de', 'ADP'), ('la', 'DET'), ('famille', 'NOUN'), ('de', 'ADP'), ('les', 'DET'), ('langues', 'NOUN'), ('romanes', 'ADJ'), ('dont', 'PRON'), ('les', 'DET'), ('locuteurs', 'NOUN'), ('sont', 'AUX'), ('appelés', 'VERB'), ('«', 'PUNCT'), ('francophones', 'NOUN'), ('»', 'PUNCT'), ('.', 'PUNCT')]
2323

2424
test_stanza.wl_test_stanza(
2525
lang = 'fra',
26-
results_sentence_tokenize = ['Le français est une langue indo-européenne de la famille des langues romanes dont les locuteurs sont appelés francophones.', 'Elle est parfois surnommée la langue de Molière.'],
27-
results_word_tokenize = ['Le', 'français', 'est', 'une', 'langue', 'indo-européenne', 'de', 'la', 'famille', 'des', 'langues', 'romanes', 'dont', 'les', 'locuteurs', 'sont', 'appelés', 'francophones', '.'],
26+
results_sentence_tokenize = ['Le français est une langue indo-européenne de la famille des langues romanes dont les locuteurs sont appelés « francophones ».', "Il est la cinquième langue parlée au monde après l'anglais, le mandarin, le hindi et l'espagnol."],
27+
results_word_tokenize = ['Le', 'français', 'est', 'une', 'langue', 'indo-européenne', 'de', 'la', 'famille', 'des', 'langues', 'romanes', 'dont', 'les', 'locuteurs', 'sont', 'appelés', '«', 'francophones', ', '.'],
2828
results_pos_tag = results_pos_tag,
2929
results_pos_tag_universal = results_pos_tag,
30-
results_lemmatize = ['le', 'français', 'être', 'un', 'langue', 'indo-européen', 'de', 'le', 'famille', 'de', 'le', 'langue', 'roman', 'dont', 'le', 'locuteur', 'être', 'appeler', 'francophone', '.'],
31-
results_dependency_parse = [('Le', 'français', 'det', 1), ('français', 'langue', 'nsubj', 3), ('est', 'langue', 'cop', 2), ('une', 'langue', 'det', 1), ('langue', 'langue', 'root', 0), ('indo-européenne', 'langue', 'amod', -1), ('de', 'famille', 'case', 2), ('la', 'famille', 'det', 1), ('famille', 'langue', 'nmod', -4), ('de', 'langues', 'case', 2), ('les', 'langues', 'det', 1), ('langues', 'famille', 'nmod', -3), ('romanes', 'langues', 'amod', -1), ('dont', 'locuteurs', 'nmod', 2), ('les', 'locuteurs', 'det', 1), ('locuteurs', 'appelés', 'nsubj:pass', 2), ('sont', 'appelés', 'aux:pass', 1), ('appelés', 'langues', 'acl:relcl', -6), ('francophones', 'appelés', 'xcomp', -1), ('.', 'langue', 'punct', -15)]
30+
results_lemmatize = ['le', 'français', 'être', 'un', 'langue', 'indo-européen', 'de', 'le', 'famille', 'de', 'le', 'langue', 'roman', 'dont', 'le', 'locuteur', 'être', 'appeler', '«', 'francophone', ', '.'],
31+
results_dependency_parse = [('Le', 'français', 'det', 1), ('français', 'langue', 'nsubj', 3), ('est', 'langue', 'cop', 2), ('une', 'langue', 'det', 1), ('langue', 'langue', 'root', 0), ('indo-européenne', 'langue', 'amod', -1), ('de', 'famille', 'case', 2), ('la', 'famille', 'det', 1), ('famille', 'langue', 'nmod', -4), ('de', 'langues', 'case', 2), ('les', 'langues', 'det', 1), ('langues', 'famille', 'nmod', -3), ('romanes', 'langues', 'amod', -1), ('dont', 'locuteurs', 'nmod', 2), ('les', 'locuteurs', 'det', 1), ('locuteurs', 'appelés', 'nsubj:pass', 2), ('sont', 'appelés', 'aux:pass', 1), ('appelés', 'langue', 'acl:relcl', -13), ('«', 'francophones', 'punct', 1), ('francophones', 'appelés', 'xcomp', -2), ('»', 'francophones', 'punct', -1), ('.', 'langue', 'punct', -17)]
3232
)
3333

3434
if __name__ == '__main__':

0 commit comments

Comments
 (0)