2016-12-05 23:42:16 +00:00
|
|
|
from wordfreq import tokenize, word_frequency
|
|
|
|
|
|
|
|
|
|
|
|
def test_apostrophes():
|
2016-12-06 22:37:35 +00:00
|
|
|
# Test that we handle apostrophes in French reasonably.
|
2018-06-01 20:40:51 +00:00
|
|
|
assert tokenize("qu'un", 'fr') == ['qu', 'un']
|
|
|
|
assert tokenize("qu'un", 'fr', include_punctuation=True) == ["qu'", "un"]
|
|
|
|
assert tokenize("langues d'oïl", 'fr') == ['langues', "d", 'oïl']
|
|
|
|
assert tokenize("langues d'oïl", 'fr', include_punctuation=True) == ['langues', "d'", 'oïl']
|
|
|
|
assert tokenize("l'heure", 'fr') == ['l', 'heure']
|
2020-04-28 19:19:56 +00:00
|
|
|
assert tokenize("l'ànima", 'ca') == ['l', 'ànima']
|
2018-06-01 20:40:51 +00:00
|
|
|
assert tokenize("l'heure", 'fr', include_punctuation=True) == ["l'", 'heure']
|
|
|
|
assert tokenize("L'Hôpital", 'fr', include_punctuation=True) == ["l'", 'hôpital']
|
|
|
|
assert tokenize("aujourd'hui", 'fr') == ["aujourd'hui"]
|
|
|
|
assert tokenize("This isn't French", 'en') == ['this', "isn't", 'french']
|
2016-12-05 23:42:16 +00:00
|
|
|
|
2016-12-05 23:48:02 +00:00
|
|
|
|
2016-12-06 22:37:35 +00:00
|
|
|
def test_catastrophes():
|
|
|
|
# More apostrophes, but this time they're in Catalan, and there's other
|
|
|
|
# mid-word punctuation going on too.
|
2018-06-01 20:40:51 +00:00
|
|
|
assert tokenize("M'acabo d'instal·lar.", 'ca') == ['m', 'acabo', 'd', 'instal·lar']
|
|
|
|
assert (
|
|
|
|
tokenize("M'acabo d'instal·lar.", 'ca', include_punctuation=True) ==
|
|
|
|
["m'", 'acabo', "d'", 'instal·lar', '.']
|
|
|
|
)
|
2017-04-27 19:09:59 +00:00
|
|
|
|
|
|
|
|
|
|
|
def test_alternate_codes():
|
|
|
|
# Try over-long language codes for French and Catalan
|
2018-06-01 20:40:51 +00:00
|
|
|
assert tokenize("qu'un", 'fra') == ['qu', 'un']
|
|
|
|
assert tokenize("qu'un", 'fre') == ['qu', 'un']
|
|
|
|
assert tokenize("M'acabo d'instal·lar.", 'cat') == ['m', 'acabo', 'd', 'instal·lar']
|
2017-04-27 19:09:59 +00:00
|
|
|
|