mirror of
https://github.com/rspeer/wordfreq.git
synced 2024-12-23 17:31:41 +00:00
parent
af29fc4f88
commit
0b5d2cdca9
@ -9,7 +9,7 @@ TOKEN_RE = regex.compile(r"""
|
|||||||
# When we see characters that are Han ideographs (\p{IsIdeo}) or hiragana
|
# When we see characters that are Han ideographs (\p{IsIdeo}) or hiragana
|
||||||
# \p{Script=Hiragana}, we allow a sequence of those characters to be glued
|
# \p{Script=Hiragana}, we allow a sequence of those characters to be glued
|
||||||
# together as a single token. Without this case, the standard rule (case 2)
|
# together as a single token. Without this case, the standard rule (case 2)
|
||||||
# would make each characte a separate token. This would be the correct
|
# would make each character a separate token. This would be the correct
|
||||||
# behavior for word-wrapping, but a messy failure mode for NLP
|
# behavior for word-wrapping, but a messy failure mode for NLP
|
||||||
# tokenization.
|
# tokenization.
|
||||||
#
|
#
|
||||||
|
Loading…
Reference in New Issue
Block a user