From 532b953839832fb62e7da393e49bff8e86ae2596 Mon Sep 17 00:00:00 2001 From: Joshua Chin Date: Mon, 20 Jul 2015 10:28:11 -0400 Subject: [PATCH] updated word_frequency docstring for Chinese Former-commit-id: 01b286e8019c44e8f6ec19807c87c01d6d403d43 --- wordfreq/__init__.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/wordfreq/__init__.py b/wordfreq/__init__.py index 9867e89..1b39257 100644 --- a/wordfreq/__init__.py +++ b/wordfreq/__init__.py @@ -266,7 +266,7 @@ def word_frequency(word, lang, wordlist='combined', minimum=0.): individual tokens. It should be noted that the current tokenizer does not support - multi-character Chinese terms. + multi-word Chinese phrases. """ args = (word, lang, wordlist, minimum) try: