mirror of
https://github.com/rspeer/wordfreq.git
synced 2024-12-23 17:31:41 +00:00
update the changelog
This commit is contained in:
parent
fd0ac9a272
commit
c8229a5378
10
CHANGELOG.md
10
CHANGELOG.md
@ -1,3 +1,13 @@
|
|||||||
|
## Version 2.4 (2020-10-01)
|
||||||
|
|
||||||
|
- The Exquisite Corpus data has been updated to include Google Books Ngrams
|
||||||
|
2019, Reddit data through 2019, Wikipedia data from 2020, and Twitter-sampled
|
||||||
|
data from 2020, and somewhat more reliable language detection.
|
||||||
|
|
||||||
|
- Updated dependencies to require recent versions of `regex` and `jieba`,
|
||||||
|
to get tokenization that's consistent with the word lists. `regex` now
|
||||||
|
requires a version after 2020.04.04.
|
||||||
|
|
||||||
## Version 2.3.2 (2020-04-28)
|
## Version 2.3.2 (2020-04-28)
|
||||||
|
|
||||||
- Relaxing the dependency on regex had an unintended consequence in 2.3.1:
|
- Relaxing the dependency on regex had an unintended consequence in 2.3.1:
|
||||||
|
8
setup.py
8
setup.py
@ -28,14 +28,14 @@ README_contents = open(os.path.join(current_dir, 'README.md'),
|
|||||||
encoding='utf-8').read()
|
encoding='utf-8').read()
|
||||||
doclines = README_contents.split("\n")
|
doclines = README_contents.split("\n")
|
||||||
dependencies = [
|
dependencies = [
|
||||||
'msgpack >= 1.0', 'langcodes >= 2', 'regex'
|
'msgpack >= 1.0', 'langcodes >= 2.1', 'regex >= 2020.04.04'
|
||||||
]
|
]
|
||||||
if sys.version_info < (3, 4):
|
if sys.version_info < (3, 4):
|
||||||
dependencies.append('pathlib')
|
dependencies.append('pathlib')
|
||||||
|
|
||||||
setup(
|
setup(
|
||||||
name="wordfreq",
|
name="wordfreq",
|
||||||
version='2.3.2',
|
version='2.4.0',
|
||||||
maintainer='Robyn Speer',
|
maintainer='Robyn Speer',
|
||||||
maintainer_email='rspeer@luminoso.com',
|
maintainer_email='rspeer@luminoso.com',
|
||||||
url='http://github.com/LuminosoInsight/wordfreq/',
|
url='http://github.com/LuminosoInsight/wordfreq/',
|
||||||
@ -58,7 +58,7 @@ setup(
|
|||||||
# Similarly, jieba is required for Chinese word frequencies.
|
# Similarly, jieba is required for Chinese word frequencies.
|
||||||
extras_require={
|
extras_require={
|
||||||
'mecab': 'mecab-python3',
|
'mecab': 'mecab-python3',
|
||||||
'jieba': 'jieba'
|
'jieba': 'jieba >= 0.42'
|
||||||
},
|
},
|
||||||
tests_require=['pytest', 'mecab-python3', 'jieba'],
|
tests_require=['pytest', 'mecab-python3', 'jieba >= 0.42'],
|
||||||
)
|
)
|
||||||
|
Loading…
Reference in New Issue
Block a user