Commit Graph

33 Commits

Author SHA1 Message Date
Robyn Speer
969a024dea actually use the results of language-detection on Reddit
Former-commit-id: 75a4a92110
2016-03-24 16:27:24 -04:00
Robyn Speer
738243e244 build a bigger wordlist that we can optionally use
Former-commit-id: df8caaff7d
2016-01-12 14:05:57 -05:00
Robyn Speer
7d1719cfb4 builder: Use an optional cutoff when merging counts
This allows the Reddit-merging step to not use such a ludicrous amount
of memory.


Former-commit-id: 973caca253
2015-12-15 14:44:34 -05:00
Robyn Speer
f5e09f3f3d gzip the intermediate step of Reddit word counting
Former-commit-id: 9a5d9d66bb
2015-12-09 13:30:08 -05:00
Robyn Speer
6d2709f064 add word frequencies from the Reddit 2007-2015 corpus
Former-commit-id: b2d7546d2d
2015-11-30 16:38:11 -05:00
Robyn Speer
7494ae27a7 fix missing word in rules.ninja comment
Former-commit-id: 9b1c4d66cd
2015-09-24 17:56:06 -04:00
Robyn Speer
4aef1dc338 don't do language-specific tokenization in freqs_to_cBpack
Tokenizing in the 'merge' step is sufficient.


Former-commit-id: bc8ebd23e9
2015-09-08 14:46:04 -04:00
Robyn Speer
3fa14ded28 language-specific frequency reading; fix 't in English
Former-commit-id: 9071defb33
2015-09-08 12:49:21 -04:00
Robyn Speer
a4554fb87c tokenize Chinese using jieba and our own frequencies
Former-commit-id: 2327f2e4d6
2015-09-05 03:16:56 -04:00
Robyn Speer
7d1c2e72e4 WIP: Traditional Chinese
Former-commit-id: 7906a671ea
2015-09-04 18:52:37 -04:00
Robyn Speer
d0ada70355 add more SUBTLEX and fix its build rules
Former-commit-id: 34474939f2
2015-09-04 12:37:35 -04:00
Robyn Speer
76c751652e refer to merge_freqs command correctly
Former-commit-id: 40d82541ba
2015-09-03 23:25:46 -04:00
Robyn Speer
f66d03b1b9 Add SUBTLEX as a source of English and Chinese data
Meanwhile, fix up the dependency graph thingy. It's actually kind of
legible now.


Former-commit-id: 2d58ba94f2
2015-09-03 18:13:13 -04:00
Joshua Chin
f9742c94ca reordered command line args
Former-commit-id: 6453d864c4
2015-07-22 10:04:14 -04:00
Joshua Chin
34504eed80 fixed rules.ninja
Former-commit-id: c5f82ecac1
2015-07-20 17:20:29 -04:00
Joshua Chin
c2f3928433 fix arabic tokens
Former-commit-id: 11a1c51321
2015-07-17 15:52:12 -04:00
Joshua Chin
a340a15870 removed mkdir -p for many cases
Former-commit-id: 98a7a8093b
2015-07-17 14:45:22 -04:00
Robyn Speer
deed2f767c remove wiki2tokens and tokenize_wikipedia
These components are no longer necessary. Wikipedia output can and
should be tokenized with the standard tokenizer, instead of the
almost-equivalent one in the Nim code.
2015-06-30 15:28:01 -04:00
Robyn Speer
f17a04aa84 fix comment and whitespace involving tokenize_twitter 2015-06-30 15:18:37 -04:00
Robyn Speer
91d6edd55b Switch to a centibel scale, add a header to the data 2015-06-22 17:38:13 -04:00
Joshua Chin
6f0a082007 removed intermediate twitter file rules 2015-06-16 17:28:09 -04:00
Robyn Speer
a5954d14df give mecab a larger buffer 2015-05-26 19:34:46 -04:00
Robyn Speer
4f738ad78c correct a Leeds bug; add some comments to rules.ninja 2015-05-26 18:08:04 -04:00
Robyn Speer
4513fed60c add Google Books data for English 2015-05-11 18:44:28 -04:00
Robyn Speer
aa55e32450 Makefile should only be needed for bootstrapping Ninja 2015-05-08 12:39:31 -04:00
Robyn Speer
a5f6113824 a reasonably complete build process 2015-05-07 19:38:33 -04:00
Robyn Speer
04bde8d617 WIP on more build steps 2015-05-07 16:49:53 -04:00
Robyn Speer
7c09fec692 add rules to count wikipedia tokens 2015-05-05 15:21:24 -04:00
Robyn Speer
c55e44e486 fix the 'count' ninja rule 2015-05-05 14:06:13 -04:00
Robyn Speer
59409266ca add and adjust some build steps
- more build steps for Wikipedia
- rename 'tokenize_twitter' to 'pretokenize_twitter' to indicate that
  the results are preliminary
2015-05-05 13:59:21 -04:00
Robyn Speer
efcf436112 WIP on new build system 2015-04-30 16:24:28 -04:00
Robyn Speer
76ea7f1bd5 define some ninja rules 2015-04-29 17:13:58 -04:00
Robyn Speer
524f7c760b WIP on Ninja build automation 2015-04-29 15:59:06 -04:00