chore: fix Travis CI
Joseph Kato
7 years ago
1 | 1 |
go:
|
2 | 2 |
- 1.7
|
3 | 3 |
- 1.8
|
4 | |
matrix:
|
5 | |
include:
|
6 | |
- os: linux
|
7 | |
env: ARCH="i686"
|
8 | |
- os: linux
|
9 | |
env: ARCH="x86_64"
|
10 | |
- os: osx
|
11 | |
env: ARCH="x86_64"
|
12 | |
osx_image: xcode8
|
13 | 4 |
install:
|
14 | 5 |
- make setup
|
15 | 6 |
- make build
|
17 | 17 |
go build ${LDFLAGS} -o bin/prose.exe ./cmd/prose
|
18 | 18 |
|
19 | 19 |
test-tokenize:
|
20 | |
python3 scripts/treebank_words.py
|
21 | 20 |
go test -v ./tokenize
|
22 | 21 |
|
23 | 22 |
test: test-tokenize
|
|
41 | 40 |
./tokenize ./tag
|
42 | 41 |
|
43 | 42 |
setup:
|
|
43 |
go get -u github.com/alecthomas/gometalinter
|
44 | 44 |
go get -u github.com/jteeuwen/go-bindata/...
|
45 | 45 |
go-bindata -ignore=\\.DS_Store -pkg="model" -o internal/model/model.go internal/model/
|
46 | 46 |
gometalinter --install
|
0 | 0 |
import json
|
1 | |
import pathlib
|
|
1 |
import os
|
2 | 2 |
|
3 | 3 |
from nltk.tokenize import TreebankWordTokenizer, sent_tokenize
|
4 | 4 |
|
5 | 5 |
if __name__ == '__main__':
|
6 | 6 |
t = TreebankWordTokenizer()
|
7 | |
with open(pathlib.PurePath('testdata', 'tokenize.json')) as d:
|
|
7 |
with open(os.path.join('testdata', 'tokenize.json')) as d:
|
8 | 8 |
data = json.load(d)
|
9 | 9 |
|
10 | 10 |
words = []
|
|
14 | 14 |
sents.append(s)
|
15 | 15 |
words.append(t.tokenize(s))
|
16 | 16 |
|
17 | |
with open(pathlib.PurePath('testdata', 'treebank_words.json'), 'w') as f:
|
|
17 |
with open(os.path.join('testdata', 'treebank_words.json'), 'w') as f:
|
18 | 18 |
json.dump(words, f, indent=4)
|
19 | 19 |
|
20 | |
with open(pathlib.PurePath('testdata', 'treebank_sents.json'), 'w') as f:
|
|
20 |
with open(os.path.join('testdata', 'treebank_sents.json'), 'w') as f:
|
21 | 21 |
json.dump(sents, f, indent=4)
|