Skip to content

Commit

Permalink
Merge pull request #26 from XiaoMi/hotfix-postprocessing
Browse files Browse the repository at this point in the history
修复针对英文切分时空格丢失问题
  • Loading branch information
nepshi authored Jan 18, 2021
2 parents 7d0c578 + 4fcc5a9 commit f8e4aec
Show file tree
Hide file tree
Showing 3 changed files with 3 additions and 3 deletions.
2 changes: 1 addition & 1 deletion minlp-tokenizer/minlptokenizer/tokenizer.py
Original file line number Diff line number Diff line change
Expand Up @@ -125,7 +125,7 @@ def tag2words(self, text, y_pred_result):
word = ''
if word:
words.append(word)
return regex.split(r'\s+', ' '.join(words))
return words

def set_interfere_factor(self, interfere_factor):
"""
Expand Down
2 changes: 1 addition & 1 deletion minlp-tokenizer/requirements.txt
Original file line number Diff line number Diff line change
@@ -1,3 +1,3 @@
tensorflow>=1.15,<2
tensorflow>=1.14
pyahocorasick
regex
2 changes: 1 addition & 1 deletion minlp-tokenizer/setup.py
Original file line number Diff line number Diff line change
Expand Up @@ -32,7 +32,7 @@

setup(
name='minlp-tokenizer',
version='3.2.3',
version='3.2.4',
description='MiNLP-Tokenizer中文分词工具',
author='Yuankai Guo, Liang Shi, Yupeng Chen',
author_email='[email protected], [email protected]',
Expand Down

0 comments on commit f8e4aec

Please sign in to comment.