-
Notifications
You must be signed in to change notification settings - Fork 10.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
删除stopwords.txt文件内容后重启,自带停用词任然生效 #1253
Comments
NotionalTokenizer会过滤 |
谢谢,但我是引用doc2vec进行文本相似度计算,里面内置了NotionalTokenizer分词器,我想问,如何在python中引用doc2vec时,修改NotionalTokenizer分词器为其他分词器 |
感谢反馈,已经修复,请参考上面的commit。 新增了 |
非常感谢,用了您这边提供的补丁,有以下两个问题: |
|
非常感谢作者的耐心解答,问题已解决。 |
注意事项
请确认下列注意事项:
版本号
当前最新版本号是:1.7.4
我使用的版本是:1.7.4
我的问题
在data/dictionary目录下,删除stopwords.txt.bin文件,并将stopwords.txt文件删除为空,重启后,重新运行程序,发现原停用词表中的停用词任然生效,即删除无效。
复现问题
步骤
触发代码
期望输出
实际输出
其他信息
The text was updated successfully, but these errors were encountered: