Metadata-Version: 2.1
Name: konoha
Version: 4.2.0
Summary: A tiny sentence/word tokenizer for Japanese text written in Python
Author: himkt
Author-email: himkt@klis.tsukuba.ac.jp
Requires-Python: >=3.6.1,<4.0.0
Classifier: Programming Language :: Python :: 3
Classifier: Programming Language :: Python :: 3.7
Classifier: Programming Language :: Python :: 3.8
Provides-Extra: all
Provides-Extra: all_with_integrations
Provides-Extra: allennlp
Provides-Extra: janome
Provides-Extra: kytea
Provides-Extra: mecab
Provides-Extra: remote
Provides-Extra: sentencepiece
Provides-Extra: sudachi
Requires-Dist: allennlp (>=0.9.0,<0.10.0); extra == "allennlp" or extra == "all_with_integrations"
Requires-Dist: boto3 (>=1.11.0,<2.0.0); extra == "remote" or extra == "all" or extra == "all_with_integrations"
Requires-Dist: janome (>=0.3.10,<0.4.0); extra == "janome" or extra == "all" or extra == "all_with_integrations"
Requires-Dist: kytea (>=0.1.4,<0.2.0); extra == "kytea" or extra == "all" or extra == "all_with_integrations"
Requires-Dist: natto-py (>=0.9.0,<0.10.0); extra == "mecab" or extra == "all" or extra == "all_with_integrations"
Requires-Dist: overrides (>=2.8.0,<3.0.0)
Requires-Dist: sentencepiece (>=0.1.85,<0.2.0); extra == "sentencepiece" or extra == "all" or extra == "all_with_integrations"
Requires-Dist: sudachipy (>=0.4.2,<0.5.0); extra == "sudachi" or extra == "all" or extra == "all_with_integrations"
