Text to sentence splitter using heuristic algorithm by Philipp Koehn and Josh Schroeder.
This module allows splitting of text paragraphs into sentences. It is based on scripts developed by Philipp
Koehn and Josh Schroeder for processing the `Europarl corpus <http://www.statmt.org/europarl/>`_.
The module is a port of `Lingua::Sentence Perl module <http://search.cpan.org/perldoc?Lingua::Sentence>`_ with
some extra additions (improved non-breaking prefix lists for some languages and added support for Danish,
Finnish, Lithuanian, Norwegian (Bokmål), Romanian, and Turkish).
Raw data
{
"_id": null,
"home_page": "https://github.com/berkmancenter/mediacloud-sentence-splitter",
"name": "sentence-splitter",
"maintainer": "",
"docs_url": null,
"requires_python": ">=3.5",
"maintainer_email": "",
"keywords": "sentence splitter tokenization tokenizer tokenize",
"author": "Philip Koehn, Josh Schroeder, Digital Silk Road, Linas Valiukas",
"author_email": "lvaliukas@cyber.law.harvard.edu",
"download_url": "https://files.pythonhosted.org/packages/20/b3/86b431fe7002ba006c08b8559d2ad78e1153bfc515a453cc96d2f55a2c40/sentence_splitter-1.4.tar.gz",
"platform": "any",
"description": "\nText to sentence splitter using heuristic algorithm by Philipp Koehn and Josh Schroeder.\n\nThis module allows splitting of text paragraphs into sentences. It is based on scripts developed by Philipp\nKoehn and Josh Schroeder for processing the `Europarl corpus <http://www.statmt.org/europarl/>`_.\n\nThe module is a port of `Lingua::Sentence Perl module <http://search.cpan.org/perldoc?Lingua::Sentence>`_ with\nsome extra additions (improved non-breaking prefix lists for some languages and added support for Danish,\nFinnish, Lithuanian, Norwegian (Bokm\u00e5l), Romanian, and Turkish).\n\n\n",
"bugtrack_url": null,
"license": "LGPLv3",
"summary": "Text to sentence splitter using heuristic algorithm by Philipp Koehn and Josh Schroeder",
"version": "1.4",
"split_keywords": [
"sentence",
"splitter",
"tokenization",
"tokenizer",
"tokenize"
],
"urls": [
{
"comment_text": "",
"digests": {
"md5": "2997a3de186228e9d434f92bceb751ec",
"sha256": "5645a3ad9c348e4287f4bc73bd573d92dccd4139042fddd51fff0591f1376763"
},
"downloads": -1,
"filename": "sentence_splitter-1.4-py2.py3-none-any.whl",
"has_sig": false,
"md5_digest": "2997a3de186228e9d434f92bceb751ec",
"packagetype": "bdist_wheel",
"python_version": "py2.py3",
"requires_python": ">=3.5",
"size": 44998,
"upload_time": "2019-01-14T17:11:23",
"upload_time_iso_8601": "2019-01-14T17:11:23.974499Z",
"url": "https://files.pythonhosted.org/packages/4a/ae/3bd609c760d57849d7ddf223762f1881f3c4df6467f4eadb3a33652b7e0d/sentence_splitter-1.4-py2.py3-none-any.whl",
"yanked": false,
"yanked_reason": null
},
{
"comment_text": "",
"digests": {
"md5": "a58c1f759d0b8ce0fe1e0c99b150ecb1",
"sha256": "3d1d773d07cc733ca2955aa87d0fa1c0a7274c6bdeec1daac5c5e92efb512f63"
},
"downloads": -1,
"filename": "sentence_splitter-1.4.tar.gz",
"has_sig": false,
"md5_digest": "a58c1f759d0b8ce0fe1e0c99b150ecb1",
"packagetype": "sdist",
"python_version": "source",
"requires_python": ">=3.5",
"size": 30607,
"upload_time": "2019-01-14T17:11:25",
"upload_time_iso_8601": "2019-01-14T17:11:25.388441Z",
"url": "https://files.pythonhosted.org/packages/20/b3/86b431fe7002ba006c08b8559d2ad78e1153bfc515a453cc96d2f55a2c40/sentence_splitter-1.4.tar.gz",
"yanked": false,
"yanked_reason": null
}
],
"upload_time": "2019-01-14 17:11:25",
"github": true,
"gitlab": false,
"bitbucket": false,
"github_user": "berkmancenter",
"github_project": "mediacloud-sentence-splitter",
"travis_ci": true,
"coveralls": true,
"github_actions": false,
"requirements": [
{
"name": "regex",
"specs": [
[
">=",
"2017.12.12"
]
]
}
],
"lcname": "sentence-splitter"
}