sentence-splitter
spacy-experimental
sentence-splitter | spacy-experimental | |
---|---|---|
1 | 5 | |
216 | 94 | |
6.0% | - | |
0.0 | 3.8 | |
over 1 year ago | 27 days ago | |
Python | Python | |
GNU General Public License v3.0 or later | MIT License |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
sentence-splitter
-
Text translation question: Helsinki-NLP skips end sentences. Any good open sourced pre-trained models for large text translation?
There are plenty of sentence splitter available, like https://github.com/mediacloud/sentence-splitter for example, but sometimes you'll have to use language specific ones.
spacy-experimental
-
Newbie question with Spacy Coreference Resolution
Trying to work with the newly released coreference resolution pipeline
-
spaCy just got an experimental feature to detect co-references
I think the details are mentioned here: https://github.com/explosion/spacy-experimental/releases/tag/v0.6.0
- SpanFinder is a new experimental spaCy component that identifies span boundaries
-
Cython Is 20
I can't speak for the parent commenter, but there is ofte. code 'around' the machine learning code that benefits from high-performance implementations. To give two examples:
1. We recently implemented an edit tree lemmatizer for spaCy. The machine learning model predicts labels that map to edit trees. However, in order to lemmatize tokens, the trees need to be applied. I implemented all the tree wrangling in Cython to speed up processing and save memory (trees can be encoded as compact C unions):
https://github.com/explosion/spaCy/blob/master/spacy/pipelin...
2. I am working on a biaffine parser for spaCy. Most implementations of biaffine parsing use a Python implementation of MST decoding, which is unfortunately quite slow. Some people have reported it to dominate parsing time (rather than a rather expensive transformer + biaffine layer). I have implemented MST decoding in Cython and it barely shows up in profiles:
https://github.com/explosion/spacy-experimental/blob/master/...
-
Utilizando Neural edit-tree lemmatization para o português
Nós iremos utilizar o template do edit_tree_lemmatizer contido da pasta de projetos do repositório https://github.com/explosion/spacy-experimental e modificaremos para treinar um modelo em português em vez de alemão.
What are some alternatives?
word-piece-tokenizer - A Lightweight Word Piece Tokenizer
neuralcoref - ✨Fast Coreference Resolution in spaCy with Neural Networks
Hebrew-Tokenizer - A very simple python tokenizer for Hebrew text.
word_forms - Accurately generate all possible forms of an English word e.g "election" --> "elect", "electoral", "electorate" etc.
bitextor - Bitextor generates translation memories from multilingual websites
nanobind - nanobind: tiny and efficient C++/Python bindings
xontrib-output-search - Get identifiers, paths, URLs and words from the previous command output and use them for the next command in xonsh shell.
warp - A Python framework for high performance GPU simulation and graphics
sentimental-onix - sentiment analysis for spacy pipeline in python
jax - Composable transformations of Python+NumPy programs: differentiate, vectorize, JIT to GPU/TPU, and more
epython - EPython is a typed-subset of the Python for extending the language new builtin types and methods
projects - 🪐 End-to-end NLP workflows from prototype to production