spacy-experimental
sentence-splitter
spacy-experimental | sentence-splitter | |
---|---|---|
5 | 1 | |
94 | 216 | |
- | 6.0% | |
4.2 | 0.0 | |
18 days ago | over 1 year ago | |
Python | Python | |
MIT License | GNU General Public License v3.0 or later |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
spacy-experimental
-
Newbie question with Spacy Coreference Resolution
Trying to work with the newly released coreference resolution pipeline
-
spaCy just got an experimental feature to detect co-references
I think the details are mentioned here: https://github.com/explosion/spacy-experimental/releases/tag/v0.6.0
- SpanFinder is a new experimental spaCy component that identifies span boundaries
-
Cython Is 20
I can't speak for the parent commenter, but there is ofte. code 'around' the machine learning code that benefits from high-performance implementations. To give two examples:
1. We recently implemented an edit tree lemmatizer for spaCy. The machine learning model predicts labels that map to edit trees. However, in order to lemmatize tokens, the trees need to be applied. I implemented all the tree wrangling in Cython to speed up processing and save memory (trees can be encoded as compact C unions):
https://github.com/explosion/spaCy/blob/master/spacy/pipelin...
2. I am working on a biaffine parser for spaCy. Most implementations of biaffine parsing use a Python implementation of MST decoding, which is unfortunately quite slow. Some people have reported it to dominate parsing time (rather than a rather expensive transformer + biaffine layer). I have implemented MST decoding in Cython and it barely shows up in profiles:
https://github.com/explosion/spacy-experimental/blob/master/...
-
Utilizando Neural edit-tree lemmatization para o português
Nós iremos utilizar o template do edit_tree_lemmatizer contido da pasta de projetos do repositório https://github.com/explosion/spacy-experimental e modificaremos para treinar um modelo em português em vez de alemão.
sentence-splitter
-
Text translation question: Helsinki-NLP skips end sentences. Any good open sourced pre-trained models for large text translation?
There are plenty of sentence splitter available, like https://github.com/mediacloud/sentence-splitter for example, but sometimes you'll have to use language specific ones.
What are some alternatives?
neuralcoref - ✨Fast Coreference Resolution in spaCy with Neural Networks
word-piece-tokenizer - A Lightweight Word Piece Tokenizer
word_forms - Accurately generate all possible forms of an English word e.g "election" --> "elect", "electoral", "electorate" etc.
Hebrew-Tokenizer - A very simple python tokenizer for Hebrew text.
nanobind - nanobind: tiny and efficient C++/Python bindings
bitextor - Bitextor generates translation memories from multilingual websites
warp - A Python framework for high performance GPU simulation and graphics
xontrib-output-search - Get identifiers, paths, URLs and words from the previous command output and use them for the next command in xonsh shell.
sentimental-onix - sentiment analysis for spacy pipeline in python
jax - Composable transformations of Python+NumPy programs: differentiate, vectorize, JIT to GPU/TPU, and more
epython - EPython is a typed-subset of the Python for extending the language new builtin types and methods
projects - 🪐 End-to-end NLP workflows from prototype to production