Reverse Dependencies of seqeval
The following projects have a declared dependency on seqeval:
- OmniEvent — A tookit for event extraction.
- optimum — Optimum Library is an extension of the Hugging Face Transformers library, providing a framework to integrate third-party libraries from Hardware Partners and interface with their specific functionality.
- paddlenlp — Easy-to-use and powerful NLP library with Awesome model zoo, supporting wide-range of NLP tasks from research to industrial applications, including Neural Search, Question Answering, Information Extraction and Sentiment Analysis end-to-end system.
- papermage — Papermage. Casting magic over scientific PDFs.
- peal — A package dedicated to using PEFT for active-learning, hence PEAL.
- pre-ai-python — Microsoft AI Python Package
- pt-pump-up — Hub for Portuguese NLP resources
- pyabsa — This tool provides the state-of-the-art models for aspect term extraction (ATE), aspect polarity classification (APC), and text classification (TC).
- pynami — Deep Learning - Transformers
- qurator-sbb-tools — Qurator
- renard-pipeline — Relationships Extraction from NARrative Documents
- scandeval — Evaluation of pretrained language models on mono- or multilingual language tasks.
- SciAssist — A toolkit for Scientific Document Processing
- seqtag-keras — Easy to use BiLSTM+CRF sequence tagging for text.
- sequence-classifier — no summary
- sikufenci — NLP tool for Ancient Chinese word segmentation.
- simpletransformers — An easy-to-use wrapper library for the Transformers library.
- simpletransformers-fork-trialandsuccess — An easy-to-use wrapper library for the Transformers library. FORK: This fork adds T5TokenizerFast and umT5 support.
- simpletransformers-le — An easy-to-use wrapper library for the Transformers library.
- span-marker — Named Entity Recognition using Span Markers
- sparseml — Libraries for applying sparsification recipes to neural networks with a few lines of code, enabling faster and smaller models
- sparseml-nightly — Libraries for applying sparsification recipes to neural networks with a few lines of code, enabling faster and smaller models
- ststransformers — An easy-to-use wrapper library for using Transformers in Semantic Textual Similarity Tasks.
- tailors — no summary
- tasknet — Seamless integration of tasks with huggingface models
- tcrf — A deep learning based sequence tagging library with CRF layer on the top of transformer models.
- temporal-taggers — Neural temporal taggers with Transformer architectures
- text-machina — Text Machina: Seamless Generation of Machine-Generated Text Datasets
- tf-models-nightly — TensorFlow Official Models
- tf-models-no-deps — TensorFlow Official Models
- tf-models-official — TensorFlow Official Models
- TFOD-Automatic — Automated Object detection for Beginner using python and Tensorflow
- transformers-crf — Transformers CRF: CRF Token Classification for Transformers
- transformers-keras — Transformer-based models implemented in tensorflow 2.x(Keras)
- transquest — Transformer based translation quality estimation
- trapper — State-of-the-art NLP through transformer models in a modular design and consistent APIs.
- treform — A text mining tool for Korean and English
- turkish-lm-tuner — Implementation of the Turkish LM Tuner
- ws-benchmark — a weak supervision learning benchmark
- x-tagger — A Natural Language Processing toolkit for token classification in its simplest form.
- yonlu — A deep learning based natural language understanding module
- zhpr — no summary
- zshot — Zero and Few shot named entity recognition
1
2