🦎UTC
Overview
Prompt-based token classification model based on DeBERTaV3 and T5 encoders. A model trained on a variety of token classification tasks demonstrates great generalization capabilities. It excels in zero-shot and few-shot settings, for diverse information extraction (IE) tasks, making it a versatile tool for a range of NLP applications.
Supported IE tasks
Named-entity recognition (NER)
Relation extraction
Summarization
Q&A
Text cleaning
Coreference resolution
Models
Common features
Prompt-based The model was trained on multiple token classification tasks making it adaptable for a variety of information extraction tasks using user prompts.
Supports zero-shot and few-shot learning. Capable of performing tasks with little to no training data, making it highly adaptable to new challenges.
3K Token Capacity. Can process texts up to 3,000 tokens in length. We work on expanding model processing capacity.
Currently supports the English language only.
Usage instructions
Examples
Zero-shot NER
Question answering
Text cleaning
Relation extraction
Fine-tuning
Currently, you can fine-tune our model via the Hugging Face Auto-train feature.
Potential and limitations
Potential. The UTC-DeBERTa-small model's prompt-based approach allows for flexible adaptation to various tasks. Its strength in token-level analysis makes it highly effective for detailed text-processing tasks.
Limitations. While the model shows promise in summarization, it is currently not its strongest application. Enhancements in this area are a future development focus.
Last updated