🦎UTC

Overview

Prompt-based token classification model based on DeBERTaV3 and T5 encoders. A model trained on a variety of token classification tasks demonstrates great generalization capabilities. It excels in zero-shot and few-shot settings, for diverse information extraction (IE) tasks, making it a versatile tool for a range of NLP applications.

Supported IE tasks

  • Named-entity recognition (NER)

  • Relation extraction

  • Summarization

  • Q&A

  • Text cleaning

  • Coreference resolution

Models

Model
Base model
Size
Input capacity
Access

UTC-small

141M

3K

English

Open-sourced under Apache 2.0

UTC-large

434M

3K

English

Open-sourced under Apache 2.0

UTC-base

184M

3K

English

Open-sourced under Apache 2.0

UTC-large

783M

3K

English

Open-sourced under Apache 2.0

Common features

  • Prompt-based The model was trained on multiple token classification tasks making it adaptable for a variety of information extraction tasks using user prompts.

  • Supports zero-shot and few-shot learning. Capable of performing tasks with little to no training data, making it highly adaptable to new challenges.

  • 3K Token Capacity. Can process texts up to 3,000 tokens in length. We work on expanding model processing capacity.

  • Currently supports the English language only.

Usage instructions

Examples

Zero-shot NER

Question answering

Text cleaning

Relation extraction

Fine-tuning

Currently, you can fine-tune our model via the Hugging Face Auto-train feature.

Potential and limitations

  • Potential. The UTC-DeBERTa-small model's prompt-based approach allows for flexible adaptation to various tasks. Its strength in token-level analysis makes it highly effective for detailed text-processing tasks.

  • Limitations. While the model shows promise in summarization, it is currently not its strongest application. Enhancements in this area are a future development focus.

Last updated