Sentiment Transformer β€” tango

A small (β‰ˆ13M parameter) transformer encoder trained entirely from scratch for 3-class sentiment analysis (negative / neutral / positive).

Architecture

Pre-layer-norm transformer encoder with [CLS] pooling and a linear classification head. Built with pure torch.nn β€” no pretrained weights.

Parameter Value
Hidden dim 256
FFN dim 1024
Layers 6
Heads 8
Max seq len 256
Vocab size 16000
Labels NEGATIVE, NEUTRAL, POSITIVE
Precision bf16 mixed-precision

Training Data

Trained on a combined corpus of:

  • IMDB (50k movie reviews)
  • Sentiment140 (1M tweets)
  • Yelp (1M reviews)
  • SST-5 (fine-grained β†’ 3-class)
  • Financial PhraseBank (finance headlines)
  • TweetEval (SemEval-2017 tweets)

Usage

from transformers import AutoModelForSequenceClassification, AutoTokenizer, pipeline

model = AutoModelForSequenceClassification.from_pretrained(
    "Impulse2000/sentiment-transformer", trust_remote_code=True
)
tokenizer = AutoTokenizer.from_pretrained("Impulse2000/sentiment-transformer")
pipe = pipeline("text-classification", model=model, tokenizer=tokenizer)
print(pipe("This movie was absolutely fantastic!"))
Downloads last month
-
Safetensors
Model size
8.92M params
Tensor type
F32
Β·
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Datasets used to train Impulse2000/sentiment-transformer