Analyze sentiment of text
Analyze financial sentiment in text
Analyze text for sentiment in real-time
Analyze sentiment of articles related to a trading asset
Predict sentiment of a text comment
Analyze cryptocurrency articles for sentiment
Analyze sentiment of Twitter tweets
Predict emotion from text
Analyze sentiment in your text
Analyze financial news sentiment from text or URL
Text_Classification_App
Analyze YouTube comments' sentiment
Enter your mood for yoga recommendations
Distilbert Distilbert Base Uncased Finetuned Sst 2 English is a fine-tuned version of the DistilBERT model, specifically trained for sentiment analysis in English. It is based on the DistilBERT Base Uncased model, which is a more efficient and compact alternative to the original BERT model. This model has been further fine-tuned on the SST-2 dataset, a widely used benchmark for sentiment analysis, making it particularly effective for binary sentiment classification tasks (positive or negative sentiment).
• Pre-trained on large-scale corpus: The model benefits from the extensive pre-training of DistilBERT, which captures general language understanding. • Fine-tuned for sentiment analysis: It is specialized for sentiment classification, making it highly accurate for detecting positive or negative sentiment in text. • Compact architecture: DistilBERT uses 6 layers compared to BERT's 12, reducing computational requirements while maintaining strong performance. • Uncased version: It treats all text as lowercase, simplifying preprocessing. • Hugging Face compatible: Can be easily integrated into workflows using the Hugging Face Transformers library.
Install the Hugging Face Transformers library if not already installed:
pip install transformers
Import necessary components:
from transformers import AutoModelForSequenceClassification, AutoTokenizer
Load the model and tokenizer:
model_name = "distilbert-base-uncased-finetuned-sst-2-english"
model = AutoModelForSequenceClassification.from_pretrained(model_name)
tokenizer = AutoTokenizer.from_pretrained(model_name)
Prepare your input text:
text = "I really enjoyed this movie!"
Tokenize the text:
inputs = tokenizer(text, return_tensors="pt", truncation=True)
Run the model to get predictions:
with torch.no_grad():
outputs = model(**inputs)
logits = outputs.logits
** Interpret the results**: Sentiment is classified as positive (logit[0] < logit[1]) or negative (logit[0] > logit[1]).
What type of input does this model expect?
This model expects raw English text as input, which will be tokenized and processed internally.
What does the output of the model represent?
The output consists of logits, which are raw (unnormalized) prediction scores. To get probabilities, apply a softmax function to the logits.
How accurate is this model?
This model achieves state-of-the-art performance on the SST-2 dataset, with an accuracy of over 92%, making it highly reliable for sentiment analysis tasks.