Edit model card

Model Name: DistilBERT for Sentiment Analysis

Model Description

Overview

This model is a fine-tuned version of distilbert-base-uncased on a social media dataset for the purpose of sentiment analysis. It can classify text into non-negative and negative sentiments.

Intended Use

This model is intended for sentiment analysis tasks, particularly for analyzing social media texts.

Model Architecture

This model is based on the DistilBertForSequenceClassification architecture, a distilled version of BERT that maintains comparable performance on downstream tasks while being more computationally efficient.

Training

Training Data

The model was trained on a dataset consisting of social media posts, surveys and interviews, labeled for sentiment (non-negative and negative). The dataset includes texts from a variety of sources and demographics.

Training Procedure

The model was trained using the following parameters:

  • Optimizer: AdamW
  • Learning Rate: 5e-5
  • Batch Size: 32
  • Epochs: 30

Training was conducted on Kaggle, utilizing two GPUs for accelerated training.

Downloads last month
330
Safetensors
Model size
67M params
Tensor type
F32
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.