Edit model card

Model Card for DistilBERT Text Classification

This is a DistilBERT model fine-tuned for text classification tasks.

Model Details

Model Description

This DistilBERT model is fine-tuned for text classification tasks. It is designed to classify texts into different categories based on the provided dataset.

  • Developed by: Thiago Adriano
  • Model type: DistilBERT for Sequence Classification
  • Language(s) (NLP): Portuguese
  • License: MIT License
  • Finetuned from model: distilbert-base-uncased

Model Sources

How to Get Started with the Model

Use the code below to get started with the model.

from transformers import AutoTokenizer, AutoModelForSequenceClassification

tokenizer = AutoTokenizer.from_pretrained("tadrianonet/distilbert-text-classification")
model = AutoModelForSequenceClassification.from_pretrained("tadrianonet/distilbert-text-classification")

inputs = tokenizer("Sample text for classification", return_tensors="pt")
outputs = model(**inputs)

Training Details

Training Data

The training data consists of text-label pairs in Portuguese. The data is preprocessed to tokenize the text and convert labels to numerical format.

Training Procedure

The model is fine-tuned using the Hugging Face Trainer API with the following hyperparameters:

  • Training regime: fp32
  • Learning rate: 2e-5
  • Batch size: 16
  • Epochs: 3

Speeds, Sizes, Times

  • Training time: Approximately 10 minutes on a single GPU

Evaluation

Testing Data, Factors & Metrics

Testing Data

The testing data is a separate set of text-label pairs used to evaluate the model's performance.

Factors

The evaluation is disaggregated by accuracy and loss.

Metrics

  • Accuracy: Measures the proportion of correct predictions
  • Loss: Measures the error in the model's predictions

Results

  • Evaluation Results:
    • Loss: 0.692
    • Accuracy: 50%

Summary

The model achieves 50% accuracy on the evaluation dataset, indicating that further fine-tuning and evaluation on a more diverse dataset may be necessary.

Model Examination

[More Information Needed]

Environmental Impact

Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).

  • Hardware Type: GPU
  • Hours used: 0.2 hours
  • Cloud Provider: [More Information Needed]
  • Compute Region: [More Information Needed]
  • Carbon Emitted: [More Information Needed]

Technical Specifications

Model Architecture and Objective

The model is based on DistilBERT, a smaller, faster, and cheaper version of BERT, designed for efficient text classification.

Compute Infrastructure

Hardware

  • Hardware Type: Single GPU
  • GPU Model: [More Information Needed]

Software

  • Framework: Transformers 4.x
  • Library: PyTorch

Citation

BibTeX:

1 bibtex @misc{thiago_adriano_2024_distilbert, author = {Thiago Adriano}, title = {DistilBERT Text Classification}, year = {2024}, publisher = {Hugging Face}, howpublished = {\url{https://huggingface.co/tadrianonet/distilbert-text-classification}}, } 1

APA:

Thiago Adriano. (2024). DistilBERT Text Classification. Hugging Face. https://huggingface.co/tadrianonet/distilbert-text-classification

More Information

For more details, visit the Hugging Face model page.

Model Card Authors

Thiago Adriano

Model Card Contact

For more information, contact Thiago Adriano at [[email protected]]

Downloads last month
4
Safetensors
Model size
67M params
Tensor type
F32
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.