|
--- |
|
language: en |
|
tags: |
|
- autotrain |
|
- text-generation |
|
- llm |
|
- memes |
|
library_name: transformers |
|
library_version: [latest version at the time of training] |
|
model_type: llama 2 |
|
widget: |
|
- text: "When you try to code without coffee, " |
|
--- |
|
|
|
# Llama 2 Meme Generator |
|
|
|
## Model Description |
|
|
|
This model is a fine-tuned version of the `llama 2` model, specifically tailored for generating meme captions. It captures the essence and humor commonly found in popular internet memes and offers a unique approach to meme creation. Just provide a prompt or a meme context, and let the model generate a fitting caption! |
|
|
|
## Training Data |
|
|
|
The model was trained using a diverse dataset of meme captions, spanning various internet trends, jokes, and pop culture references. This ensures a wide range of meme generation capabilities, from classic meme formats to contemporary internet humor. |
|
|
|
## Training Procedure |
|
|
|
The model was fine-tuned using the `autotrain llm` command with optimal hyperparameters for meme generation. Special care was taken to avoid overfitting, ensuring the model can generalize well across various meme contexts. |
|
|
|
## Usage |
|
|
|
To generate a meme caption using this model, you can use the following code: |
|
|
|
```python |
|
from transformers import AutoTokenizer, AutoModelWithLMHead |
|
|
|
tokenizer = AutoTokenizer.from_pretrained("bickett/meme-llama") |
|
model = AutoModelWithLMHead.from_pretrained("bickett/meme-llama") |
|
|
|
input_text = "When you try to code without coffee" |
|
input_ids = tokenizer.encode(input_text, return_tensors="pt") |
|
output = model.generate(input_ids) |
|
|
|
print(tokenizer.decode(output[0], skip_special_tokens=True)) |
|
|