--- datasets: - tatsu-lab/alpaca language: - en pipeline_tag: text2text-generation --- # Model Details - **Model name:** alpaca-flan-T5-large - **Model type:** - Text2Text Generation - **Parent Model:** [google/flan-t5-large](https://huggingface.co/google/flan-t5-large) - **Training dataset:** [Alpaca](https://huggingface.co/datasets/tatsu-lab/alpaca) - **Language:** English - **Framework:** PyTorch - **Model version:** 1.0 # Intended Use # How to Use ```python from transformers import pipeline ``` # Training Details The model was trained on 3xV100 GPUs Hyperparameters: learning_rate = 5e-5 batch_size = 128 epochs = 3 ``` # Limitations and Bias The model is based on a large and diverse dataset, but it may still have limitations and biases in certain areas. Some limitations include: - Language: The model is designed to work with English text only and may not perform as well in other languages. In addition, the model may have some bias in terms of the data it was trained on. The dataset includes questions from a variety of sources, but it may not be representative of all populations or perspectives. As a result, the model may perform better or worse for certain types of questions or on certain types of texts.