zirui3's picture
Upload README.md
57383e9
|
raw
history blame
589 Bytes
# Model summary
Train flan-T5-large on alpaca dataset with LoRA
# training
* torch==2.0.0+cu117
* transformers==4.28.0.dev0
* 8 x V100 32G
# How to use
```python
import transformers
from peft import PeftModel
base_model = transformers.AutoModelForSeq2SeqLM.from_pretrained("google/flan-t5-large")
peft_model = PeftModel.from_pretrained("zirui3/flan-t5-large-alpaca")
inputs = tokenizer("Any instruction that you like.", return_tensors="pt")
outputs = peft_model.generate(**inputs, max_length=128, do_sample=True)
print(tokenizer.batch_decode(outputs, skip_special_tokens=True)
```