Barcenas 27b
Based on the SillyTilly/google-gemma-2-27b-it and trained with the dataset pinzhenchen/alpaca-cleaned-es in the Spanish language.
The goal of this model is to have a relatively large model optimized in Spanish and that was at the level of the first versions of GPT-4.
I am proud of this model for being the biggest and most powerful one I have done, no doubt it is the result of my short stay in the AI world.
Made with ❤️ in Guadalupe, Nuevo Leon, Mexico 🇲🇽
- Downloads last month
- 43
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.