Suparious's picture
Update README.md
98d2528 verified
|
raw
history blame
1.7 kB
---
language:
- en
license: apache-2.0
library_name: transformers
tags:
- 4-bit
- AWQ
- text-generation
- autotrain_compatible
- endpoints_compatible
- text-generation-inference
- transformers
- unsloth
- mistral
- trl
- code
- 'medical '
- farmer
- doctor
- Mega-Series
- Cyber-Series
- Role-Play
- Self-Rag
- ThinkingBot
- milestone
- mega-series
- SpydazWebAI
base_model: LeroyDyer/Mixtral_AI_CyberTron_Ultra
pipeline_tag: text-generation
inference: false
metrics:
- accuracy
- bertscore
- bleu
- brier_score
- cer
- character
- charcut_mt
- chrf
- code_eval
datasets:
- gretelai/synthetic_text_to_sql
- HuggingFaceTB/cosmopedia
- teknium/OpenHermes-2.5
- Open-Orca/SlimOrca
- Open-Orca/OpenOrca
- cognitivecomputations/dolphin-coder
- databricks/databricks-dolly-15k
- yahma/alpaca-cleaned
- uonlp/CulturaX
- mwitiderrick/SwahiliPlatypus
- swahili
- Rogendo/English-Swahili-Sentence-Pairs
- ise-uiuc/Magicoder-Evol-Instruct-110K
- meta-math/MetaMathQA
quantized_by: Suparious
---
# LeroyDyer/Mixtral_AI_CyberTron_Ultra AWQ
- Model creator: [LeroyDyer](https://huggingface.co/LeroyDyer)
- Original model: [Mixtral_AI_CyberTron_Ultra](https://huggingface.co/LeroyDyer/Mixtral_AI_CyberTron_Ultra)
## Model Summary
What does he NOT KNOW ! that is the question!
### MOTTO FOR MODEL!
## Models are the same as loras , take them with light weight they are like tablets of knowledge!
Exactly ! ( models / loras ? is there a difference ? only mega merges make a true difference !
the small merges are just applying an adapter lol - Its in there somewhere?)
### Ok Its a Great MODEL ! (My Favorite Goto Brain now ! - will be fine tuned even more ! (if i get cloud credits))