seldonium-2x7b-MoE-v0.1 / mergekit_moe_config.yml
jomangbp's picture
Upload folder using huggingface_hub
51a0cbd verified
raw
history blame
963 Bytes
base_model: NousResearch/Hermes-2-Pro-Mistral-7B
gate_mode: cheap_embed # Use hidden state representations to determine MoE gates
dtype: bfloat16 # Output data type
experts_per_token: 2 # Number of experts per token
experts:
- source_model: Locutusque/Hercules-4.0-Mistral-v0.2-7B
positive_prompts:
- "Write a Python function to calculate the factorial of a number."
- "Implement a quicksort algorithm to sort a list of integers."
- "Design a Python class to represent a binary search tree."
- source_model: Open-Orca/Mistral-7B-OpenOrca
positive_prompts:
- "Solve the logic puzzle: 'If Ann is older than Belinda, and Belinda is younger than Cathy, who is the oldest?'"
- "Analyze the argument: 'All cats are animals. Some animals are pets. Therefore, all cats are pets.' Determine if the conclusion follows logically from the premises."
- "Evaluate the validity of the statements: 'A is true. A is false.'"