Edit model card

T3Q-MSlerp-13B

T3Q-MSlerp-13B is a merge of the following models using mergekit:

Model Developers Chihoon Lee(chlee10), T3Q

  
  slices:
      - sources:
        - model: zhengr/MixTAO-7Bx2-MoE-Instruct-v7.0
          layer_range: [0, 32]
        - model: yunconglong/13B_MATH_DPO
          layer_range: [0, 32]
  
  merge_method: slerp
  base_model: zhengr/MixTAO-7Bx2-MoE-Instruct-v7.0
  
  parameters:
   t:
      - filter: self_attn
        value: [0, 0.5, 0.3, 0.7, 1]
      - filter: mlp
        value: [1, 0.5, 0.7, 0.3, 0]
      - value: 0.5 # fallback for rest of tensors
      
  dtype: float16
Downloads last month
14
Safetensors
Model size
12.9B params
Tensor type
F32
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.