Configuration Parsing
Warning:
In config.json: "quantization_config.bits" must be an integer
original model Meta-Llama-3-70B-Instruct
EXL2 quants of Meta-Llama-3-70B-Instruct
Located in the main branch
- 2.55 bits per weight
- measurement.json
์๋ณธ ๋ชจ๋ธ Meta-Llama-3-70B-Instruct
Meta-Llama-3-70B-Instruct ๋ชจ๋ธ EXL2 ์์ํ
๋ฉ์ธ branch์ ์๋ ํ์ผ
- 2.55 bits per weight
- measurement.json
- Downloads last month
- 6
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.