task
stringclasses 1
value | org
stringclasses 7
values | model
stringclasses 9
values | hardware
stringclasses 2
values | date
stringclasses 9
values | prefill
dict | decode
dict | preprocess
dict |
---|---|---|---|---|---|---|---|
text_generation | NousResearch | Hermes-3-Llama-3.1-8B | a100-large | 2024-10-30-23-07-10 | {
"efficency": {
"unit": "tokens/kWh",
"value": 7896383.83303204
},
"energy": {
"cpu": 0.010771365165906027,
"gpu": 0.03499995191660332,
"ram": 0.000042438637758223256,
"total": 0.04581375572026757,
"unit": "kWh"
}
} | {
"efficiency": {
"unit": "tokens/kWh",
"value": 383715.9018998692
},
"energy": {
"cpu": 0.006909195393007252,
"gpu": 0.016518416020279855,
"ram": 0.000027241396927033754,
"total": 0.023454852810214137,
"unit": "kWh"
}
} | {
"efficiency": {
"unit": "samples/kWh",
"value": 19314595.743536726
},
"energy": {
"cpu": 0.00003136402598271767,
"gpu": 0.000020315571802598242,
"ram": 9.47190721638247e-8,
"total": 0.000051774316857479744,
"unit": "kWh"
}
} |
text_generation | meta-llama | Llama-3.1-8B-Instruct | a100-large | 2024-10-30-19-38-09 | {
"efficency": {
"unit": "tokens/kWh",
"value": 7875937.087230494
},
"energy": {
"cpu": 0.010796036717298752,
"gpu": 0.035094443769978056,
"ram": 0.00004221247271874213,
"total": 0.04593269295999555,
"unit": "kWh"
}
} | {
"efficiency": {
"unit": "tokens/kWh",
"value": 373534.5783406723
},
"energy": {
"cpu": 0.006907515534659228,
"gpu": 0.01715961075545351,
"ram": 0.000027028633857338117,
"total": 0.024094154923970088,
"unit": "kWh"
}
} | {
"efficiency": {
"unit": "samples/kWh",
"value": 18903508.14097379
},
"energy": {
"cpu": 0.000031577332566181814,
"gpu": 0.000021227516981525696,
"ram": 9.538423654320006e-8,
"total": 0.000052900233784250705,
"unit": "kWh"
}
} |
text_generation | EleutherAI | pythia-1.4b | a10g-large | 2024-10-25-14-19-11 | {
"efficency": {
"unit": "tokens/kWh",
"value": 54946841.77704233
},
"energy": {
"cpu": 0.0006950257205501962,
"gpu": 0.004783902188229749,
"ram": 0.000007229079517594969,
"total": 0.00548615698829754,
"unit": "kWh"
}
} | {
"efficiency": {
"unit": "tokens/kWh",
"value": 1185499.0738062232
},
"energy": {
"cpu": 0.001532609044249083,
"gpu": 0.006043184112321255,
"ram": 0.00001594622041976216,
"total": 0.0075917393769901025,
"unit": "kWh"
}
} | {
"efficiency": {
"unit": "samples/kWh",
"value": 33069647.189926323
},
"energy": {
"cpu": 0.000011265942924405358,
"gpu": 0.00001888834844376852,
"ram": 8.491845032760329e-8,
"total": 0.00003023920981850148,
"unit": "kWh"
}
} |
text_generation | microsoft | phi-2 | a10g-large | 2024-10-25-00-12-06 | {
"efficency": {
"unit": "tokens/kWh",
"value": 25532119.335130304
},
"energy": {
"cpu": 0.001485621188480941,
"gpu": 0.010244187250898752,
"ram": 0.000014692518964651205,
"total": 0.011744500958344343,
"unit": "kWh"
}
} | {
"efficiency": {
"unit": "tokens/kWh",
"value": 653974.0928690574
},
"energy": {
"cpu": 0.002664958041045172,
"gpu": 0.011070685912097256,
"ram": 0.000026369096206358968,
"total": 0.013762013049348782,
"unit": "kWh"
}
} | {
"efficiency": {
"unit": "samples/kWh",
"value": 2594586258.0208464
},
"energy": {
"cpu": 3.829965067173665e-7,
"gpu": 0,
"ram": 2.4213983180210174e-9,
"total": 3.8541790503538754e-7,
"unit": "kWh"
}
} |
text_generation | allenai | OLMo-1B-hf | a10g-large | 2024-10-24-18-23-56 | {
"efficency": {
"unit": "tokens/kWh",
"value": 62297846.58810162
},
"energy": {
"cpu": 0.0006078373521613038,
"gpu": 0.004224683213077207,
"ram": 0.0000062823229751878135,
"total": 0.0048388028882136985,
"unit": "kWh"
}
} | {
"efficiency": {
"unit": "tokens/kWh",
"value": 1390874.1413723528
},
"energy": {
"cpu": 0.0012634316170816028,
"gpu": 0.005194258099847594,
"ram": 0.000013061109683948865,
"total": 0.006470750826613145,
"unit": "kWh"
}
} | {
"efficiency": {
"unit": "samples/kWh",
"value": 35562851.946256354
},
"energy": {
"cpu": 0.00001108087887082042,
"gpu": 0.000016956680231938748,
"ram": 8.167052233473494e-8,
"total": 0.000028119229625093902,
"unit": "kWh"
}
} |
text_generation | openai-community | gpt2-large | a10g-large | 2024-10-25-15-05-20 | {
"efficency": {
"unit": "tokens/kWh",
"value": 77907728.2032169
},
"energy": {
"cpu": 0.00042952262899040045,
"gpu": 0.0027924934839930414,
"ram": 0.000004223514289139063,
"total": 0.0032262396272725808,
"unit": "kWh"
}
} | {
"efficiency": {
"unit": "tokens/kWh",
"value": 1343266.704857417
},
"energy": {
"cpu": 0.0016164460666061049,
"gpu": 0.005067730581958951,
"ram": 0.00001590753237599822,
"total": 0.006700084180941058,
"unit": "kWh"
}
} | {
"efficiency": {
"unit": "samples/kWh",
"value": 34837547.66517286
},
"energy": {
"cpu": 0.000010432217698123875,
"gpu": 0.000018202236783837478,
"ram": 7.020663859453463e-8,
"total": 0.00002870466112055589,
"unit": "kWh"
}
} |
text_generation | HuggingFaceTB | SmolLM-135M | a10g-large | 2024-10-23-19-09-15 | {
"efficency": {
"unit": "tokens/kWh",
"value": 218923397.1364946
},
"energy": {
"cpu": 0.00028216661832921095,
"gpu": 0.0011204146741087939,
"ram": 0.000002878352463094162,
"total": 0.001405459644901099,
"unit": "kWh"
}
} | {
"efficiency": {
"unit": "tokens/kWh",
"value": 1478710.1205686298
},
"energy": {
"cpu": 0.0020209151121617142,
"gpu": 0.004044847485875458,
"ram": 0.000020623012510417966,
"total": 0.00608638561054759,
"unit": "kWh"
}
} | {
"efficiency": {
"unit": "samples/kWh",
"value": 35187752.05342098
},
"energy": {
"cpu": 0.000010718486329682895,
"gpu": 0.000017620569652265772,
"ram": 7.992339706452299e-8,
"total": 0.00002841897937901319,
"unit": "kWh"
}
} |
text_generation | HuggingFaceTB | SmolLM-1.7B | a10g-large | 2024-10-24-15-16-02 | {
"efficency": {
"unit": "tokens/kWh",
"value": 41960978.223952904
},
"energy": {
"cpu": 0.0009210179793032473,
"gpu": 0.006402295760721532,
"ram": 0.000009403775706541059,
"total": 0.007332717515731321,
"unit": "kWh"
}
} | {
"efficiency": {
"unit": "tokens/kWh",
"value": 986792.3688060087
},
"energy": {
"cpu": 0.0017958659208215804,
"gpu": 0.007306251844996492,
"ram": 0.000018341901791392123,
"total": 0.009120459667609458,
"unit": "kWh"
}
} | {
"efficiency": {
"unit": "samples/kWh",
"value": 37376068.58504866
},
"energy": {
"cpu": 0.000010438337355784218,
"gpu": 0.000016238901880072376,
"ram": 7.78486547950319e-8,
"total": 0.000026755087890651626,
"unit": "kWh"
}
} |
text_generation | HuggingFaceTB | SmolLM-360M | a10g-large | 2024-10-24-14-22-22 | {
"efficency": {
"unit": "tokens/kWh",
"value": 124481811.63546507
},
"energy": {
"cpu": 0.00038417323045180534,
"gpu": 0.0020836598613710013,
"ram": 0.000003917567631278797,
"total": 0.002471750659454085,
"unit": "kWh"
}
} | {
"efficiency": {
"unit": "tokens/kWh",
"value": 1315130.308155782
},
"energy": {
"cpu": 0.0021575579513180274,
"gpu": 0.004663859981084995,
"ram": 0.00002201039115281833,
"total": 0.0068434283235558405,
"unit": "kWh"
}
} | {
"efficiency": {
"unit": "samples/kWh",
"value": 36616831.97480331
},
"energy": {
"cpu": 0.000010503983233461947,
"gpu": 0.000016727513381886716,
"ram": 7.834823109019483e-8,
"total": 0.00002730984484643886,
"unit": "kWh"
}
} |
Analysis of energy usage for HUGS models
Based on the energy_star branch of optimum-benchmark, and using codecarbon.
Fields
- task: Task the model was benchmarked on.
- org: Organization hosting the model.
- model: The specific model. Model names at HF are usually constructed with {org}/{model}.
- date: The date that the benchmark was run.
- prefill: The esimated energy and efficiency for prefilling.
- decode: The estimated energy and efficiency for decoding.
- preprocess: The estimated energy and efficiency for preprocessing.
Code to Reproduce
As I'm devving, I'm hopping between https://huggingface.co/spaces/AIEnergyScore/benchmark-hugs-models and https://huggingface.co/spaces/meg/CalculateCarbon
From there, python code/make_pretty_dataset.py
(included in this repository) takes the raw results and uploads them to the dataset here.
- Downloads last month
- 19