zli12321 commited on
Commit
a9c1982
1 Parent(s): 59ceaa0

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -1
README.md CHANGED
@@ -17,9 +17,11 @@ pipeline_tag: text-classification
17
  QA-Evaluation-Metrics is a fast and lightweight Python package for evaluating question-answering models and prompting of black-box and open-source large language models. It provides various basic and efficient metrics to assess the performance of QA models.
18
 
19
  ### Updates
20
- - Uopdated to version 0.2.8
21
  - Supports prompting OPENAI GPT-series models and Claude Series models now. (Assuimg OPENAI version > 1.0)
22
  - Supports prompting various open source models such as LLaMA-2-70B-chat, LLaVA-1.5 etc by calling API from [deepinfra](https://deepinfra.com/models).
 
 
23
 
24
 
25
  ## Installation
 
17
  QA-Evaluation-Metrics is a fast and lightweight Python package for evaluating question-answering models and prompting of black-box and open-source large language models. It provides various basic and efficient metrics to assess the performance of QA models.
18
 
19
  ### Updates
20
+ - Uopdated to version 0.2.17
21
  - Supports prompting OPENAI GPT-series models and Claude Series models now. (Assuimg OPENAI version > 1.0)
22
  - Supports prompting various open source models such as LLaMA-2-70B-chat, LLaVA-1.5 etc by calling API from [deepinfra](https://deepinfra.com/models).
23
+ - Added trained tiny-bert for QA evaluation. Model size is 18 MB.
24
+ - Pass huggingface repository name to download model directly for TransformerMatcher
25
 
26
 
27
  ## Installation