@@ -4,7 +4,7 @@ build-backend = "setuptools.build_meta"
4
4
5
5
[project ]
6
6
name = " lm_eval"
7
- version = " 0.4.0 "
7
+ version = " 0.4.1 "
8
8
authors = [
9
9
{name =" EleutherAI" , email =" contact@eleuther.ai" }
10
10
]
@@ -56,16 +56,14 @@ Repository = "https://github.com/EleutherAI/lm-evaluation-harness"
56
56
[project .optional-dependencies ]
57
57
anthropic = [" anthropic" ]
58
58
dev = [" pytest" , " pytest-cov" , " pytest-xdist" , " pre-commit" , " mypy" ]
59
- gptq = [" auto-gptq[triton] @ git+https://github.com/PanQiWei/AutoGPTQ " ]
59
+ gptq = [" auto-gptq[triton]>=0.6.0 " ]
60
60
ifeval = [" langdetect" , " immutabledict" ]
61
61
mamba = [" mamba_ssm" , " causal-conv1d==1.0.2" ]
62
62
math = [" sympy>=1.12" , " antlr4-python3-runtime==4.11" ]
63
63
multilingual = [" nagisa>=0.2.7" , " jieba>=0.42.1" , " pycountry" ]
64
64
openai = [" openai==1.3.9" , " tiktoken" ]
65
65
optimum = [" optimum[openvino]" ]
66
- promptsource = [
67
- " promptsource @ git+https://github.com/bigscience-workshop/promptsource.git#egg=promptsource"
68
- ]
66
+ promptsource = [" promptsource>=0.2.3" ]
69
67
sentencepiece = [" sentencepiece>=0.1.98" , " protobuf>=4.22.1" ]
70
68
testing = [" pytest" , " pytest-cov" , " pytest-xdist" ]
71
69
vllm = [" vllm<=0.2.5" ]
0 commit comments