Model doesnt load in vllm version 0.5.0 : ValueError: `rope_scaling`'s type field must be one of ['su', 'yarn'], got longrope
#1
by
sujjosep
- opened
llm = LLM(model=model_id, trust_remote_code=True, max_model_len=8196, tensor_parallel_size=number_gpus)
With trust remote code True
File ~/.local/lib/python3.11/site-packages/vllm/config.py:1287, in _get_and_verify_max_len(hf_config, max_model_len, disable_sliding_window, sliding_window_len)
1280 if disable_sliding_window:
1281 # TODO(robertgshaw): Find a model that supports rope_scaling
1282 # with sliding window to see if this case should be allowed.
1283 raise NotImplementedError(
1284 "Disabling sliding window is not supported for models "
1285 "with rope_scaling. Please raise an issue so we can "
1286 "investigate.")
-> 1287 assert "factor" in rope_scaling
1288 scaling_factor = rope_scaling["factor"]
1289 if rope_scaling["type"] == "yarn":
AssertionError:
Without trust remote code True :
File ~/.local/lib/python3.11/site-packages/transformers/models/phi3/configuration_phi3.py:185, in Phi3Config._rope_scaling_validation(self)
183 rope_scaling_long_factor = self.rope_scaling.get("long_factor", None)
184 if rope_scaling_type is None or rope_scaling_type not in ["su", "yarn"]:
--> 185 raise ValueError(f"`rope_scaling`'s type field must be one of ['su', 'yarn'], got {rope_scaling_type}")
186 if not (
187 isinstance(rope_scaling_short_factor, list)
188 and all(isinstance(x, (int, float)) for x in rope_scaling_short_factor)
189 ):
190 raise ValueError(
191 f"`rope_scaling`'s short_factor field must be a list of numbers, got {rope_scaling_short_factor}"
192 )
ValueError: `rope_scaling`'s type field must be one of ['su', 'yarn'], got longrope
Please update your version of vLLM to 0.5.2