Skip to content

Commit

Permalink
Pixtral fails with guided_json etc.
Browse files Browse the repository at this point in the history
  • Loading branch information
pseudotensor committed Sep 12, 2024
1 parent 84d8ec1 commit 193c93a
Show file tree
Hide file tree
Showing 2 changed files with 8 additions and 1 deletion.
7 changes: 7 additions & 0 deletions src/gradio_runner.py
Original file line number Diff line number Diff line change
Expand Up @@ -5798,6 +5798,8 @@ def _get_model_names(local_model_states):
inference_server_split = inference_server.split(':')
inference_server_type = inference_server_split[0].strip() if len(
inference_server_split) > 0 else inference_server
if 'api.together.xyz' in inference_server:
inference_server_type = 'together.ai'
from gradio_utils.grclient import GradioClient
if isinstance(model_state3.get('model', ''), GradioClient):
inference_server_type = 'gradio'
Expand All @@ -5819,6 +5821,11 @@ def _get_model_names(local_model_states):
json_vllm = model_state3.get('json_vllm', False)
model_state3['strict_json_schema'] = get_supports_schema(inference_server, base_model,
json_vllm=json_vllm, just_test=True)
if 'Pixtral' in base_model:
# https://github.com/vllm-project/vllm/issues/8429
model_state3['guided_vllm'] = False
model_state3['strict_json_schema'] = False
model_state3['json_vllm'] = False
key_list = ['display_name', 'base_model', 'inference_server_type',
'strict_json_schema',
'prompt_type', 'prompt_dict', 'chat_template'] + list(
Expand Down
2 changes: 1 addition & 1 deletion src/version.py
Original file line number Diff line number Diff line change
@@ -1 +1 @@
__version__ = "a9415940e581b56ebbbf2b1b32a0e824134ce6a5"
__version__ = "84d8ec119c6b07348535f91af1dc3dc50ea58782"

0 comments on commit 193c93a

Please sign in to comment.