Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

LM Format Enforcer Guided Decoding Support #3868

Merged
merged 29 commits into from
Apr 16, 2024
Merged
Changes from 1 commit
Commits
Show all changes
29 commits
Select commit Hold shift + click to select a range
ff46b1d
WIP: Refactoring before lm-format-enforcer integration
noamgat Mar 31, 2024
8fa1d0d
Integrated LM Format enforcer decoding
noamgat Apr 5, 2024
d2ef17c
Merge branch 'main' into lmfe-generation
noamgat Apr 5, 2024
0c3ab87
format.sh
noamgat Apr 5, 2024
04c457f
Making pip install work
noamgat Apr 5, 2024
3c2f45b
LMFE integration fixes
noamgat Apr 5, 2024
657a5d5
Ruff error
noamgat Apr 5, 2024
2b797ce
isort fix
noamgat Apr 5, 2024
d637a3e
Fixing test import
noamgat Apr 5, 2024
bb078dc
Merge branch 'main' into lmfe-generation
noamgat Apr 5, 2024
f39ec42
Refactor : introduced DecodingConfig
noamgat Apr 5, 2024
04062b5
Added tests for get_guided_processors wrapper and LMFE implementation
noamgat Apr 5, 2024
11e11be
format.sh
noamgat Apr 5, 2024
ae21046
Fixing test_guided_logits_processor_black_box()
noamgat Apr 5, 2024
13d792d
Merge branch 'main' into lmfe-generation
noamgat Apr 5, 2024
0cfbb1b
Update vllm/config.py based on PR review
noamgat Apr 6, 2024
11cb2db
Update vllm/config.py based on PR review
noamgat Apr 6, 2024
8b9bfc1
format.sh
noamgat Apr 6, 2024
ddb1a8b
Added the option to override guided_decoding_backend on a per-request…
noamgat Apr 6, 2024
dc6c919
Modifying outlines logits processor to return the tokenizer to its no…
noamgat Apr 6, 2024
bee7147
Ruff fixes
noamgat Apr 6, 2024
87bb454
yapf fixes
noamgat Apr 6, 2024
88a200a
Test fixes
noamgat Apr 6, 2024
1baedd3
Ruff fixes
noamgat Apr 6, 2024
2d35e61
yapf
noamgat Apr 6, 2024
041a16b
isort
noamgat Apr 6, 2024
3e5eaa0
Merge branch 'main' into lmfe-generation
noamgat Apr 10, 2024
105c3d0
Doc change for CI retrigger
noamgat Apr 10, 2024
79f060a
Merge branch 'main' of github.com:vllm-project/vllm into lmfe-generation
simon-mo Apr 16, 2024
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Prev Previous commit
Next Next commit
format.sh
  • Loading branch information
noamgat committed Apr 6, 2024
commit 8b9bfc10582c1cfc991f61bd2485036ba6728cbd
38 changes: 20 additions & 18 deletions vllm/config.py
Original file line number Diff line number Diff line change
Expand Up @@ -73,24 +73,26 @@ class ModelConfig:
to eager mode.
"""

def __init__(self,
model: str,
tokenizer: str,
tokenizer_mode: str,
trust_remote_code: bool,
download_dir: Optional[str],
load_format: str,
dtype: Union[str, torch.dtype],
seed: int,
revision: Optional[str] = None,
code_revision: Optional[str] = None,
tokenizer_revision: Optional[str] = None,
max_model_len: Optional[int] = None,
quantization: Optional[str] = None,
quantization_param_path: Optional[str] = None,
enforce_eager: bool = False,
max_context_len_to_capture: Optional[int] = None,
max_logprobs: int = 5,) -> None:
def __init__(
self,
model: str,
tokenizer: str,
tokenizer_mode: str,
trust_remote_code: bool,
download_dir: Optional[str],
load_format: str,
dtype: Union[str, torch.dtype],
seed: int,
revision: Optional[str] = None,
code_revision: Optional[str] = None,
tokenizer_revision: Optional[str] = None,
max_model_len: Optional[int] = None,
quantization: Optional[str] = None,
quantization_param_path: Optional[str] = None,
enforce_eager: bool = False,
max_context_len_to_capture: Optional[int] = None,
max_logprobs: int = 5,
) -> None:
self.model = model
self.tokenizer = tokenizer
self.tokenizer_mode = tokenizer_mode
Expand Down
Loading