@@ -128,7 +128,7 @@ class TopkTokenModel:
128
128
topk_per_token : int = 5 # number of topk tokens to generate for each token
129
129
generated_answer : str = None # generated answer from model, to display in frontend
130
130
131
- def genearte_topk_per_token (self , text : str ) -> List [Token ]:
131
+ def generate_topk_per_token (self , text : str ) -> List [Token ]:
132
132
raise NotImplementedError
133
133
134
134
def html_to_visualize (self , tokens : List [Token ]) -> str :
@@ -151,7 +151,7 @@ def get_model_tokenizer(self):
151
151
self .rev_vocab = format_reverse_vocab (self .tokenizer )
152
152
return self .model , self .tokenizer
153
153
154
- def genearte_topk_per_token (self , text : str ) -> List [Token ]:
154
+ def generate_topk_per_token (self , text : str ) -> List [Token ]:
155
155
model , tokenizer = self .get_model_tokenizer ()
156
156
rev_vocab = self .rev_vocab
157
157
topk_tokens , topk_probs , sequences = generate_topk_token_prob (
@@ -180,7 +180,7 @@ class TGIModel(TopkTokenModel):
180
180
# tgi support top_n_tokens, reference below:
181
181
# https://github.com/huggingface/text-generation-inference/blob/7dbaf9e9013060af52024ea1a8b361b107b50a69/proto/generate.proto#L108-L109
182
182
183
- def genearte_topk_per_token (self , text : str ) -> List [Token ]:
183
+ def generate_topk_per_token (self , text : str ) -> List [Token ]:
184
184
raise NotImplementedError
185
185
186
186
@@ -201,7 +201,7 @@ def __post_init__(self):
201
201
assert self .api_key is not None , "Please provide api key to access openai api."
202
202
self .client = OpenAI (api_key = self .api_key , base_url = self .base_url )
203
203
204
- def genearte_topk_per_token (self , text : str , ** kwargs ) -> List [Token ]:
204
+ def generate_topk_per_token (self , text : str , ** kwargs ) -> List [Token ]:
205
205
kwargs = {
206
206
"temperature" : self .temperature ,
207
207
"top_p" : self .topp ,
@@ -264,7 +264,7 @@ def openai_api_call(self, payload):
264
264
data = json .loads (response .text )
265
265
return data
266
266
267
- def genearte_topk_per_token (self , text : str , ** kwargs ) -> List [Token ]:
267
+ def generate_topk_per_token (self , text : str , ** kwargs ) -> List [Token ]:
268
268
kwargs = {
269
269
"temperature" : self .temperature ,
270
270
"top_p" : self .topp ,
0 commit comments