@@ -78,6 +78,8 @@ def __call__(
78
78
mirostat_eta : float = 0.1 ,
79
79
logits_processor : Optional [llama .LogitsProcessorList ] = None ,
80
80
grammar : Optional [llama .LlamaGrammar ] = None ,
81
+ logprobs : Optional [bool ] = None ,
82
+ top_logprobs : Optional [int ] = None ,
81
83
** kwargs , # type: ignore
82
84
) -> Union [
83
85
llama_types .CreateChatCompletionResponse ,
@@ -339,7 +341,7 @@ def _convert_completion_to_chat_function(
339
341
}
340
342
],
341
343
},
342
- "logprobs" : None ,
344
+ "logprobs" : completion [ "choices" ][ 0 ][ "logprobs" ] ,
343
345
"finish_reason" : "tool_calls" ,
344
346
}
345
347
],
@@ -392,7 +394,7 @@ def _stream_response_to_function_stream(
392
394
{
393
395
"index" : 0 ,
394
396
"finish_reason" : None ,
395
- "logprobs" : None ,
397
+ "logprobs" : chunk [ "choices" ][ 0 ][ "logprobs" ] ,
396
398
"delta" : {
397
399
"role" : None ,
398
400
"content" : None ,
@@ -427,7 +429,7 @@ def _stream_response_to_function_stream(
427
429
{
428
430
"index" : 0 ,
429
431
"finish_reason" : None ,
430
- "logprobs" : None ,
432
+ "logprobs" : chunk [ "choices" ][ 0 ][ "logprobs" ] ,
431
433
"delta" : {
432
434
"role" : None ,
433
435
"content" : None ,
@@ -492,7 +494,6 @@ def chat_completion_handler(
492
494
temperature : float = 0.2 ,
493
495
top_p : float = 0.95 ,
494
496
top_k : int = 40 ,
495
- logprobs : int = 0 ,
496
497
min_p : float = 0.05 ,
497
498
typical_p : float = 1.0 ,
498
499
stream : bool = False ,
@@ -514,6 +515,8 @@ def chat_completion_handler(
514
515
logits_processor : Optional [llama .LogitsProcessorList ] = None ,
515
516
grammar : Optional [llama .LlamaGrammar ] = None ,
516
517
logit_bias : Optional [Dict [str , float ]] = None ,
518
+ logprobs : Optional [bool ] = None ,
519
+ top_logprobs : Optional [int ] = None ,
517
520
** kwargs , # type: ignore
518
521
) -> Union [
519
522
llama_types .CreateChatCompletionResponse ,
@@ -583,7 +586,7 @@ def chat_completion_handler(
583
586
top_k = top_k ,
584
587
min_p = min_p ,
585
588
typical_p = typical_p ,
586
- logprobs = logprobs ,
589
+ logprobs = top_logprobs if logprobs else None ,
587
590
stream = stream ,
588
591
stop = stop ,
589
592
seed = seed ,
@@ -1634,7 +1637,7 @@ def message_to_str(msg: llama_types.ChatCompletionRequestMessage):
1634
1637
}
1635
1638
],
1636
1639
},
1637
- "logprobs" : None ,
1640
+ "logprobs" : completion [ "choices" ][ 0 ][ "logprobs" ] ,
1638
1641
"finish_reason" : "tool_calls" ,
1639
1642
}
1640
1643
],
@@ -2094,7 +2097,7 @@ def create_completion(stop):
2094
2097
choices = [
2095
2098
{
2096
2099
"index" : 0 ,
2097
- "logprobs" : None ,
2100
+ "logprobs" : completion [ "choices" ][ 0 ][ "logprobs" ] ,
2098
2101
"message" : {
2099
2102
"role" : "assistant" ,
2100
2103
"content" : None if content == "" else content ,
@@ -2323,11 +2326,14 @@ def chatml_function_calling(
2323
2326
model : Optional [str ] = None ,
2324
2327
logits_processor : Optional [llama .LogitsProcessorList ] = None ,
2325
2328
grammar : Optional [llama .LlamaGrammar ] = None ,
2329
+ logprobs : Optional [bool ] = None ,
2330
+ top_logprobs : Optional [int ] = None ,
2326
2331
** kwargs , # type: ignore
2327
2332
) -> Union [
2328
2333
llama_types .CreateChatCompletionResponse ,
2329
2334
Iterator [llama_types .CreateChatCompletionStreamResponse ],
2330
2335
]:
2336
+ print (logprobs )
2331
2337
function_calling_template = (
2332
2338
"{% for message in messages %}"
2333
2339
"<|im_start|>{{ message.role }}\n "
@@ -2450,6 +2456,7 @@ def chatml_function_calling(
2450
2456
model = model ,
2451
2457
logits_processor = logits_processor ,
2452
2458
grammar = grammar ,
2459
+ logprobs = top_logprobs if logprobs else None ,
2453
2460
),
2454
2461
stream = stream ,
2455
2462
)
@@ -2564,6 +2571,7 @@ def chatml_function_calling(
2564
2571
typical_p = typical_p ,
2565
2572
stream = stream ,
2566
2573
stop = ["<|im_end|>" ],
2574
+ logprobs = top_logprobs if logprobs else None ,
2567
2575
max_tokens = None ,
2568
2576
min_tokens = min_tokens ,
2569
2577
presence_penalty = presence_penalty ,
@@ -2678,7 +2686,7 @@ def chatml_function_calling(
2678
2686
{
2679
2687
"finish_reason" : "tool_calls" ,
2680
2688
"index" : 0 ,
2681
- "logprobs" : None ,
2689
+ "logprobs" : completion [ "choices" ][ 0 ][ "logprobs" ] ,
2682
2690
"message" : {
2683
2691
"role" : "assistant" ,
2684
2692
"content" : None ,
0 commit comments