Skip to content

Commit aa653e3

Browse files
committed
Prevent llama.cpp from being monkey patched more than once (closes oobabooga#6201)
1 parent a210e61 commit aa653e3

File tree

1 file changed

+6
-1
lines changed

1 file changed

+6
-1
lines changed

modules/llama_cpp_python_hijack.py

+6-1
Original file line numberDiff line numberDiff line change
@@ -100,9 +100,11 @@ def eval_with_progress(self, tokens: Sequence[int]):
100100

101101

102102
def monkey_patch_llama_cpp_python(lib):
103+
if getattr(lib.Llama, '_is_patched', False):
104+
# If the patch is already applied, do nothing
105+
return
103106

104107
def my_generate(self, *args, **kwargs):
105-
106108
if shared.args.streaming_llm:
107109
new_sequence = args[0]
108110
past_sequence = self._input_ids
@@ -116,3 +118,6 @@ def my_generate(self, *args, **kwargs):
116118
lib.Llama.eval = eval_with_progress
117119
lib.Llama.original_generate = lib.Llama.generate
118120
lib.Llama.generate = my_generate
121+
122+
# Set the flag to indicate that the patch has been applied
123+
lib.Llama._is_patched = True

0 commit comments

Comments
 (0)