Skip to content

Commit 869233f

Browse files
authored
fix: Adding an LLM param to fix broken generator from llamacpp (#1519)
1 parent e326126 commit 869233f

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

private_gpt/components/llm/llm_component.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -42,7 +42,7 @@ def __init__(self, settings: Settings) -> None:
4242
context_window=settings.llm.context_window,
4343
generate_kwargs={},
4444
# All to GPU
45-
model_kwargs={"n_gpu_layers": -1},
45+
model_kwargs={"n_gpu_layers": -1, "offload_kqv": True},
4646
# transform inputs into Llama2 format
4747
messages_to_prompt=prompt_style.messages_to_prompt,
4848
completion_to_prompt=prompt_style.completion_to_prompt,

0 commit comments

Comments
 (0)