Skip to content

Commit d1339c9

Browse files
committed
Run factuality benchmark with low temp
1 parent bda1b1a commit d1339c9

File tree

4 files changed

+7
-1
lines changed

4 files changed

+7
-1
lines changed

fact.py

+5-1
Original file line numberDiff line numberDiff line change
@@ -44,7 +44,11 @@ def main():
4444
for swap in (False, True):
4545
prompt = format_prompt(task, swap)
4646
print(f'{i}. {prompt}')
47-
output = llm.create_completion(prompt, max_tokens=2)
47+
output = llm.create_completion(
48+
prompt,
49+
max_tokens=20,
50+
temperature=1e-6,
51+
)
4852
answer = output['choices'][0]['text'].strip().split()[0]
4953
print(answer)
5054
result = dict(task, answer=answer, swap=swap)

requirements-llama-2.txt

+1
Original file line numberDiff line numberDiff line change
@@ -0,0 +1 @@
1+
llama-cpp-python==0.1.77 # Works with Llama 2 GGML

requirements-wizard.txt

+1
Original file line numberDiff line numberDiff line change
@@ -0,0 +1 @@
1+
llama-cpp-python==0.1.83 # Works with Wizard GGUF
File renamed without changes.

0 commit comments

Comments
 (0)