Skip to content

What is the complete inference process? #4

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Open
zhiyuan5986 opened this issue Feb 27, 2025 · 1 comment
Open

What is the complete inference process? #4

zhiyuan5986 opened this issue Feb 27, 2025 · 1 comment

Comments

@zhiyuan5986
Copy link

Hi, this is a great job!

My question is: In the paper's `Inference' part, you introduce the process to generate C compressed . Then which model are used in the following step (i.e. get the output answer based on query and C_\text{compressed}$)?

Thanks a lot!

@zhiyuan5986 zhiyuan5986 changed the title How is the lantancy evaluation implemented? What is the completed inference process? Feb 27, 2025
@zhiyuan5986 zhiyuan5986 changed the title What is the completed inference process? What is the complete inference process? Feb 27, 2025
@No13Judas
Copy link

Refer to the script mentioned in the README, I guess it's gpt-3.5-turbo, and I get 92% of the longbench avg score which was shown in the paper using the model, also they showed the scores with gpt-4o in the paper.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants