Replies: 1 comment 2 replies
-
As far as I understood: Gatt is a special algorithm that was used when LLaMA 2 Chat was fine tuned. The GAtt is not something that requires a change in how inference is done. |
Beta Was this translation helpful? Give feedback.
2 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Chapter 3.3 of "Llama 2: Open Foundation and Fine-Tuned Chat Models" talks about using GAtt to avoid having to repeat instructions in multi-turn dialogue. Is this something that would need to be implemented in llama.cpp, or does it already exist?
Beta Was this translation helpful? Give feedback.
All reactions