Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
fix: attention implementation initialization (#9)
when initializing the model with no explicit declare which attention implementation to use, the original implementation will throw an error. This is because the llama init function will change the attn implementation to sdpa, which is not implemented in lckv yet. We fix it by passing a copy of the config to the llama init function.
- Loading branch information