Skip to content

Conversation

@ggerganov
Copy link
Member

@ggerganov ggerganov commented May 26, 2024

Certain combinations of [EXPERIMENTAL] llama_context_params are not always supported:

    struct llama_context_params {
        ...

        enum ggml_type type_k; // data type for K cache [EXPERIMENTAL]
        enum ggml_type type_v; // data type for V cache [EXPERIMENTAL]

        bool flash_attn;  // whether to use flash attention [EXPERIMENTAL]

        ...
    };

Here is a list of known incompatibilities (we can try to update it in the future):

@ggerganov ggerganov merged commit eaf6e03 into master May 27, 2024
@ggerganov ggerganov deleted the gg/fattn-warn branch May 27, 2024 06:24
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants