Skip to content

Commit

Permalink
llama : remove redundant GQA check (ggerganov#4796)
Browse files Browse the repository at this point in the history
  • Loading branch information
ggerganov authored and jordankanter committed Feb 3, 2024
1 parent 3fe02e0 commit 3b99c26
Showing 1 changed file with 0 additions and 8 deletions.
8 changes: 0 additions & 8 deletions llama.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -4776,7 +4776,6 @@ struct llm_build_context {
const int64_t n_embd_head = hparams.n_embd_head_v;
const int64_t n_embd_gqa = hparams.n_embd_v_gqa();
GGML_ASSERT(n_embd_head == hparams.n_embd_head_k);
GGML_ASSERT(n_embd_gqa == n_embd);

struct ggml_tensor * cur;
struct ggml_tensor * inpL;
Expand Down Expand Up @@ -4900,7 +4899,6 @@ struct llm_build_context {
const int64_t n_embd_head = hparams.n_embd_head_v;
const int64_t n_embd_gqa = hparams.n_embd_v_gqa();
GGML_ASSERT(n_embd_head == hparams.n_embd_head_k);
GGML_ASSERT(n_embd_gqa == n_embd);

struct ggml_tensor * cur;
struct ggml_tensor * pos;
Expand Down Expand Up @@ -5001,7 +4999,6 @@ struct llm_build_context {
const int64_t n_embd_head = hparams.n_embd_head_v;
const int64_t n_embd_gqa = hparams.n_embd_v_gqa();
GGML_ASSERT(n_embd_head == hparams.n_embd_head_k);
GGML_ASSERT(n_embd_gqa == n_embd);

const int64_t n_rot = n_embd_head_k / 2;

Expand Down Expand Up @@ -5215,7 +5212,6 @@ struct llm_build_context {
const int64_t n_embd_head = hparams.n_embd_head_v;
const int64_t n_embd_gqa = hparams.n_embd_v_gqa();
GGML_ASSERT(n_embd_head == hparams.n_embd_head_k);
GGML_ASSERT(n_embd_gqa == n_embd);

struct ggml_tensor * cur;
struct ggml_tensor * inpL;
Expand Down Expand Up @@ -5308,7 +5304,6 @@ struct llm_build_context {
const int64_t n_embd_head = hparams.n_embd_head_v;
const int64_t n_embd_gqa = hparams.n_embd_v_gqa();
GGML_ASSERT(n_embd_head == hparams.n_embd_head_k);
GGML_ASSERT(n_embd_gqa == n_embd);

struct ggml_tensor * cur;
struct ggml_tensor * inpL;
Expand Down Expand Up @@ -5404,7 +5399,6 @@ struct llm_build_context {
const int64_t n_embd_head = hparams.n_embd_head_v;
const int64_t n_embd_gqa = hparams.n_embd_v_gqa();
GGML_ASSERT(n_embd_head == hparams.n_embd_head_k);
GGML_ASSERT(n_embd_gqa == n_embd);

struct ggml_tensor * cur;
struct ggml_tensor * inpL;
Expand Down Expand Up @@ -5731,7 +5725,6 @@ struct llm_build_context {
const int64_t n_embd_head = hparams.n_embd_head_v;
const int64_t n_embd_gqa = hparams.n_embd_v_gqa();
GGML_ASSERT(n_embd_head == hparams.n_embd_head_k);
GGML_ASSERT(n_embd_gqa == n_embd);

struct ggml_tensor * cur;
struct ggml_tensor * attn_norm_output;
Expand Down Expand Up @@ -5955,7 +5948,6 @@ struct llm_build_context {
const int64_t n_embd_head = hparams.n_embd_head_v;
const int64_t n_embd_gqa = hparams.n_embd_v_gqa();
GGML_ASSERT(n_embd_head == hparams.n_embd_head_k);
GGML_ASSERT(n_embd_gqa == n_embd);

struct ggml_tensor * cur;
struct ggml_tensor * pos;
Expand Down

0 comments on commit 3b99c26

Please sign in to comment.