Edit model card

h2oGPT clone of Meta's Llama 2 13B Chat.

Try it live on our h2oGPT demo with side-by-side LLM comparisons and private document chat!

See how it compares to other models on our LLM Leaderboard!

See more at H2O.ai

Model Architecture

LlamaForCausalLM(
  (model): LlamaModel(
    (embed_tokens): Embedding(32000, 5120, padding_idx=0)
    (layers): ModuleList(
      (0-39): 40 x LlamaDecoderLayer(
        (self_attn): LlamaAttention(
          (q_proj): Linear(in_features=5120, out_features=5120, bias=False)
          (k_proj): Linear(in_features=5120, out_features=5120, bias=False)
          (v_proj): Linear(in_features=5120, out_features=5120, bias=False)
          (o_proj): Linear(in_features=5120, out_features=5120, bias=False)
          (rotary_emb): LlamaRotaryEmbedding()
        )
        (mlp): LlamaMLP(
          (gate_proj): Linear(in_features=5120, out_features=13824, bias=False)
          (up_proj): Linear(in_features=5120, out_features=13824, bias=False)
          (down_proj): Linear(in_features=13824, out_features=5120, bias=False)
          (act_fn): SiLUActivation()
        )
        (input_layernorm): LlamaRMSNorm()
        (post_attention_layernorm): LlamaRMSNorm()
      )
    )
    (norm): LlamaRMSNorm()
  )
  (lm_head): Linear(in_features=5120, out_features=32000, bias=False)
)
Downloads last month
19,646
Safetensors
Model size
13B params
Tensor type
F32
Β·
FP16
Β·
Inference Examples
Inference API (serverless) has been turned off for this model.

Model tree for h2oai/h2ogpt-4096-llama2-13b-chat

Quantizations
1 model

Spaces using h2oai/h2ogpt-4096-llama2-13b-chat 13