MISHANM/meta-Llama-3.1-8B-Instruct.gguf

This model is a GGUF version of the meta-llama/Llama-3.1-8B-Instruct model, optimized for use with the llama.cpp framework. It is designed to run efficiently on CPUs and can be used for various natural language processing tasks.

Model Details

  1. Language: English
  2. Tasks: Text generation
  3. Base Model:meta-llama/Llama-3.1-8B-Instruct

Building and Running the Model

To build and run the model using llama.cpp, follow these steps:

Build llama.cpp Locally

git clone https://github.com/ggerganov/llama.cpp  
cd llama.cpp  
cmake -B build  
cmake --build build --config Release  

Run the Model

Navigate to the build directory and run the model with a prompt:

cd llama.cpp/build/bin   

Inference with llama.cpp

./llama-cli -m /path/to/model/ -p "Your prompt here" -n 128  

Citation Information

@misc{MISHANM/meta-Llama-3.1-8B-Instruct.gguf,
  author = {Mishan Maurya},
  title = {Introducing meta-llama/Llama-3.1-8B-Instruct GGUF Model},
  year = {2025},
  publisher = {Hugging Face},
  journal = {Hugging Face repository},
  
}
Downloads last month
-
GGUF
Model size
8.03B params
Architecture
llama
Hardware compatibility
Log In to view the estimation

We're not able to determine the quantization variants.

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for MISHANM/meta-Llama-3.1-8B-Instruct.gguf

Quantized
(493)
this model