Triangle104's picture
Update README.md
d0410cf verified
---
base_model: ruggsea/Llama3.1-8B-SEP-Chat
datasets:
- ruggsea/stanford-encyclopedia-of-philosophy_chat_multi_turn
language:
- en
- it
license: other
tags:
- llama-cpp
- gguf-my-repo
---
# Triangle104/Llama3.1-8B-SEP-Chat-Q4_K_M-GGUF
This model was converted to GGUF format from [`ruggsea/Llama3.1-8B-SEP-Chat`](https://huggingface.co/ruggsea/Llama3.1-8B-SEP-Chat) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
Refer to the [original model card](https://huggingface.co/ruggsea/Llama3.1-8B-SEP-Chat) for more details on the model.
---
Model details:
-
This model is a LoRA finetune of meta-llama/Meta-Llama-3.1-8B trained on multi-turn philosophical conversations. It is designed to engage in philosophical discussions in a conversational yet rigorous manner, maintaining academic standards while being accessible.
Model description
The model was trained using the TRL (Transformer Reinforcement Learning) library's chat template, enabling it to handle multi-turn conversations in a natural way. It builds upon the capabilities of its predecessor Llama3-stanford-encyclopedia-philosophy-QA but extends it to handle more interactive, back-and-forth philosophical discussions.
Chat Format
The model uses the standard chat format with roles:
<|system|>
{{system_prompt}}
<|user|>
{{user_message}}
<|assistant|>
{{assistant_response}}
Training Details
The model was trained with the following system prompt:
You are an expert and informative yet accessible Philosophy university professor. Students will engage with you in philosophical discussions. Respond to their questions and comments in a correct and rigorous but accessible way, maintaining academic standards while fostering understanding.
Training hyperparameters
The following hyperparameters were used during training:
Learning rate: 2e-5
Train batch size: 1
Gradient accumulation steps: 4
Effective batch size: 4
Optimizer: paged_adamw_8bit
LR scheduler: cosine with warmup
Warmup ratio: 0.03
Training epochs: 5
LoRA config:
r: 256
alpha: 128
Target modules: all-linear
Dropout: 0.05
Framework versions
PEFT 0.10.0
Transformers 4.40.1
PyTorch 2.2.2+cu121
TRL latest
Datasets 2.19.0
Tokenizers 0.19.1
Intended Use
This model is designed for:
Multi-turn philosophical discussions
Academic philosophical inquiry
Teaching and learning philosophy
Exploring philosophical concepts through dialogue
Limitations
The model should not be used as a substitute for professional philosophical advice or formal philosophical education
While the model aims to be accurate, its responses should be verified against authoritative sources
The model may occasionally generate plausible-sounding but incorrect philosophical arguments
As with all language models, it may exhibit biases present in its training data
License
This model is subject to the Meta Llama 2 license agreement. Please refer to Meta's licensing terms for usage requirements and restrictions.
How to use
Here's an example of how to use the model:
from transformers import AutoModelForCausalLM, AutoTokenizer
# Load model and tokenizer
model = AutoModelForCausalLM.from_pretrained("ruggsea/Llama3.1-SEP-Chat")
tokenizer = AutoTokenizer.from_pretrained("ruggsea/Llama3.1-SEP-Chat")
# Example conversation
messages = [
{"role": "user", "content": "What is the difference between ethics and morality?"}
]
# Format prompt using chat template
prompt = tokenizer.apply_chat_template(
messages,
add_generation_prompt=True,
tokenize=False
)
# Generate response
inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
outputs = model.generate(**inputs, max_new_tokens=512)
response = tokenizer.decode(outputs[0], skip_special_tokens=True)
---
## Use with llama.cpp
Install llama.cpp through brew (works on Mac and Linux)
```bash
brew install llama.cpp
```
Invoke the llama.cpp server or the CLI.
### CLI:
```bash
llama-cli --hf-repo Triangle104/Llama3.1-8B-SEP-Chat-Q4_K_M-GGUF --hf-file llama3.1-8b-sep-chat-q4_k_m.gguf -p "The meaning to life and the universe is"
```
### Server:
```bash
llama-server --hf-repo Triangle104/Llama3.1-8B-SEP-Chat-Q4_K_M-GGUF --hf-file llama3.1-8b-sep-chat-q4_k_m.gguf -c 2048
```
Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
Step 1: Clone llama.cpp from GitHub.
```
git clone https://github.com/ggerganov/llama.cpp
```
Step 2: Move into the llama.cpp folder and build it with `LLAMA_CURL=1` flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux).
```
cd llama.cpp && LLAMA_CURL=1 make
```
Step 3: Run inference through the main binary.
```
./llama-cli --hf-repo Triangle104/Llama3.1-8B-SEP-Chat-Q4_K_M-GGUF --hf-file llama3.1-8b-sep-chat-q4_k_m.gguf -p "The meaning to life and the universe is"
```
or
```
./llama-server --hf-repo Triangle104/Llama3.1-8B-SEP-Chat-Q4_K_M-GGUF --hf-file llama3.1-8b-sep-chat-q4_k_m.gguf -c 2048
```