NotImplementedError: Could not run 'aten::_local_scalar_dense' with arguments from the 'Meta' backend.
➕
2
3
#54 opened about 1 year ago
by
duccio84
Some of you might be interested in my 'silly' experiment.
🧠
1
2
#52 opened about 1 year ago
by
ZeroWw
Updated config.json
#51 opened about 1 year ago
by
WestM
🚀 LMDeploy support Llama3.1 and its Tool Calling. An example of calling "Wolfram Alpha" to perform complex mathematical calculations can be found from here!
#50 opened about 1 year ago
by
vansin

HF pro subscription for llama 3.1-8b
4
#49 opened about 1 year ago
by
ostoslista
Significant bias
👀
👍
3
6
#48 opened about 1 year ago
by
stutteringp0et
`rope_scaling` must be a dictionary with two fields
➕
🤝
8
4
#46 opened about 1 year ago
by
thunderdagger
Unable to load Llama 3.1 to Text-Genration WebUI
4
#45 opened about 1 year ago
by
keeeeesz
BUG Chat template doesn't respect `add_generation_prompt`flag from transformers tokenizer
👍
3
1
#44 opened about 1 year ago
by
ilu000

How to use the ASR on LLama3.1
🔥
👀
1
1
#43 opened about 1 year ago
by
andrygasy
Tokenizer 'apply_chat_template' issue
1
#42 opened about 1 year ago
by
Ksgk-fy

Function Calling Evaluation bench Nexus (0-shot)
#41 opened about 1 year ago
by
WateBear
Error: json: cannot unmarshal array into Go struct field Params.eos_token_id of type int
❤️
1
2
#40 opened about 1 year ago
by
SadeghPouriyanZadeh

ValueError: Pipeline with tokenizer without pad_token cannot do batching. You can try to set it with `pipe.tokenizer.pad_token_id = model.config.eos_token_id`.
👍
👀
2
4
#39 opened about 1 year ago
by
jsemrau

Run this on CPU and use tool calling
1
#38 opened about 1 year ago
by
J22
!!Access Problem
➕
15
13
#37 opened about 1 year ago
by
minglingfeng
LLama-3.1-8B generates way to long answers!
👍
1
3
#36 opened about 1 year ago
by
ayyylemao
Tokenizer error and/or 'rope_scaling' problem
5
#35 opened about 1 year ago
by
fazayjo
Deployment to Inference Endpoints
➕
🔥
3
6
#34 opened about 1 year ago
by
stcat
Best practice for tool calling with meta-llama/Meta-Llama-3.1-8B-Instruct
1
#33 opened about 1 year ago
by
zzclynn
The model often enters infinite generation loops
👍
5
13
#32 opened about 1 year ago
by
sszymczyk
unable to load 4-bit quantized varient with llama.cpp
#31 opened about 1 year ago
by
sunnykusawa

Garbage output ?
10
#30 opened about 1 year ago
by
danielus

Question about chat template and fine-tuning
👍
9
4
#23 opened about 1 year ago
by
tblattner
Issues loading model with ooabooga textgenwebui
👍
3
5
#20 opened about 1 year ago
by
Kenji776
what is the right tokenizer should I use for llama 3.1 8B?
3
#19 opened about 1 year ago
by
calebl
The sample code on the model card page is not right
#18 opened about 1 year ago
by
kmtao
My alternative quantizations.
7
#16 opened about 1 year ago
by
ZeroWw
ValueError: `rope_scaling` must be a dictionary with two fields
🔥
➕
106
52
#15 opened about 1 year ago
by
jsemrau

Independently Benchmarked Humaneval and Evalplus scores
2
#13 opened about 1 year ago
by
VaibhavSahai
DO NOT MERGE v2 make sure vllm and transformers work
#12 opened about 1 year ago
by
ArthurZ

DO NOT MERGE test for vllm
2
#11 opened about 1 year ago
by
ArthurZ

Please do not include original PTH files.
👍
❤️
5
4
#10 opened about 1 year ago
by
Qubitium

Utterly based
😎
🚀
8
1
#9 opened about 1 year ago
by
llama-anon
