Hugging Face's logo Hugging Face
  • Models
  • Datasets
  • Spaces
  • Posts
  • Docs
  • Enterprise
  • Pricing

  • Log In
  • Sign Up

1bitLLM
/
bitnet_b1_58-3B

Text Generation
Transformers
Safetensors
llama
text-generation-inference
Model card Files Files and versions Community
12
New discussion
Resources
  • PR & discussions documentation
  • Code of Conduct
  • Hub documentation

Add metadata and link to paper

#12 opened about 1 month ago by
nielsr

can you provide wikitest ppl and c4 ppl separately?

#11 opened 12 months ago by
sheropen-2

Can you provide more details on the training?

1
#10 opened 12 months ago by
dequ777

Any plans to use MQA (multi-query attention) or GQA (grouped-query attention) in the future?

#9 opened about 1 year ago by
graefics

Efficient Inference Kernel Support for 1.58bit.

9
#8 opened about 1 year ago by
LeiWang1999

This code from BitLinear doesn't make sense

1
#7 opened about 1 year ago by
qmsoqm

Is it bitnet {-1,0,1}?

4
#6 opened about 1 year ago by
Remek

ValueError: Tokenizer class BitnetTokenizer does not exist or is not currently imported.

2
4
#5 opened about 1 year ago by
RZJournal

Longer inference time

2
#4 opened about 1 year ago by
dittops

Why are these models fp32?

2
5
#2 opened about 1 year ago by
supercharge19

Is there a chat/instruct model in plans?

2
#1 opened about 1 year ago by
MrVodnik
Company
TOS Privacy About Jobs
Website
Models Datasets Spaces Pricing Docs