Hugging Face
Models
Datasets
Spaces
Community
Docs
Enterprise
Pricing
Log In
Sign Up
bnunticha
/
sentence-tokenizer-th
like
0
Token Classification
Transformers
PyTorch
camembert
Generated from Trainer
Model card
Files
Files and versions
Community
1
Train
Deploy
Use this model
main
sentence-tokenizer-th
Ctrl+K
Ctrl+K
1 contributor
History:
208 commits
bnunticha
End of training
5453845
over 1 year ago
.gitattributes
1.52 kB
initial commit
over 1 year ago
README.md
1.76 kB
End of training
over 1 year ago
added_tokens.json
126 Bytes
Training in progress, step 500
over 1 year ago
config.json
783 Bytes
Training in progress, step 500
over 1 year ago
pytorch_model.bin
419 MB
LFS
End of training
over 1 year ago
sentencepiece.bpe.model
905 kB
LFS
Training in progress, step 500
over 1 year ago
special_tokens_map.json
252 Bytes
Training in progress, step 500
over 1 year ago
tokenizer.json
2.18 MB
Training in progress, step 500
over 1 year ago
tokenizer_config.json
1.79 kB
Training in progress, step 500
over 1 year ago
training_args.bin
4.09 kB
LFS
Training in progress, step 500
over 1 year ago