Model Card for AnkiGPT-small
Model Details
Model Description
- Developed by: anktechsol.com
- Shared by: [anktechsol]
- Model type: Causal Language Model
- Language(s) (NLP): English, potentially aspects of Indian languages/Hinglish due to fine-tuning data.
- License: (Specify the license of the fine-tuned model, often inherited from the base model or dataset. DialoGPT uses MIT license, check the dataset license.)
- Finetuned from model:
microsoft/DialoGPT-small
Model Sources
- Repository:
https://huggingface.co/anktechsol/ankiGPT-small
(This will be the link after pushing to the hub)
Uses
Direct Use
This model can be used for text generation, particularly in conversational or narrative contexts, with a potential bias towards topics and linguistic styles present in the fine-tuning dataset (Indian conversational data).
Downstream Use
This model could potentially be used as a base for further fine-tuning on more specific Indian language tasks or domains.
Out-of-Scope Use
This model is not suitable for generating factual information, performing critical tasks requiring high accuracy, or deployment in sensitive applications without extensive further evaluation and mitigation of potential biases.
Bias, Risks, and Limitations
Based on initial testing, the model may exhibit repetitive text generation, especially for longer sequences. The model's knowledge and linguistic style are heavily influenced by the fine-tuning dataset, which may not cover all aspects of Indian languages or culture comprehensively. Biases present in the training data may be reflected in the model's output.
Recommendations
Users should be aware of the model's limitations in generating coherent long text and potential biases. It is recommended to experiment with different generation parameters (max_length
, no_repeat_ngram_size
, sampling strategies) to improve output quality. For any critical applications, thorough testing and human review of generated content are essential.
How to Get Started with the Model
Use the code below to get started with the model using the transformers library.
from transformers import pipeline
generator = pipeline("text-generation", model="anktechsol/ankiGPT-small")
prompt = "Write a short story about a day in the life of a student in a bustling Indian city, describing their commute, interactions at school, and a cultural event they attend in the evening."
generated_text = generator(prompt, max_length=300, num_return_sequences=1)
print(generated_text[0]['generated_text'])
- Downloads last month
- 37
Model tree for anktechsol/ankiGPT-small
Base model
microsoft/DialoGPT-small