Spaces:
Sleeping
Sleeping
| from transformers import BertTokenizer, BertForSequenceClassification | |
| tokenizer = BertTokenizer.from_pretrained('juridics/bertimbaulaw-base-portuguese-sts-scale') | |
| model = BertForSequenceClassification.from_pretrained('juridics/bertimbaulaw-base-portuguese-sts-scale') | |
| def generate_answers(query): | |
| inputs = tokenizer(query, return_tensors='pt', padding='max_length', truncation=True, max_length=512) | |
| attention_mask = inputs['attention_mask'] | |
| input_ids = inputs['input_ids'] | |
| generated_ids = model.generate( | |
| input_ids, | |
| attention_mask=attention_mask, | |
| max_length=len(input_ids[0]) + 100, # Aumentar o limite de geração | |
| temperature=0.7, # Ajustar a criatividade | |
| top_p=0.9, # Usar nucleus sampling | |
| no_repeat_ngram_size=2 # Evitar repetições desnecessárias | |
| ) | |
| generated_text = tokenizer.decode(generated_ids[0], skip_special_tokens=True) | |
| return generated_text |