abdalraheemdmd commited on
Commit
9475b0d
·
verified ·
1 Parent(s): 09dbf6d

Update story_generator.py

Browse files
Files changed (1) hide show
  1. story_generator.py +8 -8
story_generator.py CHANGED
@@ -29,12 +29,12 @@ def generate_story(theme, reading_level, max_new_tokens=400, temperature=0.7):
29
  with torch.no_grad():
30
  output = story_model.generate(
31
  input_ids,
32
- max_new_tokens=max_new_tokens, # ⏩ Reduced to 250 for speed (previously 700)
33
  temperature=temperature,
34
- top_k=20, # ⏩ Lowered from 50 → More focused, fewer unnecessary words
35
- top_p=0.7, # ⏩ Reduced randomness (previously 0.95)
36
  do_sample=True,
37
- early_stopping=True, # ⏩ Stops at logical sentence breaks
38
  pad_token_id=story_tokenizer.pad_token_id,
39
  eos_token_id=story_tokenizer.eos_token_id,
40
  attention_mask=input_ids.ne(story_tokenizer.pad_token_id)
@@ -50,12 +50,12 @@ def generate_questions(story, max_new_tokens=150, temperature=0.7):
50
  with torch.no_grad():
51
  output = question_model.generate(
52
  input_ids,
53
- max_new_tokens=max_new_tokens, # ⏩ Reduced to 100 for speed (previously 300)
54
  temperature=temperature,
55
- top_k=20, # ⏩ More focused question generation
56
- top_p=0.7, # ⏩ Less randomness
57
  do_sample=True,
58
- early_stopping=True, # ⏩ Stops at logical breakpoints
59
  pad_token_id=question_tokenizer.pad_token_id,
60
  eos_token_id=question_tokenizer.eos_token_id,
61
  attention_mask=input_ids.ne(question_tokenizer.pad_token_id)
 
29
  with torch.no_grad():
30
  output = story_model.generate(
31
  input_ids,
32
+ max_new_tokens=max_new_tokens,
33
  temperature=temperature,
34
+ top_k=20,
35
+ top_p=0.7,
36
  do_sample=True,
37
+ early_stopping=True,
38
  pad_token_id=story_tokenizer.pad_token_id,
39
  eos_token_id=story_tokenizer.eos_token_id,
40
  attention_mask=input_ids.ne(story_tokenizer.pad_token_id)
 
50
  with torch.no_grad():
51
  output = question_model.generate(
52
  input_ids,
53
+ max_new_tokens=max_new_tokens,
54
  temperature=temperature,
55
+ top_k=20,
56
+ top_p=0.7,
57
  do_sample=True,
58
+ early_stopping=True,
59
  pad_token_id=question_tokenizer.pad_token_id,
60
  eos_token_id=question_tokenizer.eos_token_id,
61
  attention_mask=input_ids.ne(question_tokenizer.pad_token_id)