PEFT
bryanchrist commited on
Commit
a4b46ae
·
verified ·
1 Parent(s): 7243ed0

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +16 -6
README.md CHANGED
@@ -28,11 +28,21 @@ The following `bitsandbytes` quantization config was used during training:
28
 
29
  ## Citation
30
  ```bash
31
- @inproceedings{christ_mathwell_2024,
32
- title = {{MATHWELL}: {Generating} {Educational} {Math} {Word} {Problems} {Using} {Teacher} {Annotations}},
33
- url = {https://openreview.net/forum?id=jNsjlRfpk0},
34
- booktitle = {The 2024 {Conference} on {Empirical} {Methods} in {Natural} {Language} {Processing}},
35
- author = {Christ, Bryan R. and Kropko, Jonathan and Hartvigsen, Thomas},
36
- year = {2024},
 
 
 
 
 
 
 
 
 
 
37
  }
38
  ```
 
28
 
29
  ## Citation
30
  ```bash
31
+ @inproceedings{christ-etal-2024-mathwell,
32
+ title = "{MATHWELL}: Generating Educational Math Word Problems Using Teacher Annotations",
33
+ author = "Christ, Bryan R and
34
+ Kropko, Jonathan and
35
+ Hartvigsen, Thomas",
36
+ editor = "Al-Onaizan, Yaser and
37
+ Bansal, Mohit and
38
+ Chen, Yun-Nung",
39
+ booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2024",
40
+ month = nov,
41
+ year = "2024",
42
+ address = "Miami, Florida, USA",
43
+ publisher = "Association for Computational Linguistics",
44
+ url = "https://aclanthology.org/2024.findings-emnlp.696",
45
+ pages = "11914--11938",
46
+ abstract = "Math word problems are critical K-8 educational tools, but writing them is time consuming and requires extensive expertise. To be educational, problems must be solvable, have accurate answers, and, most importantly, be educationally appropriate. We propose that language models have potential to support K-8 math education by automatically generating word problems. However, evaluating educational appropriateness is hard to quantify. We fill this gap by having teachers evaluate problems generated by LLMs, who find existing models and data often fail to be educationally appropriate. We then explore automatically generating *educational* word problems, ultimately using our expert annotations to finetune a 70B language model. Our model, MATHWELL, is the first K-8 word problem generator targeted at educational appropriateness. Further expert studies find MATHWELL generates problems far more solvable, accurate, and appropriate than public models. MATHWELL also matches GPT-4{'}s problem quality while attaining more appropriate reading levels for K-8 students and avoiding generating harmful questions.",
47
  }
48
  ```