Text Generation
GGUF
English
conversational
xzuyn commited on
Commit
fc7eb81
·
verified ·
1 Parent(s): 0cf28f1

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +429 -0
README.md ADDED
@@ -0,0 +1,429 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: PJMixers-Dev/Granite-3.1-Earthen-v0.3-1B-A400M
3
+ license: apache-2.0
4
+ pipeline_tag: text-generation
5
+ language:
6
+ - en
7
+ datasets:
8
+ - BeaverAI/REDACTED1
9
+ - BeaverAI/REDACTED2
10
+ - BeaverAI/REDACTED3
11
+ - BeaverAI/REDACTED4
12
+ - BeaverAI/REDACTED5
13
+ - BeaverAI/REDACTED6
14
+ - PJMixers-Dev/Lit-axo-Shuffled
15
+ - PJMixers-Dev/Mielikki_Erebus-87k-axo
16
+ - PJMixers/RyokoAI_Honeyfeed3600-Cleanish
17
+ - PJMixers-Dev/allura-org_fujin-cleaned-stage-2-axo
18
+ - Nelathan/synthetic-sugar-quill
19
+ - PJMixers-Dev/winglian_visual-novels-json-axo-dropped-long
20
+ - PJMixers-Dev/recursal_SCP-RECURSAL-Cleaned
21
+ - PJMixers-Dev/Subtitles
22
+ - PJMixers-Dev/KaraKaraWitch_AnimeSubtitle-axo
23
+ - PJMixers/AP-News-2024
24
+ - PJMixers-Dev/Fundus-AP-News-Formatted
25
+ - PJMixers-Dev/Fundus-AP-News-2-Formatted
26
+ - PJMixers-Dev/goodwiki-2024-12-04-axo
27
+ - epfl-llm/guidelines
28
+ - PJMixers-Dev/allenai_tulu-3-sft-mixture-filtered-2-ShareGPT
29
+ - OpenLeecher/lmsys_chat_1m_clean
30
+ - PJMixers-Dev/Gryphe-Aesir-RPG-Charcards-Opus-Mixed
31
+ - allura-org/gryphe-sonnet-3.5-charcards-names-added
32
+ - anthracite-org/c2_logs_32k_llama3_qwen2_v1.3
33
+ - PJMixers-Dev/MinervaAI_Aesir-Preview-Anon
34
+ - PJMixers-Dev/lemonilia_LimaRP-Simple-CustomShareGPT-Shuffled
35
+ - Epiculous/SynthRP-Gens-v1.1-Filtered-n-Cleaned
36
+ - PJMixers-Dev/NyxKrage_chub-logs-sharegpt-longest-CustomShareGPT
37
+ - PJMixers/OpenLeecher_Teatime_all_logs_longest-ShareGPT
38
+ - grimulkan/aicg-logs-augmented
39
+ - grimulkan/PIPPA-augmented-dedup
40
+ - PJMixers/grimulkan_bluemoon_Karen_cleaned-carded-formatted
41
+ - PJMixers/lodrick-the-lafted_OpusStories-ShareGPT
42
+ - Gryphe/ChatGPT-4o-Writing-Prompts
43
+ - Gryphe/Opus-WritingPrompts
44
+ - anthracite-org/nopm_claude_writing_fixed
45
+ - PJMixers-Dev/Tiefighter-13B-Fake-Distill-ShareGPT
46
+ - allura-org/fujin-instruct-v2
47
+ - ToastyPigeon/gutenberg-sft
48
+ - PocketDoc/Dans-Prosemaxx-Adventure
49
+ - PocketDoc/Dans-Failuremaxx-Adventure-3
50
+ - TheDrummer/AmoralQA-v2
51
+ ---
52
+ # Granite-3.1-Earthen-v0.3-1B-A400M-GGUF
53
+
54
+ [`ibm-granite/granite-3.1-1b-a400m-instruct`](https://huggingface.co/ibm-granite/granite-3.1-1b-a400m-instruct) was trained at 8K with batch size 4 gradient accumulation 4, so each step was 131,072 tokens (including any padding tokens). It was trained for 560 steps, adding up to a total of 73,400,320 unique tokens seen.
55
+
56
+ This is a small test run. A larger version is planned.
57
+
58
+ ## Quants
59
+
60
+ - [GGUF](https://huggingface.co/PJMixers-Dev/Granite-3.1-Earthen-v0.3-1B-A400M-GGUF)
61
+
62
+ ## Prompt Format
63
+
64
+ This model uses Granite-3.1 Instruct format.
65
+
66
+ ```
67
+ <|start_of_role|>system<|end_of_role|>example system prompt<|end_of_text|>
68
+ <|start_of_role|>user<|end_of_role|>example user turn 1<|end_of_text|>
69
+ <|start_of_role|>assistant<|end_of_role|>example assistant turn 1<|end_of_text|>
70
+ <|start_of_role|>user<|end_of_role|>example user turn 2<|end_of_text|>
71
+ <|start_of_role|>assistant<|end_of_role|>example assistant turn 2<|end_of_text|>
72
+ ```
73
+
74
+ ## Training Details
75
+
76
+ [<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl)
77
+
78
+ ```yaml
79
+ # Requirements before running
80
+ # - Get latest commit of axolotl (currently c0a0c75)
81
+ # - Download these to axolotl/src/axolotl/prompt_formatters
82
+ # - https://github.com/xzuyn/axolotl/blob/came-plus-formatters/src/axolotl/prompt_strategies/formatter_regex.py
83
+ # - https://github.com/xzuyn/axolotl/blob/came-plus-formatters/src/axolotl/prompt_strategies/customcompletion-regex.py
84
+ # - https://github.com/xzuyn/axolotl/blob/came-plus-formatters/src/axolotl/prompt_strategies/customgranite-regex.py
85
+ # - pip install ftfy
86
+ # - pip install git+https://github.com/xzuyn/CAME.git@sr-grams-cautious-8bit
87
+
88
+ # Weights and Biases logging config
89
+ wandb_project: Granite-3.1-1B-A400M
90
+ wandb_name: Granite-3.1-Earthen-v0.3-1B-A400M-QLoRA-run1
91
+
92
+ # Model checkpointing config
93
+ output_dir: ./Outputs/Granite-3.1-Earthen-v0.3-1B-A400M-QLoRA-run1
94
+ resume_from_checkpoint:
95
+ save_steps: 10
96
+ save_safetensors: true
97
+ save_total_limit: 2
98
+ save_only_model: false
99
+
100
+ # Model architecture config
101
+ base_model: ibm-granite/granite-3.1-1b-a400m-instruct
102
+ model_type: AutoModelForCausalLM
103
+ tokenizer_type: AutoTokenizer
104
+
105
+ # Mixed precision training config
106
+ bf16: true
107
+ fp16: false
108
+ tf32: false
109
+
110
+ # Model loading config
111
+ load_in_8bit: false
112
+ load_in_4bit: true
113
+ strict: false
114
+
115
+ # Sequence config
116
+ sequence_len: 8192
117
+ min_sample_len: 256
118
+ sample_packing: true
119
+ eval_sample_packing: true
120
+ pad_to_sequence_len: true
121
+ train_on_inputs: false
122
+ group_by_length: false
123
+
124
+ # LoRA adapter config
125
+ adapter: qlora
126
+ lora_r: 128
127
+ lora_alpha: 128
128
+ lora_dropout: 0.125
129
+ lora_target_linear: true
130
+ embeddings_skip_upcast: true
131
+
132
+ # Dataset config
133
+ datasets:
134
+ # Completion
135
+ # Story-like Data
136
+ - path: BeaverAI/REDACTED1
137
+ split: train[:4000]
138
+ type: customcompletion-regex
139
+ - path: PJMixers-Dev/Lit-axo-Shuffled
140
+ split: train[:4000]
141
+ type: customcompletion-regex
142
+ - path: PJMixers-Dev/Mielikki_Erebus-87k-axo
143
+ split: train[:4000]
144
+ type: customcompletion-regex
145
+ - path: PJMixers/RyokoAI_Honeyfeed3600-Cleanish
146
+ split: train[:4000]
147
+ type: customcompletion-regex
148
+ - path: BeaverAI/REDACTED2
149
+ type: customcompletion-regex
150
+ - path: PJMixers-Dev/allura-org_fujin-cleaned-stage-2-axo
151
+ split: train[:4000]
152
+ type: customcompletion-regex
153
+ - path: Nelathan/synthetic-sugar-quill
154
+ split: train[:4000]
155
+ type: customcompletion-regex
156
+ - path: PJMixers-Dev/winglian_visual-novels-json-axo-dropped-long
157
+ split: train[:4000]
158
+ type: customcompletion-regex
159
+ - path: BeaverAI/REDACTED3
160
+ type: customcompletion-regex
161
+ - path: PJMixers-Dev/recursal_SCP-RECURSAL-Cleaned
162
+ split: train[:4000]
163
+ type: customcompletion-regex
164
+ # Subtitle Data
165
+ - path: PJMixers-Dev/Subtitles
166
+ type: customcompletion-regex
167
+ - path: PJMixers-Dev/KaraKaraWitch_AnimeSubtitle-axo
168
+ split: train[:4000]
169
+ type: customcompletion-regex
170
+ # News Data
171
+ - path: PJMixers/AP-News-2024
172
+ type: customcompletion-regex
173
+ - path: PJMixers-Dev/Fundus-AP-News-Formatted
174
+ split: train[:4000]
175
+ type: customcompletion-regex
176
+ - path: PJMixers-Dev/Fundus-AP-News-2-Formatted
177
+ type: customcompletion-regex
178
+ # Misc Data
179
+ - path: PJMixers-Dev/goodwiki-2024-12-04-axo
180
+ split: train[:4000]
181
+ type: customcompletion-regex
182
+ - path: epfl-llm/guidelines
183
+ split: train[:4000]
184
+ field: clean_text
185
+ type: customcompletion-regex
186
+ # Granite-3.1 Instruct
187
+ # Instruction Data
188
+ - path: PJMixers-Dev/allenai_tulu-3-sft-mixture-filtered-2-ShareGPT
189
+ split: train[:4000]
190
+ type: customgranite-regex
191
+ - path: OpenLeecher/lmsys_chat_1m_clean
192
+ split: train[:4000]
193
+ type: customgranite-regex
194
+ # RP Data
195
+ - path: PJMixers-Dev/Gryphe-Aesir-RPG-Charcards-Opus-Mixed
196
+ type: customgranite-regex
197
+ - path: allura-org/gryphe-sonnet-3.5-charcards-names-added
198
+ type: customgranite-regex
199
+ - path: anthracite-org/c2_logs_32k_llama3_qwen2_v1.3
200
+ type: customgranite-regex
201
+ - path: BeaverAI/REDACTED4
202
+ type: customgranite-regex
203
+ - path: PJMixers-Dev/MinervaAI_Aesir-Preview-Anon
204
+ type: customgranite-regex
205
+ - path: PJMixers-Dev/lemonilia_LimaRP-Simple-CustomShareGPT-Shuffled
206
+ type: customgranite-regex
207
+ - path: Epiculous/SynthRP-Gens-v1.1-Filtered-n-Cleaned
208
+ type: customgranite-regex
209
+ - path: PJMixers-Dev/NyxKrage_chub-logs-sharegpt-longest-CustomShareGPT
210
+ type: customgranite-regex
211
+ - path: PJMixers/OpenLeecher_Teatime_all_logs_longest-ShareGPT
212
+ type: customgranite-regex
213
+ - path: grimulkan/aicg-logs-augmented
214
+ type: customgranite-regex
215
+ - path: grimulkan/PIPPA-augmented-dedup
216
+ type: customgranite-regex
217
+ - path: PJMixers/grimulkan_bluemoon_Karen_cleaned-carded-formatted
218
+ type: customgranite-regex
219
+ # InstStory Data
220
+ - path: PJMixers/lodrick-the-lafted_OpusStories-ShareGPT
221
+ type: customgranite-regex
222
+ - path: Gryphe/ChatGPT-4o-Writing-Prompts
223
+ type: customgranite-regex
224
+ - path: Gryphe/Opus-WritingPrompts
225
+ type: customgranite-regex
226
+ - path: anthracite-org/nopm_claude_writing_fixed
227
+ type: customgranite-regex
228
+ - path: PJMixers-Dev/Tiefighter-13B-Fake-Distill-ShareGPT
229
+ type: customgranite-regex
230
+ - path: allura-org/fujin-instruct-v2
231
+ type: customgranite-regex
232
+ - path: ToastyPigeon/gutenberg-sft
233
+ type: customgranite-regex
234
+ # Adventure Data
235
+ - path: PocketDoc/Dans-Prosemaxx-Adventure
236
+ type: customgranite-regex
237
+ - path: PocketDoc/Dans-Failuremaxx-Adventure-3
238
+ type: customgranite-regex
239
+ # Decensoring Data
240
+ - path: TheDrummer/AmoralQA-v2
241
+ type: customgranite-regex
242
+ - path: BeaverAI/REDACTED5
243
+ type: customgranite-regex
244
+ - path: BeaverAI/REDACTED6
245
+ type: customgranite-regex
246
+ val_set_size: 256
247
+ eval_strategy: steps
248
+ eval_steps: 10
249
+ dataset_prepared_path: ./00-Tokenized-Datasets/Granite-3.1-Earthen-v0.3-1B-A400M-LoRA-seed42
250
+ shuffle_merged_datasets: true
251
+
252
+ # Training hyperparameters
253
+ num_epochs: 1
254
+ gradient_accumulation_steps: 4
255
+ micro_batch_size: 4
256
+ eval_batch_size: 4
257
+ warmup_steps: 0
258
+ optimizer: came_pytorch
259
+ optim_args:
260
+ enable_stochastic_rounding: true
261
+ enable_cautious: true
262
+ enable_8bit: true
263
+ lr_scheduler: rex
264
+ learning_rate: 2.5e-7
265
+ cosine_min_lr_ratio: 0.05
266
+ weight_decay: 0.01
267
+ max_grad_norm: 0.5
268
+ logging_steps: 1
269
+
270
+ # Model optimization
271
+ gradient_checkpointing: offload
272
+ sdp_attention: true
273
+ plugins:
274
+ - axolotl.integrations.liger.LigerPlugin
275
+ liger_rope: true
276
+ liger_rms_norm: true
277
+ liger_layer_norm: true
278
+ liger_glu_activation: true
279
+ liger_cross_entropy: true
280
+ lora_mlp_kernel: false
281
+ lora_qkv_kernel: false
282
+ lora_o_kernel: false
283
+
284
+ # Debug config
285
+ debug: true
286
+ seed: 42
287
+
288
+ # Token config
289
+ special_tokens:
290
+ bos_token: "<|end_of_text|>"
291
+ eos_token: "<|end_of_text|>"
292
+ pad_token: "<|end_of_text|>"
293
+ tokens:
294
+ ```
295
+
296
+ ## Citations
297
+
298
+ <details><summary>Show Citations</summary>
299
+
300
+ ```bib
301
+ @misc{wolf2020huggingfacestransformersstateoftheartnatural,
302
+ title={HuggingFace's Transformers: State-of-the-art Natural Language Processing},
303
+ author={Thomas Wolf and Lysandre Debut and Victor Sanh and Julien Chaumond and Clement Delangue and Anthony Moi and Pierric Cistac and Tim Rault and Rémi Louf and Morgan Funtowicz and Joe Davison and Sam Shleifer and Patrick von Platen and Clara Ma and Yacine Jernite and Julien Plu and Canwen Xu and Teven Le Scao and Sylvain Gugger and Mariama Drame and Quentin Lhoest and Alexander M. Rush},
304
+ year={2020},
305
+ eprint={1910.03771},
306
+ archivePrefix={arXiv},
307
+ primaryClass={cs.CL},
308
+ url={https://arxiv.org/abs/1910.03771},
309
+ }
310
+ @misc{hu2021loralowrankadaptationlarge,
311
+ title={LoRA: Low-Rank Adaptation of Large Language Models},
312
+ author={Edward J. Hu and Yelong Shen and Phillip Wallis and Zeyuan Allen-Zhu and Yuanzhi Li and Shean Wang and Lu Wang and Weizhu Chen},
313
+ year={2021},
314
+ eprint={2106.09685},
315
+ archivePrefix={arXiv},
316
+ primaryClass={cs.CL},
317
+ url={https://arxiv.org/abs/2106.09685},
318
+ }
319
+ @misc{dettmers2023qloraefficientfinetuningquantized,
320
+ title={QLoRA: Efficient Finetuning of Quantized LLMs},
321
+ author={Tim Dettmers and Artidoro Pagnoni and Ari Holtzman and Luke Zettlemoyer},
322
+ year={2023},
323
+ eprint={2305.14314},
324
+ archivePrefix={arXiv},
325
+ primaryClass={cs.LG},
326
+ url={https://arxiv.org/abs/2305.14314},
327
+ }
328
+ @misc{dao2023flashattention2fasterattentionbetter,
329
+ title={FlashAttention-2: Faster Attention with Better Parallelism and Work Partitioning},
330
+ author={Tri Dao},
331
+ year={2023},
332
+ eprint={2307.08691},
333
+ archivePrefix={arXiv},
334
+ primaryClass={cs.LG},
335
+ url={https://arxiv.org/abs/2307.08691},
336
+ }
337
+ @misc{hsu2024ligerkernelefficienttriton,
338
+ title={Liger Kernel: Efficient Triton Kernels for LLM Training},
339
+ author={Pin-Lun Hsu and Yun Dai and Vignesh Kothapalli and Qingquan Song and Shao Tang and Siyu Zhu and Steven Shimizu and Shivam Sahni and Haowen Ning and Yanning Chen},
340
+ year={2024},
341
+ eprint={2410.10989},
342
+ archivePrefix={arXiv},
343
+ primaryClass={cs.LG},
344
+ url={https://arxiv.org/abs/2410.10989},
345
+ }
346
+ @misc{chen2021rexrevisitingbudgetedtraining,
347
+ title={REX: Revisiting Budgeted Training with an Improved Schedule},
348
+ author={John Chen and Cameron Wolfe and Anastasios Kyrillidis},
349
+ year={2021},
350
+ eprint={2107.04197},
351
+ archivePrefix={arXiv},
352
+ primaryClass={cs.LG},
353
+ url={https://arxiv.org/abs/2107.04197},
354
+ }
355
+ @misc{luo2023cameconfidenceguidedadaptivememory,
356
+ title={CAME: Confidence-guided Adaptive Memory Efficient Optimization},
357
+ author={Yang Luo and Xiaozhe Ren and Zangwei Zheng and Zhuo Jiang and Xin Jiang and Yang You},
358
+ year={2023},
359
+ eprint={2307.02047},
360
+ archivePrefix={arXiv},
361
+ primaryClass={cs.CL},
362
+ url={https://arxiv.org/abs/2307.02047},
363
+ }
364
+ @misc{zamirai2021revisitingbfloat16training,
365
+ title={Revisiting BFloat16 Training},
366
+ author={Pedram Zamirai and Jian Zhang and Christopher R. Aberger and Christopher De Sa},
367
+ year={2021},
368
+ eprint={2010.06192},
369
+ archivePrefix={arXiv},
370
+ primaryClass={cs.LG},
371
+ url={https://arxiv.org/abs/2010.06192},
372
+ }
373
+ @misc{liang2025cautiousoptimizersimprovingtraining,
374
+ title={Cautious Optimizers: Improving Training with One Line of Code},
375
+ author={Kaizhao Liang and Lizhang Chen and Bo Liu and Qiang Liu},
376
+ year={2025},
377
+ eprint={2411.16085},
378
+ archivePrefix={arXiv},
379
+ primaryClass={cs.LG},
380
+ url={https://arxiv.org/abs/2411.16085},
381
+ }
382
+ @misc{xie2025sana15efficientscaling,
383
+ title={SANA 1.5: Efficient Scaling of Training-Time and Inference-Time Compute in Linear Diffusion Transformer},
384
+ author={Enze Xie and Junsong Chen and Yuyang Zhao and Jincheng Yu and Ligeng Zhu and Chengyue Wu and Yujun Lin and Zhekai Zhang and Muyang Li and Junyu Chen and Han Cai and Bingchen Liu and Daquan Zhou and Song Han},
385
+ year={2025},
386
+ eprint={2501.18427},
387
+ archivePrefix={arXiv},
388
+ primaryClass={cs.CV},
389
+ url={https://arxiv.org/abs/2501.18427},
390
+ }
391
+ @misc{dallabetta2024fundussimpletousenewsscraper,
392
+ title={Fundus: A Simple-to-Use News Scraper Optimized for High Quality Extractions},
393
+ author={Max Dallabetta and Conrad Dobberstein and Adrian Breiding and Alan Akbik},
394
+ year={2024},
395
+ eprint={2403.15279},
396
+ archivePrefix={arXiv},
397
+ primaryClass={cs.CL},
398
+ url={https://arxiv.org/abs/2403.15279},
399
+ }
400
+ @misc{lambert2025tulu3pushingfrontiers,
401
+ title={Tulu 3: Pushing Frontiers in Open Language Model Post-Training},
402
+ author={Nathan Lambert and Jacob Morrison and Valentina Pyatkin and Shengyi Huang and Hamish Ivison and Faeze Brahman and Lester James V. Miranda and Alisa Liu and Nouha Dziri and Shane Lyu and Yuling Gu and Saumya Malik and Victoria Graf and Jena D. Hwang and Jiangjiang Yang and Ronan Le Bras and Oyvind Tafjord and Chris Wilhelm and Luca Soldaini and Noah A. Smith and Yizhong Wang and Pradeep Dasigi and Hannaneh Hajishirzi},
403
+ year={2025},
404
+ eprint={2411.15124},
405
+ archivePrefix={arXiv},
406
+ primaryClass={cs.CL},
407
+ url={https://arxiv.org/abs/2411.15124},
408
+ }
409
+ @misc{zheng2024lmsyschat1mlargescalerealworldllm,
410
+ title={LMSYS-Chat-1M: A Large-Scale Real-World LLM Conversation Dataset},
411
+ author={Lianmin Zheng and Wei-Lin Chiang and Ying Sheng and Tianle Li and Siyuan Zhuang and Zhanghao Wu and Yonghao Zhuang and Zhuohan Li and Zi Lin and Eric P. Xing and Joseph E. Gonzalez and Ion Stoica and Hao Zhang},
412
+ year={2024},
413
+ eprint={2309.11998},
414
+ archivePrefix={arXiv},
415
+ primaryClass={cs.CL},
416
+ url={https://arxiv.org/abs/2309.11998},
417
+ }
418
+ @misc{gosling2023pippapartiallysyntheticconversational,
419
+ title={PIPPA: A Partially Synthetic Conversational Dataset},
420
+ author={Tear Gosling and Alpin Dale and Yinhe Zheng},
421
+ year={2023},
422
+ eprint={2308.05884},
423
+ archivePrefix={arXiv},
424
+ primaryClass={cs.CL},
425
+ url={https://arxiv.org/abs/2308.05884},
426
+ }
427
+ ```
428
+
429
+ </details>