Sebastian Gehrmann
commited on
Commit
·
a4619f9
1
Parent(s):
b1e3133
Data Card.
Browse files- OrangeSum.json +6 -2
- README.md +22 -8
OrangeSum.json
CHANGED
|
@@ -14,7 +14,7 @@
|
|
| 14 |
"additional-splits-capacicites": "N/A"
|
| 15 |
},
|
| 16 |
"starting": {
|
| 17 |
-
"research-pointers": "Papers about abstractive summarization using seq2seq models:\
|
| 18 |
"technical-terms": "No unique technical words in this data card."
|
| 19 |
}
|
| 20 |
},
|
|
@@ -79,7 +79,11 @@
|
|
| 79 |
"languages": {
|
| 80 |
"is-multilingual": "no",
|
| 81 |
"license": "other: Other license",
|
| 82 |
-
"task
|
|
|
|
|
|
|
|
|
|
|
|
|
| 83 |
},
|
| 84 |
"credit": {},
|
| 85 |
"structure": {}
|
|
|
|
| 14 |
"additional-splits-capacicites": "N/A"
|
| 15 |
},
|
| 16 |
"starting": {
|
| 17 |
+
"research-pointers": "Papers about abstractive summarization using seq2seq models:\n\n- [Abstractive Text Summarization using Sequence-to-sequence RNNs and Beyond](https://aclanthology.org/K16-1028/)\n- [Get To The Point: Summarization with Pointer-Generator Networks](https://aclanthology.org/P17-1099/)\n- [BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension](https://aclanthology.org/2020.acl-main.703)\n- [BARThez: a Skilled Pretrained French Sequence-to-Sequence Model](https://aclanthology.org/2021.emnlp-main.740/)\n\nPapers about (pretrained) Transformers:\n\n- [Attention is All you Need](https://papers.nips.cc/paper/2017/hash/3f5ee243547dee91fbd053c1c4a845aa-Abstract.html)\n- [BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding](https://aclanthology.org/N19-1423/)",
|
| 18 |
"technical-terms": "No unique technical words in this data card."
|
| 19 |
}
|
| 20 |
},
|
|
|
|
| 79 |
"languages": {
|
| 80 |
"is-multilingual": "no",
|
| 81 |
"license": "other: Other license",
|
| 82 |
+
"task": "Summarization",
|
| 83 |
+
"task-other": "N/A",
|
| 84 |
+
"language-names": [
|
| 85 |
+
"French"
|
| 86 |
+
]
|
| 87 |
},
|
| 88 |
"credit": {},
|
| 89 |
"structure": {}
|
README.md
CHANGED
|
@@ -15,7 +15,7 @@ size_categories:
|
|
| 15 |
source_datasets:
|
| 16 |
- original
|
| 17 |
task_categories:
|
| 18 |
-
-
|
| 19 |
task_ids:
|
| 20 |
- unknown
|
| 21 |
---
|
|
@@ -102,6 +102,13 @@ no
|
|
| 102 |
<!-- scope: telescope -->
|
| 103 |
no
|
| 104 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 105 |
#### License
|
| 106 |
|
| 107 |
<!-- quick -->
|
|
@@ -109,6 +116,12 @@ no
|
|
| 109 |
<!-- scope: telescope -->
|
| 110 |
other: Other license
|
| 111 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 112 |
|
| 113 |
### Credit
|
| 114 |
|
|
@@ -152,15 +165,16 @@ no
|
|
| 152 |
<!-- info: Getting started with in-depth research on the task. Add relevant pointers to resources that researchers can consult when they want to get started digging deeper into the task. -->
|
| 153 |
<!-- scope: microscope -->
|
| 154 |
Papers about abstractive summarization using seq2seq models:
|
| 155 |
-
|
| 156 |
-
https://aclanthology.org/
|
| 157 |
-
https://aclanthology.org/
|
| 158 |
-
https://aclanthology.org/
|
|
|
|
| 159 |
|
| 160 |
Papers about (pretrained) Transformers:
|
| 161 |
-
|
| 162 |
-
https://
|
| 163 |
-
https://aclanthology.org/
|
| 164 |
|
| 165 |
#### Technical Terms
|
| 166 |
|
|
|
|
| 15 |
source_datasets:
|
| 16 |
- original
|
| 17 |
task_categories:
|
| 18 |
+
- summarization
|
| 19 |
task_ids:
|
| 20 |
- unknown
|
| 21 |
---
|
|
|
|
| 102 |
<!-- scope: telescope -->
|
| 103 |
no
|
| 104 |
|
| 105 |
+
#### Covered Languages
|
| 106 |
+
|
| 107 |
+
<!-- quick -->
|
| 108 |
+
<!-- info: What languages/dialects are covered in the dataset? -->
|
| 109 |
+
<!-- scope: telescope -->
|
| 110 |
+
`French`
|
| 111 |
+
|
| 112 |
#### License
|
| 113 |
|
| 114 |
<!-- quick -->
|
|
|
|
| 116 |
<!-- scope: telescope -->
|
| 117 |
other: Other license
|
| 118 |
|
| 119 |
+
#### Primary Task
|
| 120 |
+
|
| 121 |
+
<!-- info: What primary task does the dataset support? -->
|
| 122 |
+
<!-- scope: telescope -->
|
| 123 |
+
Summarization
|
| 124 |
+
|
| 125 |
|
| 126 |
### Credit
|
| 127 |
|
|
|
|
| 165 |
<!-- info: Getting started with in-depth research on the task. Add relevant pointers to resources that researchers can consult when they want to get started digging deeper into the task. -->
|
| 166 |
<!-- scope: microscope -->
|
| 167 |
Papers about abstractive summarization using seq2seq models:
|
| 168 |
+
|
| 169 |
+
- [Abstractive Text Summarization using Sequence-to-sequence RNNs and Beyond](https://aclanthology.org/K16-1028/)
|
| 170 |
+
- [Get To The Point: Summarization with Pointer-Generator Networks](https://aclanthology.org/P17-1099/)
|
| 171 |
+
- [BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension](https://aclanthology.org/2020.acl-main.703)
|
| 172 |
+
- [BARThez: a Skilled Pretrained French Sequence-to-Sequence Model](https://aclanthology.org/2021.emnlp-main.740/)
|
| 173 |
|
| 174 |
Papers about (pretrained) Transformers:
|
| 175 |
+
|
| 176 |
+
- [Attention is All you Need](https://papers.nips.cc/paper/2017/hash/3f5ee243547dee91fbd053c1c4a845aa-Abstract.html)
|
| 177 |
+
- [BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding](https://aclanthology.org/N19-1423/)
|
| 178 |
|
| 179 |
#### Technical Terms
|
| 180 |
|