Datasets:
id
stringlengths 1
4
| tokens
sequence | ner_tags
sequence |
---|---|---|
0 | [
"I",
"hate",
"the",
"words",
"chunder",
",",
"vomit",
"and",
"puke",
".",
"BUUH",
"."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
1 | [
"♥",
".",
".",
")",
")",
"(",
"♫",
".",
"(",
"ړײ",
")",
"♫",
".",
"♥",
".",
"«",
"▓",
"»",
"♥",
".",
"♫",
".",
".",
"╝",
"╚",
".",
".",
"♫",
"Happy",
"New",
"Year"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2 | [
"Alesan",
"kenapa",
"mlm",
"kita",
"lbh",
"srg",
"galau",
"Poconggg",
"'\"'",
"TwitFAKTA",
":",
"Otak",
"lebih",
"aktif",
"di",
"malam",
"hari",
"dari",
"pada",
"di",
"pagi",
"hari",
".",
"#",
"TwitFAKTA",
"'\"'"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
3 | [
"Complete",
"Tosca",
"on",
"the",
"tube",
"http://t.co/O90deSLB"
] | [
0,
0,
0,
0,
0,
0
] |
4 | [
"Think",
"you",
"call",
"that",
"smash",
"and",
"grab",
".",
"#",
"Gateshead",
"'s",
"media",
"man",
"just",
"admitted",
"to",
"me",
"it",
"was",
"'\"'",
"daylight",
"robbery",
"'\"'",
".",
"Shaw",
"'s",
"only",
"touch",
"was",
"his",
"goal",
"."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
5,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
1,
0,
0,
0,
0,
0,
0,
0
] |
5 | [
"Happy",
"New",
"Year",
"world",
"!"
] | [
0,
0,
0,
0,
0
] |
6 | [
"middle",
"aged",
"man",
"band",
"playing",
"blink",
"182",
".",
"l0",
"l",
"."
] | [
0,
0,
0,
0,
0,
3,
4,
4,
0,
0,
0
] |
7 | [
"14",
"days",
"till",
"my",
"birthday",
":",
")"
] | [
0,
0,
0,
0,
0,
0,
0
] |
8 | [
"Happy",
"New",
"Year",
"!",
"It",
"'s",
"time",
"to",
"start",
"thinking",
"where",
"you",
"'d",
"like",
"to",
"visit",
"in",
"#",
"2012",
".",
"Need",
"some",
"ideas",
"?",
"http://t.co/NpsUeTAG"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
9 | [
"i",
"'ve",
"got",
"dressed",
"but",
"only",
"because",
"I",
"need",
"biscuits"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
10 | [
"Happy",
"New",
"Year",
"2012",
"#",
"fireworks",
"#",
"HappyNewYear",
"#",
"photography",
"http://t.co/ImHeLldm"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
11 | [
"Happy",
"New",
"Year",
"to",
"you",
"all",
".",
"Feeling",
"pretty",
"virtuous",
"after",
"organising",
"a",
"row",
"for",
"some",
"mates",
"and",
"myself",
"at",
"10",
"this",
"morning",
".",
"It",
"was",
"great",
".",
".",
".",
"ish",
"!"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
12 | [
"My",
"dad",
"thinks",
"I",
"ca",
"n't",
"hear",
"him",
"singing",
"in",
"the",
"kitchen",
".",
".",
"."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
13 | [
"How",
"did",
"Dorothy",
"Gale",
"come",
"back",
"*",
"younger",
"*",
"in",
"Return",
"to",
"Oz",
"?",
"?"
] | [
0,
0,
1,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
14 | [
"Need",
"to",
"stop",
"eating",
"so",
"much",
"!"
] | [
0,
0,
0,
0,
0,
0,
0
] |
15 | [
"a",
"little",
"confused",
"on",
"who",
"I",
"did",
"the",
"countdown",
"last",
"night",
"."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
16 | [
"looking",
"4",
"a",
"new",
"dancehall",
"albu",
"to",
"download",
".",
".",
".",
".",
".",
".",
"nebody",
"hav",
"ne",
"ideas",
"?",
"?",
"?",
"?"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
17 | [
"Check",
"out",
"the",
"new",
"#",
"TantamountToTreaser",
"preview",
"on",
"our",
"website",
"!",
"http://t.co/NiqVnls3"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
18 | [
"I",
"gotta",
"feel",
"alive",
"even",
"if",
"it",
"kills",
"me",
",",
"promise",
"to",
"always",
"give",
"you",
"me",
"the",
"real",
"me"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
19 | [
"TD",
"Chargers",
"=",
"\\",
".",
"If",
"we",
"ca",
"n't",
"beat",
"the",
"Chargers",
"we",
"do",
"n't",
"deserve",
"to",
"go",
"to",
"the",
"playoffs",
"."
] | [
3,
4,
0,
0,
0,
0,
0,
0,
0,
0,
3,
4,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
20 | [
"well",
"I",
"dunno",
"whether",
"this",
"makes",
"me",
"uncool",
"or",
"uncultured",
"or",
"what",
"but",
"I",
"blimmin",
"LOVED",
"sherlock",
",",
"more",
"more",
"please",
".",
"yay",
"good",
"telly",
"finally"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
21 | [
"Random",
"boy",
"I",
"'ve",
"never",
"met",
"before",
"said",
"to",
"me",
"last",
"night",
"'",
"you",
"'re",
"saffron",
"coe",
",",
"you",
"'re",
"19",
"and",
"you",
"live",
"in",
"stamford",
"'",
"erm",
"#",
"lockingmydoors"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
1,
2,
0,
0,
0,
0,
0,
0,
0,
0,
5,
0,
0,
0,
0
] |
22 | [
"Deciding",
"whether",
"or",
"not",
"to",
"add",
"my",
"parents",
"in",
"facebook",
"or",
"not",
".",
".",
".",
"What",
"do",
"you",
"guys",
"think",
"?"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
23 | [
"Wadey",
"Is",
"3",
"Sets",
"To",
"1",
"Now",
".",
"3",
"More",
"Sets",
"&",
"His",
"In",
"The",
"Final",
"!",
"!"
] | [
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
24 | [
"#",
"DownloadDYFWIF",
"please",
"please",
"please",
"JLSOfficial"
] | [
0,
0,
0,
0,
0,
0
] |
25 | [
"It",
"'s",
"hard",
"to",
"find",
"a",
"boy",
"like",
"you",
",",
"and",
"I",
"guess",
"it",
"'s",
"hard",
"to",
"find",
"a",
"girl",
"like",
"me",
".",
"But",
"I",
"just",
"suppose",
"that",
"'s",
"the",
"way",
"things",
"are",
"."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
26 | [
"Well",
",",
"this",
"made",
"My",
"Day",
"Thank",
"Apple",
":",
")",
"http://t.co/YOj7iZrn",
"#",
"RelationshipSongs",
"#",
"NoStupidHeadsAllowed",
"#",
"InMyFreeTime"
] | [
0,
0,
0,
0,
0,
0,
0,
3,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
27 | [
"Retweet",
"if",
"you",
"'ve",
"ever",
"had",
"/",
"been",
"to",
"a",
"party",
"where",
"the",
"stupid",
"neighbours",
"complained",
"and",
"called",
"fed"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
28 | [
"I",
"think",
"it",
"'s",
"11",
",",
"it",
"'s",
"pretty",
"hard",
"to",
"count",
".",
"Maybe",
"it",
"'s",
"10",
".",
"I",
"went",
"to",
"the",
"theatre",
"17",
"times",
"in",
"total",
"though",
"."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
29 | [
"This",
"life",
".",
".",
".",
".",
".",
"#",
"Ones"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
30 | [
"What",
"sleeping",
"pattern",
"?"
] | [
0,
0,
0,
0
] |
31 | [
"Why",
"remain",
"in",
"a",
"relationship",
"that",
"has",
"no",
"benefits",
".",
".",
".",
"?",
"Why",
"stay",
"in",
"a",
"relationship",
"when",
"you",
"have",
"no",
"reasons",
"to",
"?"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
32 | [
"The",
"music",
"sounds",
"better",
"with",
"you",
"#",
"fb"
] | [
0,
0,
0,
0,
0,
0,
0,
0
] |
33 | [
"[",
"VIDEO",
"]",
":",
"How",
"to",
"Get",
"Promoted",
"at",
"Work",
"http://t.co/hakhXCcN",
"#",
"Career",
"#",
"Featured"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
34 | [
"people",
"who",
"send",
"there",
"tweets",
"to",
"facebook",
"omg",
"fuck",
"of"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
35 | [
"I",
"used",
"to",
"be",
"so",
"bad",
"at",
"naming",
"any",
"k",
"idol",
"group",
"members",
"pmsl",
"I",
"would",
"get",
"so",
"confused",
"and",
"now",
"I",
"'m",
"pro",
";",
")",
")",
")"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
3,
4,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
36 | [
"Neighbors",
"Occupy",
"Road",
",",
"Blockade",
"Sludge",
"Trucks",
":",
"Recently",
",",
"a",
"group",
"of",
"farmers",
"and",
"neighbors",
"in",
"Salmon",
"Valley",
",",
"near",
".",
".",
".",
"http://t.co/dvNVix5J"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
5,
6,
0,
0,
0,
0,
0,
0
] |
37 | [
"Well",
"that",
"could",
"of",
"been",
"a",
"lot",
"more",
"painful",
".",
"3",
"miles",
",",
"fairly",
"slowly",
"but",
"my",
"lungs",
"did",
"n't",
"try",
"to",
"leap",
"out",
"of",
"my",
"chest",
".",
"My",
"thighs",
"though",
".",
".",
".",
".",
".",
"OWW"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
38 | [
"smart",
"move",
"of",
"the",
"J",
".",
"Edgar",
"Promo",
"team",
"showing",
"exclusively",
"young",
"shots",
"of",
"leo",
"to",
"grab",
"the",
"fan",
"base",
".",
".",
"them",
"be",
"some",
"confused",
"punters",
";",
")"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
39 | [
"Photo",
":",
"http://t.co/RGBss9WK"
] | [
0,
0,
0
] |
40 | [
"Pride",
"and",
"Prejudice",
"is",
"on",
".",
"Oh",
"my",
"oh",
"my",
"be",
"still",
"my",
"beating",
"heart",
"."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
41 | [
"I",
"'ve",
"been",
"making",
"shapes",
"in",
"this",
"for",
"2",
"days",
"solid",
".",
"#",
"simplethingsinlife",
"http://t.co/Aa0i40NI"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
42 | [
"Chfhc",
"vm"
] | [
0,
0
] |
43 | [
"FUUUUUULHAAAAAAM",
"!",
"!",
"!"
] | [
3,
0,
0,
0
] |
44 | [
"Wengers",
"big",
"mistake",
"is",
"not",
"being",
"ruthless",
"enough",
"with",
"shit",
"players",
".",
"Do",
"they",
"want",
"to",
"go",
"?",
"Probably",
"no",
"but",
"when",
"your",
"the",
"manager",
"you",
"hold",
"cards"
] | [
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
45 | [
"I",
"lobe",
"Hobo",
"Rob",
"#",
"JS"
] | [
0,
0,
0,
1,
0,
0
] |
46 | [
"Not",
"sure",
"if",
"I",
"like",
"#",
"bankjob",
"or",
"not",
"?",
"Maybe",
"it",
"'s",
"because",
"I",
"do",
"n't",
"understand",
"it",
"#",
"confused"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
47 | [
"If",
"I",
"have",
"to",
"yell",
"at",
"Otis",
"one",
"more",
"time",
"tonight",
"for",
"doing",
"something",
"dangerous",
",",
"he",
"'s",
"getting",
"sent",
"back",
"to",
"Dad",
"'s",
".",
"Boy",
"'s",
"got",
"a",
"death",
"wish"
] | [
0,
0,
0,
0,
0,
0,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
48 | [
"need",
"a",
"new",
"tattoo",
".",
".",
".",
".",
"any",
"suggestions",
"guys",
"?",
"?",
"tweet",
"me"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
49 | [
"Fresh",
"sheets",
"&",
"lt",
";",
"3"
] | [
0,
0,
0,
0,
0,
0
] |
50 | [
"I",
"'m",
"going",
"to",
"regret",
"staying",
"up",
"this",
"late",
"."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
51 | [
"Fat",
"or",
"thin",
"am",
"still",
"fit",
"hahahaha"
] | [
0,
0,
0,
0,
0,
0,
0
] |
52 | [
"#",
"lessambitiousbooks",
"The",
"Green",
"inch",
"."
] | [
0,
0,
0,
0,
0,
0
] |
53 | [
"same",
"shit",
"different",
"day",
"."
] | [
0,
0,
0,
0,
0
] |
54 | [
"Kane",
"loves",
"coming",
"through",
"the",
"ring",
".",
"What",
"'s",
"this",
"new",
"move",
"all",
"about",
"?",
"It",
"'s",
"a",
"shit",
"mandible",
"claw",
"#",
"Raw",
"#",
"wrestlezone"
] | [
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
55 | [
"It",
"'s",
"too",
"bloody",
"early",
",",
"I",
"want",
"to",
"go",
"back",
"to",
"bed",
"as",
"it",
"'s",
"cold",
",",
"it",
"'s",
"wet",
"and",
"it",
"'s",
"blowing",
"a",
"bloody",
"gale",
"out",
"there",
":",
"("
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
56 | [
"10",
"years",
"ago",
"today",
"I",
"won",
"my",
"world",
"title",
"with",
"my",
"dancing",
".",
"I",
"'m",
"now",
"seeing",
"the",
"results",
"for",
"the",
"2012",
"world",
"championships",
"at",
"disneyworld",
",",
"Florida"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
5,
0,
5
] |
57 | [
"Good",
"morning",
"all",
".",
"(",
":"
] | [
0,
0,
0,
0,
0,
0
] |
58 | [
"feel",
"sorry",
"for",
"my",
"dad",
",",
"he",
"has",
"to",
"work",
"in",
"this",
"weather",
"all",
"day"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
59 | [
"Do",
"n't",
"even",
"care",
"about",
"mocks",
",",
"they",
"'re",
"mocks",
"ffs",
"!",
"#",
"pointless",
"."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
60 | [
"You",
"just",
"piss",
"my",
"actual",
"life",
"off"
] | [
0,
0,
0,
0,
0,
0,
0
] |
61 | [
"Home",
"Business",
"Marketing",
"Through",
"Your",
"Social",
"Media",
"Profile"
] | [
0,
0,
0,
0,
0,
0,
0,
0
] |
62 | [
"#",
"ICD",
"Daniel",
"James",
"Evans",
",",
"Woolwich",
"Room",
":",
"3",
"at",
"10",
":",
"39"
] | [
0,
0,
1,
2,
2,
0,
5,
6,
0,
0,
0,
0,
0,
0
] |
63 | [
"Ski",
"Resorts",
"Note",
"Increase",
"in",
"Car",
"Hire",
"Demand",
"|",
"Season",
"Car",
"Hire",
"Blog",
"http://t.co/5amtlxc3"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
64 | [
"Por",
"primera",
"vez",
"me",
"pagan",
"(",
"de",
"verdad",
")",
"productividad",
".",
"Aleluya",
"!",
"!",
"!"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
65 | [
"#",
"jobs",
"#",
"php",
"Database",
"Developer",
",",
"Oracle",
",",
"MySQL",
",",
"Perl",
",",
"OO",
",",
"Java",
":",
"Salary",
"/",
"Rate",
":",
"Market",
"RateLocation",
":",
"city",
"of",
"london",
"Da",
".",
".",
".",
"http://t.co/PhPoOza4"
] | [
0,
0,
0,
0,
0,
0,
0,
3,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
5,
6,
6,
0,
0,
0,
0,
0
] |
66 | [
"Último",
"dia",
"de",
"férias",
",",
"tempo",
"feio",
",",
"chuva",
".",
".",
".",
"boo"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
67 | [
"Who",
"is",
"stupid",
"enough",
"to",
"wear",
"uggs",
"in",
"this",
"weather",
".",
".",
".",
".",
"Me",
"!",
"That",
"'s",
"who",
".",
".",
".",
"#",
"fml",
"http://t.co/jNSVSCsl"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
68 | [
"#",
"Jobs",
"#",
"Sales",
"Telesales",
"/",
"Telemarketing",
"/",
"Business",
"Development",
":",
"Yorkshire-",
"Doncaster",
",",
"My",
"client",
"is",
"a",
"well",
"establ",
".",
".",
".",
"http://t.co/EaOHrSZ2",
"#",
"Leeds"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
5,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
5
] |
69 | [
"Sticky",
"toffee",
"pudding",
"from",
"plate",
"to",
"tummy",
"in",
"under",
"2",
"minutes",
".",
"Feeling",
"a",
"little",
"unwell",
"now",
"."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
70 | [
"Same",
"goes",
"for",
"the",
"new",
"Lamb",
"of",
"God",
"stuff",
"too",
"!"
] | [
0,
0,
0,
0,
0,
3,
4,
4,
0,
0,
0
] |
71 | [
"hopefully",
"i",
"'m",
"right"
] | [
0,
0,
0,
0
] |
72 | [
"Parents",
"in",
"law",
"are",
"cooking",
"dinner",
".",
"I",
"would",
"go",
"and",
"help",
"but",
"I",
"can",
"hear",
"father",
"in",
"law",
"is",
"getting",
"shouted",
"at",
",",
"which",
"means",
"I",
"will",
"too",
".",
"#",
"hides"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
73 | [
"#",
"ManCity",
"v",
"#",
"Liverpool",
"kicks",
"off",
"at",
"8",
",",
"will",
"be",
"a",
"tough",
"game",
"for",
"both",
"sides",
"but",
"I",
"reckon",
"City",
"will",
"take",
"all",
"3",
"points",
"away",
"from",
"Liverpool",
".",
"#",
"PL"
] | [
0,
3,
0,
0,
3,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
3,
0,
0,
0,
0,
0,
0,
0,
3,
0,
0,
0
] |
74 | [
"I",
"JUST",
"LOST",
"ALL",
"MY",
"MUSIC",
"ON",
"MY",
"PLAYLIST",
"!",
"!",
"!",
"GAME",
"OVER",
"#",
"itunes",
"#",
"twats"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
75 | [
"literally",
"all",
"my",
"trousers",
"are",
"too",
"big",
"for",
"me",
":",
"(",
":",
"("
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
76 | [
"im",
"gonna",
"learn",
"to",
"be",
"a",
"lifegaurd",
"hopfully",
"so",
"while",
"everyone",
"else",
"is",
"working",
"in",
"a",
"shop",
"actually",
"doing",
"stuff",
"il",
"be",
"sitting",
"on",
"a",
"pool",
"side",
".",
"yay"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
77 | [
"Kuyt",
"responsible",
"for",
"giving",
"ball",
"away",
"prior",
"to",
"the",
"#",
"Aguero",
"goal",
".",
"From",
"being",
"someone",
"previously",
"so",
"reliable",
",",
"he",
"'s",
"been",
"terrible",
"this",
"year",
"."
] | [
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
78 | [
"Back",
"on",
"the",
"90210",
"buzz",
"!"
] | [
0,
0,
0,
0,
0,
0
] |
79 | [
"Am",
"I",
"the",
"only",
"one",
"who",
"thinks",
"that",
"Louise",
"Mensch",
"looks",
"like",
"a",
"female",
"version",
"of",
"Alan",
"B",
"'",
"stard",
"with",
"views",
"to",
"match",
"?"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
1,
2,
0,
0,
0,
0,
0,
0,
1,
2,
2,
2,
0,
0,
0,
0,
0
] |
80 | [
"The",
"Business",
"#",
"GreatFilm"
] | [
0,
0,
0,
0
] |
81 | [
"If",
"it",
"was",
"a",
"white",
"man",
"stabbed",
"by",
"a",
"group",
"of",
"black",
"men",
".",
".",
"."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
82 | [
"I",
"think",
"I",
"may",
"have",
"to",
"stop",
"following",
"people",
"who",
"only",
"are",
"sad",
"on",
"twitter",
",",
"it",
"makes",
"me",
"unhappy",
".",
"#",
"januaryblues",
"#",
"cheerup"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
83 | [
"if",
"i",
"changed",
"the",
"world",
"for",
"you",
"i",
"bet",
"you",
"would",
"n't",
"have",
"a",
"clue"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
84 | [
"Feddddd",
"up"
] | [
0,
0
] |
85 | [
"Random",
"thought",
"why",
"do",
"ppl",
"from",
"East",
"(",
"not",
"Hackney",
")",
"say",
"Yii",
"And",
"wt",
"does",
"it",
"mean",
":",
"S"
] | [
0,
0,
0,
0,
0,
0,
5,
0,
0,
5,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
86 | [
"Come",
"Dine",
"with",
"Me",
"-",
"Extra",
"Spicy",
"http://t.co/KS0ROHmm",
"Winner",
"of",
"Best",
"Lifestyle",
"Show",
"at",
"the",
"2011",
"TV",
"Choice",
"Awards",
"#",
"Amazon",
"#",
"UK",
"#",
"DVD",
"#",
"ITV"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
3,
0,
0,
0,
0,
0,
3
] |
87 | [
"Have",
"you",
"Liked",
"our",
"Facebook",
"page",
"?",
"http://t.co/6Vkx6tzN",
"Start",
"2012",
"with",
"a",
"Like",
"!",
"You",
"can",
"even",
"get",
"an",
"energy",
"quote",
"there",
"too",
"!"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
88 | [
"Putting",
"the",
"finishing",
"touches",
"on",
"my",
"Mindful",
"Living",
"Guide",
"mailing",
"about",
"the",
"6",
"-",
"week",
"I",
"CHING",
"course",
",",
"#",
"Lewes",
",",
"beg",
".",
"20",
"Jan",
":",
"http://t.co/rhRPOjuD"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
5,
0,
0,
0,
0,
0,
0,
0
] |
89 | [
"Do",
"n't",
"you",
"hate",
"it",
"when",
"you",
"can",
"see",
"your",
"parent",
"is",
"upset",
"/",
"confused",
",",
"you",
"have",
"an",
"idea",
"to",
"help",
"but",
"wo",
"n't",
"confront",
"them",
"about",
"it",
"and",
"you",
"just",
"+"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
90 | [
"listening",
"to",
"'\"'",
"Time",
"is",
"not",
"Enough",
"-",
"BRENDA",
"LEE",
"'\"'",
"♫",
"http://t.co/nQYdQoO8"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
1,
2,
0,
0,
0
] |
91 | [
"I",
"'m",
"still",
"opening",
"xmas",
"presents",
"#",
"ridiculous",
".",
".",
".",
"anyone",
"want",
"an",
"idiot",
"abroad",
"1",
"on",
"dvd",
"?"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
92 | [
"Rachid",
"Harkouk",
"#",
"randomformerfootballers",
"#",
"qpr",
"#",
"showsmyage"
] | [
1,
2,
0,
0,
0,
3,
0,
0
] |
93 | [
"Schaue",
"gerade",
"mal",
"wieder",
"Sex",
"and",
"the",
"City",
"und",
"stelle",
"so",
"fest",
":",
"Carrie",
"ist",
"eine",
"ganz",
"schön",
"hysterische",
"Ziege",
"."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
94 | [
"same",
"shit",
",",
"different",
"day",
"."
] | [
0,
0,
0,
0,
0,
0
] |
95 | [
"The",
"Wind",
"In",
"The",
"Willow",
":",
"Part",
"of",
"what",
"garden",
"writers",
"refer",
"to",
"as",
"'\"'",
"the",
"borrowed",
"landscape",
"…",
"http://t.co/rNNoLoxh"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
96 | [
"Still",
"find",
"it",
"creepy",
"the",
"account",
"trying",
"to",
"be",
"Casey",
"hahaha",
"freak"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
1,
0,
0
] |
97 | [
"33"
] | [
0
] |
98 | [
"murked",
"a",
"FP",
"1",
"paper",
".",
"YEAH",
"BOY",
"."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
99 | [
"Besti-vid",
"of",
"the",
"week",
":",
"Air",
"'",
"Sonic",
"Armada",
"'",
"(",
"A",
"Trip",
"To",
"The",
"Moon",
")",
"http://t.co/TmgSpWjy"
] | [
0,
0,
0,
0,
0,
3,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
Dataset Card for broad_twitter_corpus
Dataset Summary
This is the Broad Twitter corpus, a dataset of tweets collected over stratified times, places and social uses. The goal is to represent a broad range of activities, giving a dataset more representative of the language used in this hardest of social media formats to process. Further, the BTC is annotated for named entities.
See the paper, Broad Twitter Corpus: A Diverse Named Entity Recognition Resource, for details.
Supported Tasks and Leaderboards
- Named Entity Recognition
- On PWC: Named Entity Recognition on Broad Twitter Corpus
Languages
English from UK, US, Australia, Canada, Ireland, New Zealand; bcp47:en
Dataset Structure
Data Instances
Feature | Count |
---|---|
Documents | 9 551 |
Tokens | 165 739 |
Person entities | 5 271 |
Location entities | 3 114 |
Organization entities | 3 732 |
Data Fields
Each tweet contains an ID, a list of tokens, and a list of NER tags
id
: astring
feature.tokens
: alist
ofstrings
ner_tags
: alist
of class IDs (int
s) representing the NER class:
0: O
1: B-PER
2: I-PER
3: B-ORG
4: I-ORG
5: B-LOC
6: I-LOC
Data Splits
Section | Region | Collection period | Description | Annotators | Tweet count |
---|---|---|---|---|---|
A | UK | 2012.01 | General collection | Expert | 1000 |
B | UK | 2012.01-02 | Non-directed tweets | Expert | 2000 |
E | Global | 2014.07 | Related to MH17 disaster | Crowd & expert | 200 |
F | Stratified | 2009-2014 | Twitterati | Crowd & expert | 2000 |
G | Stratified | 2011-2014 | Mainstream news | Crowd & expert | 2351 |
H | Non-UK | 2014 | General collection | Crowd & expert | 2000 |
The most varied parts of the BTC are sections F and H. However, each of the remaining four sections has some specific readily-identifiable bias. So, we propose that one uses half of section H for evaluation and leaves the other half in the training data. Section H should be partitioned in the order of the JSON-format lines. Note that the CoNLL-format data is readily reconstructible from the JSON format, which is the authoritative data format from which others are derived.
Test: Section F
Development: Section H (the paper says "second half of Section H" but ordinality could be ambiguous, so it all goes in. Bonne chance)
Training: everything else
Dataset Creation
Curation Rationale
[Needs More Information]
Source Data
Initial Data Collection and Normalization
[Needs More Information]
Who are the source language producers?
[Needs More Information]
Annotations
Annotation process
[Needs More Information]
Who are the annotators?
[Needs More Information]
Personal and Sensitive Information
[Needs More Information]
Considerations for Using the Data
Social Impact of Dataset
[Needs More Information]
Discussion of Biases
[Needs More Information]
Other Known Limitations
[Needs More Information]
Additional Information
Dataset Curators
[Needs More Information]
Licensing Information
Creative Commons Attribution 4.0 International (CC BY 4.0)
Citation Information
@inproceedings{derczynski2016broad,
title={Broad twitter corpus: A diverse named entity recognition resource},
author={Derczynski, Leon and Bontcheva, Kalina and Roberts, Ian},
booktitle={Proceedings of COLING 2016, the 26th International Conference on Computational Linguistics: Technical Papers},
pages={1169--1179},
year={2016}
}
Contributions
Author-added dataset @leondz
- Downloads last month
- 119