diff --git "a/CMU Advanced NLP 2024 (12) Reinforcement Learning/transcript.srt" "b/CMU Advanced NLP 2024 (12) Reinforcement Learning/transcript.srt" new file mode 100644--- /dev/null +++ "b/CMU Advanced NLP 2024 (12) Reinforcement Learning/transcript.srt" @@ -0,0 +1,7063 @@ +1 +00:00:00,840 --> 00:00:05,920 +okay so uh let's get started um today + +2 +00:00:04,200 --> 00:00:08,000 +I'm going to be talking about learning + +3 +00:00:05,920 --> 00:00:09,480 +from Human feedback I wrote + +4 +00:00:08,000 --> 00:00:12,160 +reinforcement learning from Human + +5 +00:00:09,480 --> 00:00:14,519 +feedback because that's what um you know + +6 +00:00:12,160 --> 00:00:15,759 +a lot of people talk about nowadays but + +7 +00:00:14,519 --> 00:00:18,880 +actually there's other methods of + +8 +00:00:15,759 --> 00:00:21,840 +learning from Human feedback so first + +9 +00:00:18,880 --> 00:00:24,760 +I'm going to be talking about the ways + +10 +00:00:21,840 --> 00:00:27,920 +we can get uh human feedback for the + +11 +00:00:24,760 --> 00:00:31,039 +generations of models and mostly focus + +12 +00:00:27,920 --> 00:00:32,960 +on generation tasks because is um + +13 +00:00:31,039 --> 00:00:35,800 +generation tasks are harder than like + +14 +00:00:32,960 --> 00:00:38,559 +classification tasks that we uh we deal + +15 +00:00:35,800 --> 00:00:40,000 +with normally so I'll spend a fair + +16 +00:00:38,559 --> 00:00:42,239 +amount of time talking about how we do + +17 +00:00:40,000 --> 00:00:45,760 +that and then after I talk about how we + +18 +00:00:42,239 --> 00:00:48,360 +do that we'll move into um how we + +19 +00:00:45,760 --> 00:00:51,160 +actually learn from that + +20 +00:00:48,360 --> 00:00:53,399 +signal so normally what we've done up + +21 +00:00:51,160 --> 00:00:56,399 +until this point is maximum likelihood + +22 +00:00:53,399 --> 00:00:58,199 +training uh this is just an overview + +23 +00:00:56,399 --> 00:00:59,559 +slide so we what we want to do is we + +24 +00:00:58,199 --> 00:01:00,760 +want to maximize the likelihood of + +25 +00:00:59,559 --> 00:01:03,280 +predicting the next word and the + +26 +00:01:00,760 --> 00:01:05,960 +reference given the previous words uh + +27 +00:01:03,280 --> 00:01:08,119 +which gives us the loss of the output + +28 +00:01:05,960 --> 00:01:09,799 +given the input uh where you know the + +29 +00:01:08,119 --> 00:01:13,960 +input can be the prompt the output can + +30 +00:01:09,799 --> 00:01:16,080 +be the answer to uh the output but + +31 +00:01:13,960 --> 00:01:18,360 +there's uh lots of problems with + +32 +00:01:16,080 --> 00:01:20,439 +learning from Maximum likelihood and I'm + +33 +00:01:18,360 --> 00:01:22,079 +going to give three examples here I + +34 +00:01:20,439 --> 00:01:24,159 +think all of these are actually real + +35 +00:01:22,079 --> 00:01:26,880 +problems uh that we need to be worried + +36 +00:01:24,159 --> 00:01:30,240 +about so the first one is that some + +37 +00:01:26,880 --> 00:01:32,439 +mistakes are worse than others so um in + +38 +00:01:30,240 --> 00:01:33,560 +the end we want good outputs and some + +39 +00:01:32,439 --> 00:01:36,520 +mistaken + +40 +00:01:33,560 --> 00:01:38,200 +predictions uh can be a bigger problem + +41 +00:01:36,520 --> 00:01:42,680 +for the output being + +42 +00:01:38,200 --> 00:01:46,000 +good so to give an example uh let's say + +43 +00:01:42,680 --> 00:01:47,600 +what we actually wanted from like a + +44 +00:01:46,000 --> 00:01:49,320 +speech recognition system or a + +45 +00:01:47,600 --> 00:01:54,040 +translation system or something like + +46 +00:01:49,320 --> 00:01:54,040 +that is uh please send this package to + +47 +00:01:54,280 --> 00:01:58,920 +Pittsburgh if I write please send a + +48 +00:01:56,880 --> 00:02:01,560 +package to Pittsburgh then this is not a + +49 +00:01:58,920 --> 00:02:03,560 +huge problem + +50 +00:02:01,560 --> 00:02:06,479 +if I write uh please send this package + +51 +00:02:03,560 --> 00:02:07,719 +to Tokyo then that might be a big + +52 +00:02:06,479 --> 00:02:09,640 +problem because the package you wanted + +53 +00:02:07,719 --> 00:02:12,760 +to come to Pittsburgh goes to Tokyo + +54 +00:02:09,640 --> 00:02:13,680 +instead and uh you might not want that + +55 +00:02:12,760 --> 00:02:16,080 +to + +56 +00:02:13,680 --> 00:02:18,000 +happen you might also have it say + +57 +00:02:16,080 --> 00:02:20,400 +bleeping send this package to Pittsburgh + +58 +00:02:18,000 --> 00:02:22,200 +instead of pleas um and that would be a + +59 +00:02:20,400 --> 00:02:24,200 +problem in a customer service system + +60 +00:02:22,200 --> 00:02:28,400 +right because your customer would uh + +61 +00:02:24,200 --> 00:02:28,400 +leave and never come back + +62 +00:02:28,840 --> 00:02:32,040 +so + +63 +00:02:30,360 --> 00:02:33,720 +determiner like this is not going to + +64 +00:02:32,040 --> 00:02:35,640 +cause a huge issue U messing up other + +65 +00:02:33,720 --> 00:02:37,519 +things is going to cause a larger + +66 +00:02:35,640 --> 00:02:39,519 +issue but from the point of view of + +67 +00:02:37,519 --> 00:02:42,680 +Maximum likelihood all of these are just + +68 +00:02:39,519 --> 00:02:44,560 +tokens and messing up one token is the + +69 +00:02:42,680 --> 00:02:47,519 +same as messing up another token so + +70 +00:02:44,560 --> 00:02:50,040 +that's uh you know an + +71 +00:02:47,519 --> 00:02:52,080 +issue another problem is that the gold + +72 +00:02:50,040 --> 00:02:54,640 +standard and maximum likelihood + +73 +00:02:52,080 --> 00:02:57,480 +estimation can be bad it can be like not + +74 +00:02:54,640 --> 00:02:59,239 +what you want and uh corpa are full of + +75 +00:02:57,480 --> 00:03:02,400 +outputs that we wouldn't want a language + +76 +00:02:59,239 --> 00:03:05,400 +model producing so for example uh toxic + +77 +00:03:02,400 --> 00:03:07,799 +comments on Reddit uh + +78 +00:03:05,400 --> 00:03:09,959 +disinformation um another thing that a + +79 +00:03:07,799 --> 00:03:13,000 +lot of people don't think about uh quite + +80 +00:03:09,959 --> 00:03:15,640 +as much is a lot of the data online is + +81 +00:03:13,000 --> 00:03:17,680 +uh from is automatically generated + +82 +00:03:15,640 --> 00:03:19,720 +nowadays for example from machine + +83 +00:03:17,680 --> 00:03:24,080 +translation a lot of the translations + +84 +00:03:19,720 --> 00:03:25,720 +online are from uh 2016 Google translate + +85 +00:03:24,080 --> 00:03:27,560 +uh when Google translate was a lot less + +86 +00:03:25,720 --> 00:03:29,120 +good than it is now and so you have like + +87 +00:03:27,560 --> 00:03:31,760 +poor quality translations that were + +88 +00:03:29,120 --> 00:03:31,760 +automatically + +89 +00:03:33,040 --> 00:03:37,959 +a final problem is uh something that's + +90 +00:03:35,280 --> 00:03:40,360 +called exposure bias and exposure bias + +91 +00:03:37,959 --> 00:03:44,000 +basically what it means is mle training + +92 +00:03:40,360 --> 00:03:46,000 +doesn't consider um the necessarity the + +93 +00:03:44,000 --> 00:03:48,599 +necessity for generation and it relies + +94 +00:03:46,000 --> 00:03:51,360 +on gold standard context so if we go + +95 +00:03:48,599 --> 00:03:54,159 +back to the mle equation when we're + +96 +00:03:51,360 --> 00:03:57,200 +calculating mle this y less than T is + +97 +00:03:54,159 --> 00:03:59,200 +always correct it's always a good output + +98 +00:03:57,200 --> 00:04:01,439 +and so what the model does is it learns + +99 +00:03:59,200 --> 00:04:04,280 +to over rely on good + +100 +00:04:01,439 --> 00:04:06,079 +outputs and one example of a problem + +101 +00:04:04,280 --> 00:04:08,360 +that this causes is models tend to + +102 +00:04:06,079 --> 00:04:10,560 +repeat themselves over and over again + +103 +00:04:08,360 --> 00:04:12,319 +for example um when you use some + +104 +00:04:10,560 --> 00:04:15,079 +generation algorithms and the reason why + +105 +00:04:12,319 --> 00:04:18,519 +this happens is because in a gold + +106 +00:04:15,079 --> 00:04:22,079 +standard output if a word has appeared + +107 +00:04:18,519 --> 00:04:25,840 +previously that word is more likely to + +108 +00:04:22,079 --> 00:04:28,560 +happen next so like if you say um like I + +109 +00:04:25,840 --> 00:04:29,759 +am going um I am going to Pittsburgh + +110 +00:04:28,560 --> 00:04:31,880 +you're much more likely to say + +111 +00:04:29,759 --> 00:04:33,000 +Pittsburgh again in the future because + +112 +00:04:31,880 --> 00:04:35,720 +you're talking about Pittsburgh + +113 +00:04:33,000 --> 00:04:37,400 +topically as coherent so what you get is + +114 +00:04:35,720 --> 00:04:38,639 +you get mle trained models saying I'm + +115 +00:04:37,400 --> 00:04:40,160 +going to Pittsburgh I am going to + +116 +00:04:38,639 --> 00:04:41,680 +Pittsburgh I am going to Pittsburgh I + +117 +00:04:40,160 --> 00:04:45,280 +going to Pittsburgh you've probably seen + +118 +00:04:41,680 --> 00:04:47,320 +this before uh at some point and so um + +119 +00:04:45,280 --> 00:04:49,320 +exposure bias is basically that the + +120 +00:04:47,320 --> 00:04:51,039 +model has never been exposed to mistakes + +121 +00:04:49,320 --> 00:04:55,240 +in the past and so it can't deal with + +122 +00:04:51,039 --> 00:04:56,840 +them so what this does is um if you have + +123 +00:04:55,240 --> 00:04:58,560 +an alternative training algorithm you + +124 +00:04:56,840 --> 00:05:02,120 +can fix this by generating a whole bunch + +125 +00:04:58,560 --> 00:05:04,880 +of outputs uh down like scoring some of + +126 +00:05:02,120 --> 00:05:06,880 +them poorly and penalizing the model for + +127 +00:05:04,880 --> 00:05:09,960 +uh generating po outputs and so that can + +128 +00:05:06,880 --> 00:05:09,960 +fix these problems as + +129 +00:05:10,800 --> 00:05:18,440 +well uh any questions about this all + +130 +00:05:15,199 --> 00:05:20,800 +good Okay cool so now I'd like to get + +131 +00:05:18,440 --> 00:05:23,919 +into how we measure how good an output + +132 +00:05:20,800 --> 00:05:26,360 +is and there's different ways of doing + +133 +00:05:23,919 --> 00:05:30,319 +this um the first one is objective + +134 +00:05:26,360 --> 00:05:32,680 +assessment so for some uh tasks or for + +135 +00:05:30,319 --> 00:05:35,400 +many tasks there's kind of objectively a + +136 +00:05:32,680 --> 00:05:37,280 +correct answer there's also human + +137 +00:05:35,400 --> 00:05:40,360 +subjective annotations so you can ask + +138 +00:05:37,280 --> 00:05:42,919 +humans to do annotation for you there's + +139 +00:05:40,360 --> 00:05:45,400 +machine prediction of human + +140 +00:05:42,919 --> 00:05:48,319 +preferences and there's also use in + +141 +00:05:45,400 --> 00:05:50,840 +another system in a downstream + +142 +00:05:48,319 --> 00:05:52,960 +task so the way objective assessment + +143 +00:05:50,840 --> 00:05:54,919 +works is you have an annotated correct + +144 +00:05:52,960 --> 00:05:57,080 +answer in match against this so like if + +145 +00:05:54,919 --> 00:06:00,600 +you're solving math problems uh + +146 +00:05:57,080 --> 00:06:02,560 +answering objective questions and and + +147 +00:06:00,600 --> 00:06:04,280 +you know you can pick any arbitrary + +148 +00:06:02,560 --> 00:06:06,840 +example you can pick your classification + +149 +00:06:04,280 --> 00:06:09,800 +example from uh like your text + +150 +00:06:06,840 --> 00:06:11,880 +classification tasks an even clearer + +151 +00:06:09,800 --> 00:06:13,880 +example is if you have math problems + +152 +00:06:11,880 --> 00:06:15,639 +there's kind of objectively one answer + +153 +00:06:13,880 --> 00:06:18,080 +to any math problem and there's no other + +154 +00:06:15,639 --> 00:06:19,680 +answer that could be correct so this + +155 +00:06:18,080 --> 00:06:21,160 +makes your life easy if you're handling + +156 +00:06:19,680 --> 00:06:22,560 +this type of problem but of course + +157 +00:06:21,160 --> 00:06:24,120 +there's many other types of problems we + +158 +00:06:22,560 --> 00:06:26,039 +want to handle that don't have objective + +159 +00:06:24,120 --> 00:06:29,039 +answers like + +160 +00:06:26,039 --> 00:06:31,440 +this so let's say we're handling a gener + +161 +00:06:29,039 --> 00:06:34,680 +a generation task where we don't have an + +162 +00:06:31,440 --> 00:06:36,360 +objective answer um in this Cas kind of + +163 +00:06:34,680 --> 00:06:39,440 +one of our gold standards is human + +164 +00:06:36,360 --> 00:06:42,360 +evaluation so we might have a source + +165 +00:06:39,440 --> 00:06:44,919 +input like a prompt or an input text for + +166 +00:06:42,360 --> 00:06:47,240 +machine translation we have one or + +167 +00:06:44,919 --> 00:06:49,960 +several hypotheses and we ask a human + +168 +00:06:47,240 --> 00:06:53,280 +annotator to basically give uh a score + +169 +00:06:49,960 --> 00:06:55,759 +for them or do some sort of other + +170 +00:06:53,280 --> 00:06:59,759 +annotation and the different varieties + +171 +00:06:55,759 --> 00:07:03,080 +of annotation that we can give are um + +172 +00:06:59,759 --> 00:07:04,599 +something called direct assessment so uh + +173 +00:07:03,080 --> 00:07:06,599 +direct assessment is a term that comes + +174 +00:07:04,599 --> 00:07:09,280 +from machine translation uh so you might + +175 +00:07:06,599 --> 00:07:11,039 +not see it used uh lots of other places + +176 +00:07:09,280 --> 00:07:13,120 +but it's basically just give a score + +177 +00:07:11,039 --> 00:07:15,759 +directly to how good the output is so + +178 +00:07:13,120 --> 00:07:17,199 +you can say like if you say please send + +179 +00:07:15,759 --> 00:07:18,960 +this translation is please send this + +180 +00:07:17,199 --> 00:07:21,759 +package to Tokyo we give it a score of + +181 +00:07:18,960 --> 00:07:24,360 +two out of 10 or something like + +182 +00:07:21,759 --> 00:07:28,000 +this + +183 +00:07:24,360 --> 00:07:30,840 +so the the question here is like what + +184 +00:07:28,000 --> 00:07:32,400 +does like let's say I gave a score of + +185 +00:07:30,840 --> 00:07:34,520 +two out of 10 for please send this + +186 +00:07:32,400 --> 00:07:37,680 +package to Tokyo what score should I + +187 +00:07:34,520 --> 00:07:40,240 +give for please send a package to Tokyo + +188 +00:07:37,680 --> 00:07:42,360 +anyone have any ideas the the correct + +189 +00:07:40,240 --> 00:07:46,520 +answer is please send this package to + +190 +00:07:42,360 --> 00:07:48,000 +take out of eight out of 10 yeah but you + +191 +00:07:46,520 --> 00:07:50,440 +might disagree on that right it's kind + +192 +00:07:48,000 --> 00:07:52,159 +of like subjective um one of the + +193 +00:07:50,440 --> 00:07:54,039 +difficulties of direct assessment is + +194 +00:07:52,159 --> 00:07:55,520 +giving a number like this is pretty + +195 +00:07:54,039 --> 00:07:57,800 +difficult if you don't have a very clear + +196 +00:07:55,520 --> 00:07:59,720 +rubric and very skilled annotators and + +197 +00:07:57,800 --> 00:08:02,879 +it's hard to get consistency between + +198 +00:07:59,720 --> 00:08:04,400 +people when you do this so the advantage + +199 +00:08:02,879 --> 00:08:05,599 +is it kind of gives you an idea of how + +200 +00:08:04,400 --> 00:08:07,520 +good things are overall but the + +201 +00:08:05,599 --> 00:08:09,280 +disadvantage is it's more difficult to + +202 +00:08:07,520 --> 00:08:11,319 +annotate and get + +203 +00:08:09,280 --> 00:08:13,159 +consistency um another thing that I + +204 +00:08:11,319 --> 00:08:15,319 +should point out is often scores are + +205 +00:08:13,159 --> 00:08:18,680 +assigned separately based on desirable + +206 +00:08:15,319 --> 00:08:20,960 +traits so um we don't necessarily just + +207 +00:08:18,680 --> 00:08:23,479 +say how good is it we say how fluent is + +208 +00:08:20,960 --> 00:08:26,120 +it like is it fluent uh + +209 +00:08:23,479 --> 00:08:28,159 +English in Translation there's a concept + +210 +00:08:26,120 --> 00:08:30,720 +called adequacy which is how well does + +211 +00:08:28,159 --> 00:08:34,599 +the output reflect the input + +212 +00:08:30,720 --> 00:08:36,519 +semantics um and if you're assessing + +213 +00:08:34,599 --> 00:08:38,440 +translation systems actually it's common + +214 +00:08:36,519 --> 00:08:40,519 +to assess fluency without even looking + +215 +00:08:38,440 --> 00:08:43,200 +at the input because then you can just + +216 +00:08:40,519 --> 00:08:44,880 +say how fluent is it but for adequacy + +217 +00:08:43,200 --> 00:08:46,320 +you definitely need to understand the + +218 +00:08:44,880 --> 00:08:49,600 +input so you need to be a bilingual + +219 +00:08:46,320 --> 00:08:54,680 +speaker to be able to assess + +220 +00:08:49,600 --> 00:08:57,560 +that um factuality um and so factuality + +221 +00:08:54,680 --> 00:09:00,160 +is tricky um it can either be factuality + +222 +00:08:57,560 --> 00:09:03,880 +grounded in a particular input text in + +223 +00:09:00,160 --> 00:09:05,600 +which case um the facts would have to be + +224 +00:09:03,880 --> 00:09:07,680 +you know things that were said in the + +225 +00:09:05,600 --> 00:09:09,399 +input or it can be just kind of is the + +226 +00:09:07,680 --> 00:09:11,120 +statement factual in general in which + +227 +00:09:09,399 --> 00:09:13,720 +case you need to go online you need to + +228 +00:09:11,120 --> 00:09:16,480 +search for things and like uh check + +229 +00:09:13,720 --> 00:09:18,480 +whether the statement is factual or not + +230 +00:09:16,480 --> 00:09:20,480 +um other things are like coherence does + +231 +00:09:18,480 --> 00:09:21,480 +the output fit coherently within the + +232 +00:09:20,480 --> 00:09:23,680 +larger + +233 +00:09:21,480 --> 00:09:25,680 +discs um and there's many many other + +234 +00:09:23,680 --> 00:09:28,120 +ones of these this is also task + +235 +00:09:25,680 --> 00:09:29,760 +dependent so like the things you will + +236 +00:09:28,120 --> 00:09:31,000 +evaluate for machine transl are + +237 +00:09:29,760 --> 00:09:32,880 +different than the ones you would do for + +238 +00:09:31,000 --> 00:09:35,760 +dialog which are different than the ones + +239 +00:09:32,880 --> 00:09:38,200 +you would do for a general purpose + +240 +00:09:35,760 --> 00:09:41,279 +chatot uh which is different kind things + +241 +00:09:38,200 --> 00:09:44,120 +you would do for um summarization for + +242 +00:09:41,279 --> 00:09:46,320 +example so if you're interested in doing + +243 +00:09:44,120 --> 00:09:47,519 +something like this uh then I definitely + +244 +00:09:46,320 --> 00:09:48,800 +encourage you to look at what other + +245 +00:09:47,519 --> 00:09:51,399 +people have done for the tasks you're + +246 +00:09:48,800 --> 00:09:53,079 +interested in uh previously and uh find + +247 +00:09:51,399 --> 00:09:54,880 +out the different types of traits that + +248 +00:09:53,079 --> 00:09:58,320 +did + +249 +00:09:54,880 --> 00:10:00,760 +last uh any any questions about this + +250 +00:09:58,320 --> 00:10:03,079 +also + +251 +00:10:00,760 --> 00:10:06,920 +okay the next type of feedback is + +252 +00:10:03,079 --> 00:10:09,839 +preference ratings um and so this is uh + +253 +00:10:06,920 --> 00:10:12,600 +basically what you do is you have two or + +254 +00:10:09,839 --> 00:10:14,240 +more outputs from different models or + +255 +00:10:12,600 --> 00:10:16,440 +different Generations from an individual + +256 +00:10:14,240 --> 00:10:18,839 +model and you ask a human which one is + +257 +00:10:16,440 --> 00:10:22,320 +better like is one better than the other + +258 +00:10:18,839 --> 00:10:23,839 +or are they tied and so in this case um + +259 +00:10:22,320 --> 00:10:26,320 +you might have please send this package + +260 +00:10:23,839 --> 00:10:28,880 +to Tokyo please send a package to + +261 +00:10:26,320 --> 00:10:31,040 +Tokyo we might disagree on how like good + +262 +00:10:28,880 --> 00:10:33,959 +or bad each of them are but I think most + +263 +00:10:31,040 --> 00:10:35,959 +people would agree that this one is like + +264 +00:10:33,959 --> 00:10:37,480 +despite the fact that it got this wrong + +265 +00:10:35,959 --> 00:10:40,160 +the second one is better than the first + +266 +00:10:37,480 --> 00:10:42,240 +one so this is a little bit of an easier + +267 +00:10:40,160 --> 00:10:45,040 +task it's easier to uh get people to + +268 +00:10:42,240 --> 00:10:46,839 +annotate these things + +269 +00:10:45,040 --> 00:10:50,519 +consistently however it has the + +270 +00:10:46,839 --> 00:10:52,839 +disadvantage that you can't really tell + +271 +00:10:50,519 --> 00:10:55,360 +uh whether systems are really good or + +272 +00:10:52,839 --> 00:10:57,200 +really bad so let's say you have a bunch + +273 +00:10:55,360 --> 00:11:00,279 +of really bad systems that you're + +274 +00:10:57,200 --> 00:11:01,839 +comparing with each other um you might + +275 +00:11:00,279 --> 00:11:03,680 +find that one is better than the other + +276 +00:11:01,839 --> 00:11:06,000 +but that still doesn't mean it's ready + +277 +00:11:03,680 --> 00:11:07,399 +to be deployed or if you have a bunch of + +278 +00:11:06,000 --> 00:11:11,040 +really good systems they're all + +279 +00:11:07,399 --> 00:11:13,000 +basically you know very very similar to + +280 +00:11:11,040 --> 00:11:14,399 +another but one is like slightly more + +281 +00:11:13,000 --> 00:11:18,639 +fluent than the other you might still + +282 +00:11:14,399 --> 00:11:20,680 +get a similar result um and so that also + +283 +00:11:18,639 --> 00:11:22,760 +makes it uh you know a little bit + +284 +00:11:20,680 --> 00:11:24,880 +difficult to use practically in some + +285 +00:11:22,760 --> 00:11:27,040 +ways I didn't put it on the slide but + +286 +00:11:24,880 --> 00:11:30,680 +there's another way you can kind of get + +287 +00:11:27,040 --> 00:11:33,920 +the best of both worlds um which is a + +288 +00:11:30,680 --> 00:11:35,560 +side by side assessment and side by-side + +289 +00:11:33,920 --> 00:11:38,440 +assessment basically what you would do + +290 +00:11:35,560 --> 00:11:40,560 +is you would say um please send this + +291 +00:11:38,440 --> 00:11:43,399 +package to Tokyo please send a package + +292 +00:11:40,560 --> 00:11:47,279 +to Pittsburgh give each of them a direct + +293 +00:11:43,399 --> 00:11:48,839 +score um but you can use decimal places + +294 +00:11:47,279 --> 00:11:51,120 +and you can't use the same score for all + +295 +00:11:48,839 --> 00:11:55,920 +of them and so it's + +296 +00:11:51,120 --> 00:11:57,480 +like five 500 and 4.99 out of five or + +297 +00:11:55,920 --> 00:11:59,519 +something like that like you like one + +298 +00:11:57,480 --> 00:12:02,639 +slightly better than the other or or + +299 +00:11:59,519 --> 00:12:04,480 +something like that um so there are ways + +300 +00:12:02,639 --> 00:12:07,240 +to kind of get Best of Both Worlds if + +301 +00:12:04,480 --> 00:12:11,720 +you're interested in doing + +302 +00:12:07,240 --> 00:12:11,720 +that um + +303 +00:12:14,920 --> 00:12:20,519 +so one problem one other problem with + +304 +00:12:18,279 --> 00:12:22,519 +preference rankings is that there's a + +305 +00:12:20,519 --> 00:12:24,440 +limited number of things that humans can + +306 +00:12:22,519 --> 00:12:28,160 +compare before they get really + +307 +00:12:24,440 --> 00:12:32,360 +overwhelmed so if you say I + +308 +00:12:28,160 --> 00:12:35,560 +want like I want to + +309 +00:12:32,360 --> 00:12:36,920 +rate 15 systems or 20 systems with + +310 +00:12:35,560 --> 00:12:39,120 +respect to how good they are with + +311 +00:12:36,920 --> 00:12:40,639 +respect to each other it's going to be + +312 +00:12:39,120 --> 00:12:43,680 +impossible for humans to come up with a + +313 +00:12:40,639 --> 00:12:46,959 +good preference ranking between them and + +314 +00:12:43,680 --> 00:12:49,480 +so the typical way around this um which + +315 +00:12:46,959 --> 00:12:52,360 +is also used in uh things like the + +316 +00:12:49,480 --> 00:12:55,440 +chatbot Arena by lmis and other things + +317 +00:12:52,360 --> 00:12:58,720 +like this is to use uh something like an + +318 +00:12:55,440 --> 00:13:00,959 +ELO or true skill ranking and what these + +319 +00:12:58,720 --> 00:13:03,079 +are is these are things that were + +320 +00:13:00,959 --> 00:13:05,760 +created for the ranking of like chess + +321 +00:13:03,079 --> 00:13:09,160 +players or video game players or other + +322 +00:13:05,760 --> 00:13:11,720 +things where they like b battle against + +323 +00:13:09,160 --> 00:13:13,920 +each other in multiple matches uh + +324 +00:13:11,720 --> 00:13:16,440 +pair-wise and then you put all of the + +325 +00:13:13,920 --> 00:13:18,399 +wins and losses into these ranking + +326 +00:13:16,440 --> 00:13:20,600 +algorithms and they give you a score + +327 +00:13:18,399 --> 00:13:22,920 +about how good like each of the each of + +328 +00:13:20,600 --> 00:13:27,079 +the players are so if you do something + +329 +00:13:22,920 --> 00:13:29,480 +like this you can um get basically a + +330 +00:13:27,079 --> 00:13:32,120 +ranking of systems despite the that you + +331 +00:13:29,480 --> 00:13:35,240 +only did pairwise assessments so these + +332 +00:13:32,120 --> 00:13:35,240 +are also a good thing to know + +333 +00:13:37,399 --> 00:13:43,839 +about a final variety of human feedback + +334 +00:13:40,600 --> 00:13:45,320 +uh that we create is uh air annotation + +335 +00:13:43,839 --> 00:13:47,519 +and this can be useful for a number of + +336 +00:13:45,320 --> 00:13:49,839 +reasons um but basically the way it + +337 +00:13:47,519 --> 00:13:53,839 +works is you annotate individual errors + +338 +00:13:49,839 --> 00:13:55,639 +within the outputs and um oh one thing I + +339 +00:13:53,839 --> 00:13:58,120 +should mention is that um I'm giving a + +340 +00:13:55,639 --> 00:14:00,880 +lot of examples from machine translation + +341 +00:13:58,120 --> 00:14:02,800 +um I feel like machine translation has + +342 +00:14:00,880 --> 00:14:04,519 +been doing evaluation of generated + +343 +00:14:02,800 --> 00:14:07,600 +outputs for a lot longer than a lot of + +344 +00:14:04,519 --> 00:14:09,000 +other uh fields of NLP have and + +345 +00:14:07,600 --> 00:14:11,800 +therefore their methodology is more + +346 +00:14:09,000 --> 00:14:13,480 +developed than a lot of other fields um + +347 +00:14:11,800 --> 00:14:16,199 +but a lot of these things can also be + +348 +00:14:13,480 --> 00:14:18,079 +applied to uh other uh other tasks as + +349 +00:14:16,199 --> 00:14:19,079 +well but anyway getting back to this + +350 +00:14:18,079 --> 00:14:20,680 +there's something for machine + +351 +00:14:19,079 --> 00:14:23,639 +translation called multi-dimensional + +352 +00:14:20,680 --> 00:14:26,240 +quality metrics and the multidimensional + +353 +00:14:23,639 --> 00:14:29,160 +quality metrics basically what they do + +354 +00:14:26,240 --> 00:14:32,199 +is they annotate spans in the output + +355 +00:14:29,160 --> 00:14:34,800 +where each Span in the output is given a + +356 +00:14:32,199 --> 00:14:38,079 +severity ranking of the error and it's + +357 +00:14:34,800 --> 00:14:40,199 +given a type of the error and there's + +358 +00:14:38,079 --> 00:14:42,600 +about eight different types of Errors + +359 +00:14:40,199 --> 00:14:44,839 +like this doesn't violate or this + +360 +00:14:42,600 --> 00:14:47,399 +violates linguistic conventions of using + +361 +00:14:44,839 --> 00:14:49,880 +the word this instead of uh here by + +362 +00:14:47,399 --> 00:14:51,639 +using the word uh instead of this here + +363 +00:14:49,880 --> 00:14:55,079 +and then this is an accuracy error + +364 +00:14:51,639 --> 00:14:57,839 +because it's not accurately con uh uh + +365 +00:14:55,079 --> 00:15:01,720 +conveying the output and then this error + +366 +00:14:57,839 --> 00:15:04,600 +is minor uh this error is Major um and + +367 +00:15:01,720 --> 00:15:06,399 +then there's also like severe severe + +368 +00:15:04,600 --> 00:15:07,440 +versus major but minor and major is a + +369 +00:15:06,399 --> 00:15:09,680 +more important + +370 +00:15:07,440 --> 00:15:11,839 +distinction um so the advantage of this + +371 +00:15:09,680 --> 00:15:14,279 +is a couple fold number one it gives you + +372 +00:15:11,839 --> 00:15:16,440 +more fine grained feedback uh in that + +373 +00:15:14,279 --> 00:15:19,199 +you can say okay this system has a lot + +374 +00:15:16,440 --> 00:15:22,199 +of uh accuracy errors this system has a + +375 +00:15:19,199 --> 00:15:24,880 +lot of linguistic conventions errors um + +376 +00:15:22,199 --> 00:15:28,600 +it also can be more consistent because + +377 +00:15:24,880 --> 00:15:29,839 +if you just say to people which output + +378 +00:15:28,600 --> 00:15:31,800 +is better + +379 +00:15:29,839 --> 00:15:34,560 +or what is the score of this output + +380 +00:15:31,800 --> 00:15:36,360 +people have trouble deciding about that + +381 +00:15:34,560 --> 00:15:39,560 +because it's a more subjective + +382 +00:15:36,360 --> 00:15:41,680 +evaluation but if I say is this word + +383 +00:15:39,560 --> 00:15:43,000 +correct it's a little bit easier for + +384 +00:15:41,680 --> 00:15:44,759 +people to do so you can get more + +385 +00:15:43,000 --> 00:15:46,920 +consistent annotations + +386 +00:15:44,759 --> 00:15:49,720 +here the problem with this is this can + +387 +00:15:46,920 --> 00:15:50,839 +be very time consuming so um you know + +388 +00:15:49,720 --> 00:15:52,480 +obviously you need to go through and + +389 +00:15:50,839 --> 00:15:56,440 +annotate every single error if it's for + +390 +00:15:52,480 --> 00:15:56,440 +a long outputs or something your + +391 +00:15:56,959 --> 00:16:03,519 +problem so anyway these are just three + +392 +00:15:59,800 --> 00:16:05,680 +uh ways of collecting human feedback um + +393 +00:16:03,519 --> 00:16:08,639 +and then there's an alternative which is + +394 +00:16:05,680 --> 00:16:10,079 +automatic evaluation of outputs and um + +395 +00:16:08,639 --> 00:16:14,399 +there's a bunch of different ways we can + +396 +00:16:10,079 --> 00:16:16,800 +do this the basic idea here is we have a + +397 +00:16:14,399 --> 00:16:20,199 +source um we have a couple + +398 +00:16:16,800 --> 00:16:22,800 +hypotheses and uh we have an automatic + +399 +00:16:20,199 --> 00:16:26,000 +system that generates outputs uh like + +400 +00:16:22,800 --> 00:16:28,279 +scores and we optionally have a + +401 +00:16:26,000 --> 00:16:30,839 +reference output so the reference output + +402 +00:16:28,279 --> 00:16:33,519 +is a human created gold standard output + +403 +00:16:30,839 --> 00:16:35,120 +with respect to how good that um uh with + +404 +00:16:33,519 --> 00:16:38,240 +respect to like what the output should + +405 +00:16:35,120 --> 00:16:38,240 +be in an ideal + +406 +00:16:38,279 --> 00:16:47,079 +case and basically the goal of automatic + +407 +00:16:43,199 --> 00:16:50,199 +evaluation is to + +408 +00:16:47,079 --> 00:16:52,839 +predict human preferences or to predict + +409 +00:16:50,199 --> 00:16:56,240 +what the human scores would be um + +410 +00:16:52,839 --> 00:16:58,600 +because still at this point um we mostly + +411 +00:16:56,240 --> 00:16:59,480 +view what humans think of the output to + +412 +00:16:58,600 --> 00:17:01,680 +be + +413 +00:16:59,480 --> 00:17:03,280 +uh kind of the + +414 +00:17:01,680 --> 00:17:06,199 +standard + +415 +00:17:03,280 --> 00:17:08,439 +and this is called a variety of things + +416 +00:17:06,199 --> 00:17:10,600 +depending on what field you're in um in + +417 +00:17:08,439 --> 00:17:12,559 +machine translation and summarization + +418 +00:17:10,600 --> 00:17:13,520 +it's called automatic evaluation also a + +419 +00:17:12,559 --> 00:17:16,520 +lot in + +420 +00:17:13,520 --> 00:17:18,400 +dialogue um if you're talking about + +421 +00:17:16,520 --> 00:17:21,000 +people from reinforcement learning or + +422 +00:17:18,400 --> 00:17:24,600 +other things um or chat Bots or things + +423 +00:17:21,000 --> 00:17:28,240 +like that uh a lot of people or uh like + +424 +00:17:24,600 --> 00:17:31,280 +AGI or whatever um a lot of people call + +425 +00:17:28,240 --> 00:17:32,520 +it uh word model um because that + +426 +00:17:31,280 --> 00:17:34,480 +specifically comes from the point of + +427 +00:17:32,520 --> 00:17:36,440 +view of like learning from this feedback + +428 +00:17:34,480 --> 00:17:37,960 +but essentially they're the same thing + +429 +00:17:36,440 --> 00:17:41,080 +uh from my point of view they're trying + +430 +00:17:37,960 --> 00:17:42,520 +to predict how good an output is and how + +431 +00:17:41,080 --> 00:17:44,240 +much you should reward the model for + +432 +00:17:42,520 --> 00:17:46,559 +producing that + +433 +00:17:44,240 --> 00:17:48,679 +output + +434 +00:17:46,559 --> 00:17:50,520 +um so there's a bunch of different + +435 +00:17:48,679 --> 00:17:51,720 +methods to do this I'm not going to + +436 +00:17:50,520 --> 00:17:53,799 +cover all of them I'm just going to + +437 +00:17:51,720 --> 00:17:55,240 +cover three paradigms for doing this so + +438 +00:17:53,799 --> 00:17:57,880 +you know where to look further if you're + +439 +00:17:55,240 --> 00:18:00,039 +interested in doing these things um the + +440 +00:17:57,880 --> 00:18:02,400 +first one is embedding based + +441 +00:18:00,039 --> 00:18:04,679 +evaluation and the way embedding based + +442 +00:18:02,400 --> 00:18:06,600 +evaluation works is usually it's + +443 +00:18:04,679 --> 00:18:11,400 +unsupervised calculation based on + +444 +00:18:06,600 --> 00:18:14,880 +embeding similarity between um + +445 +00:18:11,400 --> 00:18:18,080 +the output that the model generated and + +446 +00:18:14,880 --> 00:18:20,840 +a reference output that uh you have + +447 +00:18:18,080 --> 00:18:23,400 +created so sorry this is very small but + +448 +00:18:20,840 --> 00:18:25,559 +we have a reference here that says the + +449 +00:18:23,400 --> 00:18:27,640 +weather is cold today and we have a + +450 +00:18:25,559 --> 00:18:30,240 +candidate that says it is freezing today + +451 +00:18:27,640 --> 00:18:33,000 +so this is probably you know like a good + +452 +00:18:30,240 --> 00:18:35,480 +um a reasonably good + +453 +00:18:33,000 --> 00:18:37,640 +output and we run this through some + +454 +00:18:35,480 --> 00:18:39,120 +embedding model uh it was called Bert + +455 +00:18:37,640 --> 00:18:40,679 +score and so of course you can run it + +456 +00:18:39,120 --> 00:18:42,240 +through Bert but basically it can be any + +457 +00:18:40,679 --> 00:18:43,799 +embedding model that gives you embedding + +458 +00:18:42,240 --> 00:18:46,200 +for each token in the + +459 +00:18:43,799 --> 00:18:47,640 +sequence and so there are five tokens in + +460 +00:18:46,200 --> 00:18:49,720 +this sequence four tokens in this + +461 +00:18:47,640 --> 00:18:51,960 +sequence you get five tokens and then + +462 +00:18:49,720 --> 00:18:54,799 +four sorry five embeddings and then four + +463 +00:18:51,960 --> 00:18:57,400 +embeddings you calculate carewise cosine + +464 +00:18:54,799 --> 00:18:59,880 +similarity between all of them and this + +465 +00:18:57,400 --> 00:19:03,480 +gives you cosine + +466 +00:18:59,880 --> 00:19:06,480 +similarity Matrix and then you take the + +467 +00:19:03,480 --> 00:19:09,120 +ARG Max or you take the maximum + +468 +00:19:06,480 --> 00:19:11,280 +similarity along either the + +469 +00:19:09,120 --> 00:19:15,799 +rows or the + +470 +00:19:11,280 --> 00:19:19,559 +columns and here the rows correspond + +471 +00:19:15,799 --> 00:19:22,400 +to tokens in the reference and because + +472 +00:19:19,559 --> 00:19:24,039 +the rows correspond to tokens in the + +473 +00:19:22,400 --> 00:19:26,960 +reference + +474 +00:19:24,039 --> 00:19:28,320 +the how well you find something that is + +475 +00:19:26,960 --> 00:19:31,679 +similar to each of the tokens in the + +476 +00:19:28,320 --> 00:19:34,000 +reference is like a recall based method + +477 +00:19:31,679 --> 00:19:35,919 +because it's saying how many tokens in + +478 +00:19:34,000 --> 00:19:39,520 +the reference have a good match in the + +479 +00:19:35,919 --> 00:19:41,120 +output and then if you look at the + +480 +00:19:39,520 --> 00:19:42,799 +columns if you look at the max and the + +481 +00:19:41,120 --> 00:19:44,960 +columns this is like a precision based + +482 +00:19:42,799 --> 00:19:47,000 +metric because it's saying how many of + +483 +00:19:44,960 --> 00:19:49,360 +the things in the output are similar + +484 +00:19:47,000 --> 00:19:51,240 +have a similar match in the reference so + +485 +00:19:49,360 --> 00:19:54,480 +basically you can calculate recall and + +486 +00:19:51,240 --> 00:19:56,200 +precision over all of the tokens and + +487 +00:19:54,480 --> 00:20:00,200 +then feed this into something that looks + +488 +00:19:56,200 --> 00:20:02,400 +like fmeasure and you can also use tfidf + +489 +00:20:00,200 --> 00:20:06,000 +waiting um like what I talked about in + +490 +00:20:02,400 --> 00:20:07,799 +the rag lecture uh to upweight low + +491 +00:20:06,000 --> 00:20:09,520 +frequency words because low frequency + +492 +00:20:07,799 --> 00:20:11,440 +words tend to be more content words and + +493 +00:20:09,520 --> 00:20:13,120 +going back to my example you know if you + +494 +00:20:11,440 --> 00:20:14,280 +make a mistake from Pittsburgh to Tokyo + +495 +00:20:13,120 --> 00:20:17,880 +that's going to be more painful than + +496 +00:20:14,280 --> 00:20:21,000 +making a mistake from this to um so + +497 +00:20:17,880 --> 00:20:22,520 +actually if you'll uh if you were paying + +498 +00:20:21,000 --> 00:20:25,480 +close attention to the rag lecture this + +499 +00:20:22,520 --> 00:20:27,360 +looks really similar to the co bear um + +500 +00:20:25,480 --> 00:20:29,559 +the co bear retrieval objective that I + +501 +00:20:27,360 --> 00:20:30,960 +talked about in the r lecture um I don't + +502 +00:20:29,559 --> 00:20:32,840 +think it's a coincidence they both came + +503 +00:20:30,960 --> 00:20:34,360 +out around the same time uh so people + +504 +00:20:32,840 --> 00:20:36,360 +were thinking about the same thing but + +505 +00:20:34,360 --> 00:20:37,600 +um this is one method that's pretty + +506 +00:20:36,360 --> 00:20:40,200 +widely + +507 +00:20:37,600 --> 00:20:43,480 +use the bird Square code base is also + +508 +00:20:40,200 --> 00:20:45,440 +really nice and easy to use so um if uh + +509 +00:20:43,480 --> 00:20:47,640 +you want to try it out feel free to take + +510 +00:20:45,440 --> 00:20:47,640 +a + +511 +00:20:48,159 --> 00:20:53,840 +look cool um the next one I'd like to + +512 +00:20:51,600 --> 00:20:56,080 +talk about is a regression based + +513 +00:20:53,840 --> 00:20:58,760 +evaluation and the way this works is + +514 +00:20:56,080 --> 00:21:02,600 +this is usually used in a supervised uh + +515 +00:20:58,760 --> 00:21:04,320 +setting so uh the way what you have to + +516 +00:21:02,600 --> 00:21:07,600 +do is you have to calculate a whole + +517 +00:21:04,320 --> 00:21:09,799 +bunch of like actual human + +518 +00:21:07,600 --> 00:21:12,440 +judgments and + +519 +00:21:09,799 --> 00:21:15,000 +usually these judgments can either be + +520 +00:21:12,440 --> 00:21:16,960 +direct assessment uh where you actually + +521 +00:21:15,000 --> 00:21:19,120 +have a score or they can be pairwise + +522 +00:21:16,960 --> 00:21:20,840 +judgments and then if you have direct + +523 +00:21:19,120 --> 00:21:23,640 +assessment you use a regression based + +524 +00:21:20,840 --> 00:21:26,039 +loss like uh minimum squared error if + +525 +00:21:23,640 --> 00:21:27,520 +you have pairwise uh you use a ranking + +526 +00:21:26,039 --> 00:21:29,039 +based loss that tries to upweight the + +527 +00:21:27,520 --> 00:21:31,360 +ones that are higher scoring downward + +528 +00:21:29,039 --> 00:21:33,200 +the ones that are lower scoring one + +529 +00:21:31,360 --> 00:21:35,720 +typical example of this is Comet which + +530 +00:21:33,200 --> 00:21:37,200 +is or has been at least for a very long + +531 +00:21:35,720 --> 00:21:39,880 +time the state-of-the art and machine + +532 +00:21:37,200 --> 00:21:41,279 +translation evaluation and the reason + +533 +00:21:39,880 --> 00:21:43,440 +why it works so well is because we have + +534 +00:21:41,279 --> 00:21:44,720 +a bunch of evaluations for machine + +535 +00:21:43,440 --> 00:21:46,080 +translation they've been doing + +536 +00:21:44,720 --> 00:21:47,600 +evaluation and machine translation + +537 +00:21:46,080 --> 00:21:50,480 +systems for years and you can use that + +538 +00:21:47,600 --> 00:21:52,720 +as lots of supervised training data so + +539 +00:21:50,480 --> 00:21:54,640 +basically you just take um these + +540 +00:21:52,720 --> 00:21:56,440 +evaluation data you have human + +541 +00:21:54,640 --> 00:21:59,080 +annotations you have the output + +542 +00:21:56,440 --> 00:22:00,320 +according to a model like comet um you + +543 +00:21:59,080 --> 00:22:02,679 +calculate the difference between them + +544 +00:22:00,320 --> 00:22:05,640 +and you update model + +545 +00:22:02,679 --> 00:22:07,080 +parameters um the problem this is great + +546 +00:22:05,640 --> 00:22:08,520 +if you have lots of training data the + +547 +00:22:07,080 --> 00:22:10,640 +problem with this is for a lot of tasks + +548 +00:22:08,520 --> 00:22:12,360 +we don't have lots of training data so + +549 +00:22:10,640 --> 00:22:14,720 +um you know training these is a little + +550 +00:22:12,360 --> 00:22:14,720 +bit less + +551 +00:22:15,400 --> 00:22:22,919 +feasible and now recently uh what we + +552 +00:22:19,600 --> 00:22:25,279 +have been moving into is is a QA based + +553 +00:22:22,919 --> 00:22:27,120 +evaluation which is basically where we + +554 +00:22:25,279 --> 00:22:30,760 +ask a language model how good the output + +555 +00:22:27,120 --> 00:22:32,279 +is and so uh gmba is an example one of + +556 +00:22:30,760 --> 00:22:34,559 +the early examples of this for machine + +557 +00:22:32,279 --> 00:22:37,320 +translation evaluation uh where they + +558 +00:22:34,559 --> 00:22:39,840 +basically just ask a g gp4 like score + +559 +00:22:37,320 --> 00:22:41,600 +the following translation from Source + +560 +00:22:39,840 --> 00:22:44,000 +language to target language with respect + +561 +00:22:41,600 --> 00:22:47,080 +to the human reference um on a + +562 +00:22:44,000 --> 00:22:49,200 +continuous scale from Z to 100 uh where + +563 +00:22:47,080 --> 00:22:51,320 +the score of zero means no meaning + +564 +00:22:49,200 --> 00:22:54,039 +preserved and the score of 100 means a + +565 +00:22:51,320 --> 00:22:56,880 +perfect meaning in grammar uh you feed + +566 +00:22:54,039 --> 00:22:58,760 +in the source um you feed in the T the + +567 +00:22:56,880 --> 00:23:01,000 +human reference optionally if you have a + +568 +00:22:58,760 --> 00:23:03,320 +human reference and then you feed in the + +569 +00:23:01,000 --> 00:23:06,760 +Target um and you get a + +570 +00:23:03,320 --> 00:23:09,919 +score and um so this this works pretty + +571 +00:23:06,760 --> 00:23:12,720 +well this can give you uh better results + +572 +00:23:09,919 --> 00:23:15,159 +um there's a especially if you have a + +573 +00:23:12,720 --> 00:23:16,960 +strong language model the problem is + +574 +00:23:15,159 --> 00:23:18,279 +it's very unpredictable whether this is + +575 +00:23:16,960 --> 00:23:20,120 +going to work well and it's very + +576 +00:23:18,279 --> 00:23:23,039 +dependent on the prompt that you're + +577 +00:23:20,120 --> 00:23:25,279 +using so um right now A lot of people + +578 +00:23:23,039 --> 00:23:27,279 +are using gp4 without actually + +579 +00:23:25,279 --> 00:23:29,039 +validating whether it does a good job at + +580 +00:23:27,279 --> 00:23:33,080 +evaluation and + +581 +00:23:29,039 --> 00:23:34,919 +and my the results are all across the + +582 +00:23:33,080 --> 00:23:36,880 +board it can be anywhere from very very + +583 +00:23:34,919 --> 00:23:38,640 +good to very very bad at evaluating + +584 +00:23:36,880 --> 00:23:41,320 +particular tasks so I would be at least + +585 +00:23:38,640 --> 00:23:43,559 +a little bit suspicious of whether gp4 + +586 +00:23:41,320 --> 00:23:45,679 +is doing a good job evaluating for your + +587 +00:23:43,559 --> 00:23:49,320 +task especially more complex + +588 +00:23:45,679 --> 00:23:51,960 +tests um I would especially be + +589 +00:23:49,320 --> 00:23:54,000 +suspicious if you're doing two uh any of + +590 +00:23:51,960 --> 00:23:56,760 +the two following things number one if + +591 +00:23:54,000 --> 00:23:59,880 +you're comparing gp4 or any model + +592 +00:23:56,760 --> 00:24:02,400 +against itself in another model because + +593 +00:23:59,880 --> 00:24:05,200 +gp4 really likes + +594 +00:24:02,400 --> 00:24:06,880 +gp4 it really likes its own outputs and + +595 +00:24:05,200 --> 00:24:08,120 +there are papers uh sorry I don't + +596 +00:24:06,880 --> 00:24:09,679 +actually have the references here but I + +597 +00:24:08,120 --> 00:24:11,200 +can follow up if people are interested + +598 +00:24:09,679 --> 00:24:13,080 +but there are papers that demonstrate + +599 +00:24:11,200 --> 00:24:15,799 +that gp4 likes it you know its own + +600 +00:24:13,080 --> 00:24:19,200 +outputs more than others also if you're + +601 +00:24:15,799 --> 00:24:22,120 +explicitly optimizing the outputs using + +602 +00:24:19,200 --> 00:24:24,640 +rlf um there is something called good + +603 +00:24:22,120 --> 00:24:27,120 +Hearts law which is basically anytime + +604 +00:24:24,640 --> 00:24:29,520 +you uh start optimizing towards a metric + +605 +00:24:27,120 --> 00:24:32,559 +it becomes a bad metric and that also + +606 +00:24:29,520 --> 00:24:35,000 +happens for gp4 based evaluations so if + +607 +00:24:32,559 --> 00:24:37,200 +you start optimizing for gp4 based + +608 +00:24:35,000 --> 00:24:38,960 +evaluations especially for reference + +609 +00:24:37,200 --> 00:24:41,679 +list metrics that don't use a reference + +610 +00:24:38,960 --> 00:24:44,840 +output then um you start basically + +611 +00:24:41,679 --> 00:24:47,440 +exploiting the metric + +612 +00:24:44,840 --> 00:24:49,840 +um another thing that you can do with QA + +613 +00:24:47,440 --> 00:24:53,279 +based evaluation is ask about fine grade + +614 +00:24:49,840 --> 00:24:54,919 +mistakes and so this is a paper by um uh + +615 +00:24:53,279 --> 00:24:56,480 +Patrick Fernandez who's a student who's + +616 +00:24:54,919 --> 00:25:02,080 +working with me and basically what we + +617 +00:24:56,480 --> 00:25:05,240 +did is we asked the model to um not give + +618 +00:25:02,080 --> 00:25:07,360 +a particular score but actually identify + +619 +00:25:05,240 --> 00:25:08,880 +the mistakes in the output and when we + +620 +00:25:07,360 --> 00:25:10,559 +asked it to identify the mistakes in the + +621 +00:25:08,880 --> 00:25:13,720 +output we found that this gave more + +622 +00:25:10,559 --> 00:25:17,320 +consistent uh results so kind of + +623 +00:25:13,720 --> 00:25:18,840 +interestingly we ask humans to identify + +624 +00:25:17,320 --> 00:25:21,120 +individual mistakes and the output that + +625 +00:25:18,840 --> 00:25:24,240 +gives humans more consistent results + +626 +00:25:21,120 --> 00:25:25,559 +it's the same thing for gp4 so um that + +627 +00:25:24,240 --> 00:25:27,320 +that's another paper you can look at if + +628 +00:25:25,559 --> 00:25:29,640 +you're + +629 +00:25:27,320 --> 00:25:32,679 +interested + +630 +00:25:29,640 --> 00:25:38,000 +cool um so I I mentioned that you could + +631 +00:25:32,679 --> 00:25:38,000 +or could not uh trust uh yeah sorry go + +632 +00:25:44,679 --> 00:25:51,279 +ahead uh correct so yeah B basically + +633 +00:25:47,360 --> 00:25:53,279 +just what you do is you have the source + +634 +00:25:51,279 --> 00:25:54,960 +um ideally you'll also have a reference + +635 +00:25:53,279 --> 00:25:57,840 +output that was created by skilled + +636 +00:25:54,960 --> 00:25:59,720 +humans and then you put in the Target + +637 +00:25:57,840 --> 00:26:02,279 +you know output basically you have the + +638 +00:25:59,720 --> 00:26:08,000 +input ideally a reference output created + +639 +00:26:02,279 --> 00:26:08,000 +by Good by skilled humans and uh like + +640 +00:26:15,159 --> 00:26:20,240 +hypothesis yeah I + +641 +00:26:17,919 --> 00:26:24,559 +mean it's a good question and I don't + +642 +00:26:20,240 --> 00:26:26,919 +know if we actually have a a very clear + +643 +00:26:24,559 --> 00:26:31,399 +empirical like evidence of why this is + +644 +00:26:26,919 --> 00:26:33,320 +the case but my hypothesis about this is + +645 +00:26:31,399 --> 00:26:36,159 +yes we kind of would expect models to be + +646 +00:26:33,320 --> 00:26:38,200 +more biased towards their own outputs + +647 +00:26:36,159 --> 00:26:40,919 +and the reason why is because + +648 +00:26:38,200 --> 00:26:43,080 +essentially you know models + +649 +00:26:40,919 --> 00:26:44,279 +are within their embeddings they're + +650 +00:26:43,080 --> 00:26:45,760 +encoding when they're in a high + +651 +00:26:44,279 --> 00:26:47,600 +probability part of the space and when + +652 +00:26:45,760 --> 00:26:50,200 +they're in a low probability part of the + +653 +00:26:47,600 --> 00:26:51,120 +space and like the high probability part + +654 +00:26:50,200 --> 00:26:54,600 +of the + +655 +00:26:51,120 --> 00:26:56,200 +space is going to be the high + +656 +00:26:54,600 --> 00:26:58,600 +probability part of the space is going + +657 +00:26:56,200 --> 00:27:02,559 +to be associated with good outputs + +658 +00:26:58,600 --> 00:27:07,000 +because like when + +659 +00:27:02,559 --> 00:27:08,600 +models are more sure of their outputs + +660 +00:27:07,000 --> 00:27:11,960 +they're more likely to be + +661 +00:27:08,600 --> 00:27:13,520 +good just because that indicates that + +662 +00:27:11,960 --> 00:27:15,240 +like they're closer to the training data + +663 +00:27:13,520 --> 00:27:17,760 +that it had and other things like that + +664 +00:27:15,240 --> 00:27:21,600 +so model probabilities are associated + +665 +00:27:17,760 --> 00:27:23,760 +with outputs uh with uh with good + +666 +00:27:21,600 --> 00:27:26,600 +outputs but just + +667 +00:27:23,760 --> 00:27:29,440 +correla separately from + +668 +00:27:26,600 --> 00:27:32,120 +that I believe a model can identify when + +669 +00:27:29,440 --> 00:27:33,320 +it's in a high probability segment of + +670 +00:27:32,120 --> 00:27:35,799 +the space and when it's in a low + +671 +00:27:33,320 --> 00:27:39,399 +probability segment of the space and + +672 +00:27:35,799 --> 00:27:39,399 +because of that I expect + +673 +00:27:39,519 --> 00:27:45,519 +that I like there are segments of the + +674 +00:27:43,240 --> 00:27:47,120 +embedding space where it's more likely + +675 +00:27:45,519 --> 00:27:48,360 +to answer yes about something being good + +676 +00:27:47,120 --> 00:27:50,960 +or not and those are going to be + +677 +00:27:48,360 --> 00:27:54,760 +associated with high uh like high + +678 +00:27:50,960 --> 00:27:56,159 +probability outbreaks as well and also + +679 +00:27:54,760 --> 00:27:57,760 +models are more likely to generate + +680 +00:27:56,159 --> 00:28:00,240 +outputs that are high probability + +681 +00:27:57,760 --> 00:28:02,320 +according into their model by definition + +682 +00:28:00,240 --> 00:28:03,880 +so all three of those effects together + +683 +00:28:02,320 --> 00:28:05,640 +would basically go into a model being + +684 +00:28:03,880 --> 00:28:09,120 +bios supports its own outputs compared + +685 +00:28:05,640 --> 00:28:11,559 +to that puts in another model but um + +686 +00:28:09,120 --> 00:28:13,279 +yeah this is a very handwavy explanation + +687 +00:28:11,559 --> 00:28:15,519 +but like putting the two the three + +688 +00:28:13,279 --> 00:28:18,600 +together models output high probability + +689 +00:28:15,519 --> 00:28:20,880 +things from their own probability Space + +690 +00:28:18,600 --> 00:28:23,440 +by definition + +691 +00:28:20,880 --> 00:28:25,760 +um things that are high probability are + +692 +00:28:23,440 --> 00:28:27,519 +associated with being good uh just + +693 +00:28:25,760 --> 00:28:29,279 +because otherwise a model would be + +694 +00:28:27,519 --> 00:28:31,840 +outputting garbage + +695 +00:28:29,279 --> 00:28:33,840 +and um the final thing which is more + +696 +00:28:31,840 --> 00:28:35,679 +tenuous is if the model is in a high + +697 +00:28:33,840 --> 00:28:37,919 +probability segment of the space it's + +698 +00:28:35,679 --> 00:28:39,760 +more likely to Output yes according to a + +699 +00:28:37,919 --> 00:28:41,480 +question of it being good and I I think + +700 +00:28:39,760 --> 00:28:44,360 +that's probably true but I'm not 100% + +701 +00:28:41,480 --> 00:28:44,360 +sure about the the + +702 +00:28:45,559 --> 00:28:51,039 +fin um maybe maybe someone wants to + +703 +00:28:49,000 --> 00:28:52,840 +examinate examine that as a final + +704 +00:28:51,039 --> 00:28:54,200 +project it seems like a interesting + +705 +00:28:52,840 --> 00:28:57,080 +interesting + +706 +00:28:54,200 --> 00:29:00,039 +question um cool uh were there any other + +707 +00:28:57,080 --> 00:29:00,039 +questions about these methods + +708 +00:29:00,159 --> 00:29:07,120 +here um okay so when I say like an + +709 +00:29:03,960 --> 00:29:11,080 +evaluation metric is good or not what do + +710 +00:29:07,120 --> 00:29:13,200 +I mean by this being good or not um or a + +711 +00:29:11,080 --> 00:29:16,880 +reward model or whatever else and + +712 +00:29:13,200 --> 00:29:18,440 +basically the um the way we typically do + +713 +00:29:16,880 --> 00:29:19,840 +this is by doing something called meta + +714 +00:29:18,440 --> 00:29:22,440 +evaluation so it's called meta + +715 +00:29:19,840 --> 00:29:25,799 +evaluation because it's evaluation of + +716 +00:29:22,440 --> 00:29:29,279 +evaluation and uh the way we do this is + +717 +00:29:25,799 --> 00:29:32,519 +we have human uh scores and we have + +718 +00:29:29,279 --> 00:29:34,760 +automatic scores and we usually + +719 +00:29:32,519 --> 00:29:38,640 +calculate some sort of correlation + +720 +00:29:34,760 --> 00:29:41,000 +between the scores so um typical ones + +721 +00:29:38,640 --> 00:29:46,440 +are rank correlations like Pearson's + +722 +00:29:41,000 --> 00:29:48,799 +correlation or tendle uh Tow and uh so + +723 +00:29:46,440 --> 00:29:51,200 +the more Associated the automatic scores + +724 +00:29:48,799 --> 00:29:53,960 +are with the human scores the higher + +725 +00:29:51,200 --> 00:29:55,159 +these correlations are going to be um + +726 +00:29:53,960 --> 00:29:57,559 +there's other things that you can + +727 +00:29:55,159 --> 00:30:00,080 +calculate so if you're trying to figure + +728 +00:29:57,559 --> 00:30:01,640 +out whether a model um matches human + +729 +00:30:00,080 --> 00:30:04,279 +pairwise preferences you can just + +730 +00:30:01,640 --> 00:30:06,440 +calculate accuracy so I didn't put that + +731 +00:30:04,279 --> 00:30:08,080 +on um I didn't put that on the slide + +732 +00:30:06,440 --> 00:30:10,880 +here but you can just calculate accuracy + +733 +00:30:08,080 --> 00:30:13,120 +of pairwise preferences um you can also + +734 +00:30:10,880 --> 00:30:15,360 +calculate the absolute error between the + +735 +00:30:13,120 --> 00:30:19,320 +the judgments if you want to know uh + +736 +00:30:15,360 --> 00:30:21,720 +whether the absolute error matches so um + +737 +00:30:19,320 --> 00:30:24,159 +the these are good things to do if you + +738 +00:30:21,720 --> 00:30:25,600 +want to use an evaluation metric but you + +739 +00:30:24,159 --> 00:30:27,200 +aren't sure whether it's good or not I + +740 +00:30:25,600 --> 00:30:29,640 +would check to see whether the authors + +741 +00:30:27,200 --> 00:30:32,000 +have done this sort of meta evaluation + +742 +00:30:29,640 --> 00:30:33,760 +if they haven't be a little bit + +743 +00:30:32,000 --> 00:30:36,960 +suspicious if they have be a little bit + +744 +00:30:33,760 --> 00:30:39,799 +less suspicious but um + +745 +00:30:36,960 --> 00:30:42,960 +yeah how do people do this typically uh + +746 +00:30:39,799 --> 00:30:45,640 +usually they create uh data sets like + +747 +00:30:42,960 --> 00:30:49,440 +the WM they use data sets like the WMT + +748 +00:30:45,640 --> 00:30:53,960 +shared tasks um or + +749 +00:30:49,440 --> 00:30:57,679 +uh uh like some evl um but there's also + +750 +00:30:53,960 --> 00:30:59,960 +other ways to create um uh there's also + +751 +00:30:57,679 --> 00:31:01,639 +Lots other data sets but in order to do + +752 +00:30:59,960 --> 00:31:05,639 +this reliably you need a fairly large + +753 +00:31:01,639 --> 00:31:05,639 +data set so it's one thing to be aware + +754 +00:31:07,080 --> 00:31:10,760 +of + +755 +00:31:08,720 --> 00:31:14,200 +cool + +756 +00:31:10,760 --> 00:31:16,360 +um then the final thing um all of the + +757 +00:31:14,200 --> 00:31:17,919 +automatic evaluation methods that I + +758 +00:31:16,360 --> 00:31:20,240 +talked about now are trying to match + +759 +00:31:17,919 --> 00:31:22,679 +human preferences but that's not the + +760 +00:31:20,240 --> 00:31:24,960 +only thing that you necessarily want to + +761 +00:31:22,679 --> 00:31:28,440 +do the final thing that you might want + +762 +00:31:24,960 --> 00:31:30,840 +to do is uh use the model outputs in a + +763 +00:31:28,440 --> 00:31:34,200 +downstream system and see whether they + +764 +00:31:30,840 --> 00:31:36,399 +are effective for that so there's two + +765 +00:31:34,200 --> 00:31:39,080 +concepts of intrinsic evaluation and + +766 +00:31:36,399 --> 00:31:41,720 +extrinsic evaluation so intrinsic + +767 +00:31:39,080 --> 00:31:44,159 +evaluation um evaluates the quality of + +768 +00:31:41,720 --> 00:31:45,720 +the output itself and so that would be + +769 +00:31:44,159 --> 00:31:48,639 +like asking a human directly about how + +770 +00:31:45,720 --> 00:31:50,720 +good is this output extrinsic evaluation + +771 +00:31:48,639 --> 00:31:53,679 +is evaluating output quality by its + +772 +00:31:50,720 --> 00:31:57,000 +utility um and so just to give one + +773 +00:31:53,679 --> 00:31:58,360 +example um if you can evaluate large + +774 +00:31:57,000 --> 00:32:00,200 +language model summary + +775 +00:31:58,360 --> 00:32:04,200 +through question answering + +776 +00:32:00,200 --> 00:32:05,880 +accuracy um and so you can take the + +777 +00:32:04,200 --> 00:32:07,399 +output of an llm and feed it through a + +778 +00:32:05,880 --> 00:32:09,600 +question answering model and see whether + +779 +00:32:07,399 --> 00:32:12,399 +you're able to answer questions based on + +780 +00:32:09,600 --> 00:32:15,799 +this and that kind of gives you a better + +781 +00:32:12,399 --> 00:32:18,279 +idea of whether the summary require uh + +782 +00:32:15,799 --> 00:32:20,120 +incorporates requisite information but + +783 +00:32:18,279 --> 00:32:22,120 +if you think about anything an llm can + +784 +00:32:20,120 --> 00:32:23,760 +be used for usually it's part of a + +785 +00:32:22,120 --> 00:32:26,679 +bigger system so you can evaluate it as + +786 +00:32:23,760 --> 00:32:28,399 +a part of that bigger system um the + +787 +00:32:26,679 --> 00:32:30,639 +problem with this is it's a very + +788 +00:32:28,399 --> 00:32:33,960 +indirect way of assessing things so like + +789 +00:32:30,639 --> 00:32:36,080 +let's say your QA model is just bad uh + +790 +00:32:33,960 --> 00:32:38,480 +how can you disentangle the effect of + +791 +00:32:36,080 --> 00:32:41,679 +the L summary versus the QA model that's + +792 +00:32:38,480 --> 00:32:44,120 +not a trivial thing to do so ideally + +793 +00:32:41,679 --> 00:32:47,000 +like a combination of these two is + +794 +00:32:44,120 --> 00:32:47,000 +practically the best way + +795 +00:32:48,039 --> 00:32:52,200 +go cool so + +796 +00:32:56,039 --> 00:32:59,960 +yeah yeah it wouldn't necessar + +797 +00:32:58,360 --> 00:33:05,679 +say it's harder to do it might even be + +798 +00:32:59,960 --> 00:33:05,679 +easier to do um which is like let's + +799 +00:33:06,679 --> 00:33:11,720 +say Let me let me see if I can come up + +800 +00:33:09,360 --> 00:33:11,720 +with + +801 +00:33:12,639 --> 00:33:17,600 +example what let's + +802 +00:33:15,000 --> 00:33:19,670 +say you + +803 +00:33:17,600 --> 00:33:22,979 +are trying + +804 +00:33:19,670 --> 00:33:22,979 +[Music] + +805 +00:33:24,639 --> 00:33:29,760 +to let's say you're trying to + +806 +00:33:30,559 --> 00:33:33,559 +guess + +807 +00:33:39,000 --> 00:33:45,399 +whether let's say you're trying to guess + +808 +00:33:42,399 --> 00:33:46,559 +whether a someone will be hired at a + +809 +00:33:45,399 --> 00:33:52,039 +company or + +810 +00:33:46,559 --> 00:33:53,880 +not based on an llm generated summary of + +811 +00:33:52,039 --> 00:33:58,880 +their qualifications for a position or + +812 +00:33:53,880 --> 00:34:01,799 +something like that um and + +813 +00:33:58,880 --> 00:34:03,080 +you what actually maybe this is not a + +814 +00:34:01,799 --> 00:34:04,720 +great example because whether you should + +815 +00:34:03,080 --> 00:34:06,960 +be doing this ethically is a little bit + +816 +00:34:04,720 --> 00:34:08,159 +unclear but let's say you were doing + +817 +00:34:06,960 --> 00:34:09,560 +let's say you were doing something like + +818 +00:34:08,159 --> 00:34:11,520 +that just because it's one example I can + +819 +00:34:09,560 --> 00:34:14,320 +think of right now whether they will get + +820 +00:34:11,520 --> 00:34:16,320 +hired or not is um is clear because you + +821 +00:34:14,320 --> 00:34:19,399 +have a objective answer right whether + +822 +00:34:16,320 --> 00:34:21,480 +they were hired or not um or maybe maybe + +823 +00:34:19,399 --> 00:34:23,800 +another example would be like let's say + +824 +00:34:21,480 --> 00:34:26,320 +um let's say you want to predict the + +825 +00:34:23,800 --> 00:34:29,599 +diagnosis in a medical application based + +826 +00:34:26,320 --> 00:34:32,960 +on an llm generated some of somebody's + +827 +00:34:29,599 --> 00:34:35,919 +uh you know LM generated summary of + +828 +00:34:32,960 --> 00:34:38,480 +somebody's you know past medical history + +829 +00:34:35,919 --> 00:34:40,839 +and all this stuff and here you want the + +830 +00:34:38,480 --> 00:34:43,440 +llm generated summary you definitely + +831 +00:34:40,839 --> 00:34:44,879 +want the summary because the summary is + +832 +00:34:43,440 --> 00:34:47,560 +going to be viewed by a doctor who will + +833 +00:34:44,879 --> 00:34:49,359 +make the final decision but you also + +834 +00:34:47,560 --> 00:34:50,760 +have information about the diagnoses of + +835 +00:34:49,359 --> 00:34:52,399 +all the people in your medical system + +836 +00:34:50,760 --> 00:34:54,560 +later because you know they went through + +837 +00:34:52,399 --> 00:34:56,480 +your medical system for years and you + +838 +00:34:54,560 --> 00:34:58,200 +know later like through lots of tests + +839 +00:34:56,480 --> 00:35:00,800 +and stuff uh whether how they were + +840 +00:34:58,200 --> 00:35:02,320 +diagnosed so you generate an LM based + +841 +00:35:00,800 --> 00:35:05,000 +summary and then you predict the + +842 +00:35:02,320 --> 00:35:06,599 +diagnosis from the summary so there the + +843 +00:35:05,000 --> 00:35:08,040 +evaluation of the diagnosis is very + +844 +00:35:06,599 --> 00:35:11,480 +clear because you kind of have a gold + +845 +00:35:08,040 --> 00:35:12,599 +standard answer um but the EV intrinsic + +846 +00:35:11,480 --> 00:35:14,839 +evaluation of whether it's a good + +847 +00:35:12,599 --> 00:35:16,839 +summary or not is not as clear because + +848 +00:35:14,839 --> 00:35:19,400 +you'd have pass do whether it's good and + +849 +00:35:16,839 --> 00:35:21,079 +understandable summary so the extrinsic + +850 +00:35:19,400 --> 00:35:24,920 +evaluation might be easier because it's + +851 +00:35:21,079 --> 00:35:26,480 +clearer um so there are cases like that + +852 +00:35:24,920 --> 00:35:30,720 +um the problem is you would have to have + +853 +00:35:26,480 --> 00:35:33,800 +that data in order to do that um yeah do + +854 +00:35:30,720 --> 00:35:38,240 +like evaluation yeah I was just + +855 +00:35:33,800 --> 00:35:40,800 +wondering typically the + +856 +00:35:38,240 --> 00:35:42,880 +like like how do you accomodate the + +857 +00:35:40,800 --> 00:35:47,160 +diversity oh yeah that's a great that's + +858 +00:35:42,880 --> 00:35:50,240 +a great question um so how do you how do + +859 +00:35:47,160 --> 00:35:50,240 +you get these scores + +860 +00:35:50,720 --> 00:35:55,800 +here there's a number of different + +861 +00:35:53,200 --> 00:35:59,160 +things in the WMT shared tasks what they + +862 +00:35:55,800 --> 00:36:00,280 +did is they did + +863 +00:35:59,160 --> 00:36:03,200 +the first thing they do is they + +864 +00:36:00,280 --> 00:36:06,319 +normalize by annotator and what they do + +865 +00:36:03,200 --> 00:36:10,400 +is they basically take the zcore or Z + +866 +00:36:06,319 --> 00:36:12,240 +score of the um of the human annotator's + +867 +00:36:10,400 --> 00:36:14,880 +actual scores because some people are + +868 +00:36:12,240 --> 00:36:16,400 +more harsh than other people and so what + +869 +00:36:14,880 --> 00:36:20,680 +that means is you basically normalize to + +870 +00:36:16,400 --> 00:36:22,119 +have zero mean in unit variance um and + +871 +00:36:20,680 --> 00:36:24,119 +then after they've normalized to zero + +872 +00:36:22,119 --> 00:36:29,560 +mean and unit variance then I think they + +873 +00:36:24,119 --> 00:36:29,560 +average together different humans so um + +874 +00:36:30,160 --> 00:36:36,520 +then for how do you deal with the fact + +875 +00:36:33,680 --> 00:36:38,040 +that humans disagree on things and I + +876 +00:36:36,520 --> 00:36:39,480 +think it's pretty varied I don't know if + +877 +00:36:38,040 --> 00:36:42,160 +there's any gold standard way of doing + +878 +00:36:39,480 --> 00:36:43,839 +it but sometimes you just average + +879 +00:36:42,160 --> 00:36:46,359 +sometimes you throw away examples where + +880 +00:36:43,839 --> 00:36:47,960 +humans disagree a lot um because like + +881 +00:36:46,359 --> 00:36:50,200 +you can't get the humans to agree how + +882 +00:36:47,960 --> 00:36:53,319 +could you expect how could you expect a + +883 +00:36:50,200 --> 00:36:55,119 +machine to do well um so I think it it's + +884 +00:36:53,319 --> 00:36:59,200 +a little bit test + +885 +00:36:55,119 --> 00:37:01,560 +defending yeah so for + +886 +00:36:59,200 --> 00:37:04,560 +generation inin + +887 +00:37:01,560 --> 00:37:06,280 +andin yeah so for code generation that's + +888 +00:37:04,560 --> 00:37:08,200 +I I I love this example because I've + +889 +00:37:06,280 --> 00:37:09,960 +worked on code generation a lot of + +890 +00:37:08,200 --> 00:37:12,680 +people only think about extrinsic + +891 +00:37:09,960 --> 00:37:14,400 +evaluation of code Generation Um or I + +892 +00:37:12,680 --> 00:37:16,160 +don't know if it's extrinsic but only + +893 +00:37:14,400 --> 00:37:19,160 +think about execution based evaluation + +894 +00:37:16,160 --> 00:37:20,520 +of code generation which is like you + +895 +00:37:19,160 --> 00:37:22,400 +execute the code you see whether it + +896 +00:37:20,520 --> 00:37:25,040 +passs unit tests and other things like + +897 +00:37:22,400 --> 00:37:26,839 +this but in reality actually there's a + +898 +00:37:25,040 --> 00:37:28,599 +lot of other important things for code + +899 +00:37:26,839 --> 00:37:30,560 +like readability and other stuff like + +900 +00:37:28,599 --> 00:37:32,160 +that and you should be evaluating those + +901 +00:37:30,560 --> 00:37:34,920 +things but I think a lot of people like + +902 +00:37:32,160 --> 00:37:36,520 +kind of ignore that so um there there + +903 +00:37:34,920 --> 00:37:38,880 +are a few Pap that do that but most of + +904 +00:37:36,520 --> 00:37:41,000 +the time people just execute the Cod + +905 +00:37:38,880 --> 00:37:45,520 +process + +906 +00:37:41,000 --> 00:37:47,760 +un cool okay um so yeah moving on to the + +907 +00:37:45,520 --> 00:37:51,160 +learning part so now I'd like to talk + +908 +00:37:47,760 --> 00:37:55,280 +about uh learning and the first thing + +909 +00:37:51,160 --> 00:37:59,480 +I'll cover is error and risk and so + +910 +00:37:55,280 --> 00:38:02,280 +basically um the way we calculate air is + +911 +00:37:59,480 --> 00:38:03,119 +we generate an output and we calculate + +912 +00:38:02,280 --> 00:38:07,680 +its + +913 +00:38:03,119 --> 00:38:09,480 +Badness um and so generating the output + +914 +00:38:07,680 --> 00:38:13,160 +could be argmax it could be sampling it + +915 +00:38:09,480 --> 00:38:15,800 +could be anything else like that um and + +916 +00:38:13,160 --> 00:38:18,640 +we calculate its Badness uh which is one + +917 +00:38:15,800 --> 00:38:21,040 +minus in which could be like how bad is + +918 +00:38:18,640 --> 00:38:22,720 +the output uh if you're you have a + +919 +00:38:21,040 --> 00:38:24,760 +Badness measure or it could be one minus + +920 +00:38:22,720 --> 00:38:28,400 +the evaluation Square to calculate its + +921 +00:38:24,760 --> 00:38:30,160 +Badness and this is defined as error + +922 +00:38:28,400 --> 00:38:31,440 +and generally what you want to do is you + +923 +00:38:30,160 --> 00:38:33,520 +want to minimize + +924 +00:38:31,440 --> 00:38:36,800 +error + +925 +00:38:33,520 --> 00:38:39,400 +um because in the end you're going to be + +926 +00:38:36,800 --> 00:38:42,359 +deploying A system that just outputs you + +927 +00:38:39,400 --> 00:38:46,079 +know one thing and uh you're going to + +928 +00:38:42,359 --> 00:38:49,800 +want that to be as good a thing as + +929 +00:38:46,079 --> 00:38:53,000 +possible um but the problem with this is + +930 +00:38:49,800 --> 00:38:56,400 +there's no easy way to actually optimize + +931 +00:38:53,000 --> 00:38:59,079 +this value in especially in a text + +932 +00:38:56,400 --> 00:39:01,800 +generation sty setting but even in the + +933 +00:38:59,079 --> 00:39:06,839 +classification setting we can't easily + +934 +00:39:01,800 --> 00:39:06,839 +maximize err because um if you look at + +935 +00:39:09,040 --> 00:39:14,200 +the if you look at the surface of air uh + +936 +00:39:12,760 --> 00:39:15,960 +at some point you're going to have a + +937 +00:39:14,200 --> 00:39:18,319 +non-differentiable part when you take + +938 +00:39:15,960 --> 00:39:21,119 +the argmax and or when you do sampling + +939 +00:39:18,319 --> 00:39:23,319 +or anything like that so um you're not + +940 +00:39:21,119 --> 00:39:27,119 +going to be able to do gradient based + +941 +00:39:23,319 --> 00:39:29,200 +optimization so what we do normally is + +942 +00:39:27,119 --> 00:39:33,400 +um + +943 +00:39:29,200 --> 00:39:37,000 +we instead calculate something uh called + +944 +00:39:33,400 --> 00:39:38,560 +risk and what risk looks like is uh we + +945 +00:39:37,000 --> 00:39:40,599 +talked a little bit about minimum based + +946 +00:39:38,560 --> 00:39:43,520 +risk for decoding but this is for uh + +947 +00:39:40,599 --> 00:39:46,160 +training time and what it looks like is + +948 +00:39:43,520 --> 00:39:49,040 +it's essentially the expected err of the + +949 +00:39:46,160 --> 00:39:52,359 +output and the expected err of the + +950 +00:39:49,040 --> 00:39:54,760 +output um includes a probability in the + +951 +00:39:52,359 --> 00:39:58,240 +objective function here and that + +952 +00:39:54,760 --> 00:40:01,079 +probability uh is differential basically + +953 +00:39:58,240 --> 00:40:02,319 +so we can um uh we can easily do + +954 +00:40:01,079 --> 00:40:05,720 +gradient based + +955 +00:40:02,319 --> 00:40:09,119 +optimization through it um the problem + +956 +00:40:05,720 --> 00:40:12,200 +with this is It's differentiable but for + +957 +00:40:09,119 --> 00:40:17,160 +text generation for example the sum is + +958 +00:40:12,200 --> 00:40:20,319 +intractable because we have a combinator + +959 +00:40:17,160 --> 00:40:23,880 +large number of potential outputs um + +960 +00:40:20,319 --> 00:40:25,520 +because you know if this is we've talked + +961 +00:40:23,880 --> 00:40:28,720 +about this before but if this is like + +962 +00:40:25,520 --> 00:40:30,680 +link you know 50 and we have a 30,000 + +963 +00:40:28,720 --> 00:40:32,839 +vocabul that's 30,000 to the 50 + +964 +00:40:30,680 --> 00:40:34,599 +possibilities we can't take a su over + +965 +00:40:32,839 --> 00:40:36,359 +that many + +966 +00:40:34,599 --> 00:40:38,400 +possibilities + +967 +00:40:36,359 --> 00:40:42,680 +um + +968 +00:40:38,400 --> 00:40:45,839 +so minimum R risk training uh tries to + +969 +00:40:42,680 --> 00:40:48,440 +minimize risk reinforcement learning + +970 +00:40:45,839 --> 00:40:50,040 +also many of the models especially + +971 +00:40:48,440 --> 00:40:53,599 +policy gradient models are trying to + +972 +00:40:50,040 --> 00:40:55,240 +minimize risk as well so um but the + +973 +00:40:53,599 --> 00:40:58,040 +reason why I wanted to talk about risk + +974 +00:40:55,240 --> 00:41:00,440 +first is because this is very simple to + +975 +00:40:58,040 --> 00:41:01,640 +get to from the uh the point of view of + +976 +00:41:00,440 --> 00:41:06,560 +like all the things that we've studied + +977 +00:41:01,640 --> 00:41:06,560 +so so I think it's talking about + +978 +00:41:06,760 --> 00:41:11,800 +that + +979 +00:41:08,319 --> 00:41:15,520 +um one other thing that I should mention + +980 +00:41:11,800 --> 00:41:18,400 +about is + +981 +00:41:15,520 --> 00:41:23,079 +um or no sorry I'll I'll talk about that + +982 +00:41:18,400 --> 00:41:26,880 +later so when we want to optimize risk + +983 +00:41:23,079 --> 00:41:30,560 +um what we do is we sample in order to + +984 +00:41:26,880 --> 00:41:35,520 +make this trct so a very simple way to + +985 +00:41:30,560 --> 00:41:37,640 +minimize risk is instead of um instead + +986 +00:41:35,520 --> 00:41:39,359 +of summing over all of the possible + +987 +00:41:37,640 --> 00:41:42,760 +outputs we sum over a small number of + +988 +00:41:39,359 --> 00:41:46,079 +possible outputs and we upgrade uh and + +989 +00:41:42,760 --> 00:41:47,359 +we uh sorry normalize uh to make this + +990 +00:41:46,079 --> 00:41:51,200 +all add up to + +991 +00:41:47,359 --> 00:41:52,839 +one and so this normalizer here is + +992 +00:41:51,200 --> 00:41:55,319 +basically the sum over all of the + +993 +00:41:52,839 --> 00:41:58,599 +probabilities that we have uh on the top + +994 +00:41:55,319 --> 00:42:02,119 +part here and and these samples can be + +995 +00:41:58,599 --> 00:42:05,480 +created either using sampling or n best + +996 +00:42:02,119 --> 00:42:07,040 +search we don't need to have from the + +997 +00:42:05,480 --> 00:42:11,040 +point of view of doing this sort of + +998 +00:42:07,040 --> 00:42:13,960 +minimum risk training the kind of + +999 +00:42:11,040 --> 00:42:16,880 +correct way of doing this is sampling + +1000 +00:42:13,960 --> 00:42:19,880 +using ancestral sampling uh like we + +1001 +00:42:16,880 --> 00:42:23,079 +talked about before and um in minimizing + +1002 +00:42:19,880 --> 00:42:25,839 +the output based on the the samples but + +1003 +00:42:23,079 --> 00:42:28,480 +the problem with that is um as many of + +1004 +00:42:25,839 --> 00:42:31,440 +you also might have seen when you were + +1005 +00:42:28,480 --> 00:42:33,599 +sampling from your language model uh + +1006 +00:42:31,440 --> 00:42:35,160 +from assignment one if you sample with + +1007 +00:42:33,599 --> 00:42:38,040 +temperature one it gives you a lot of + +1008 +00:42:35,160 --> 00:42:40,720 +like not very good outlets right and so + +1009 +00:42:38,040 --> 00:42:43,400 +if you're sampling with temperature one + +1010 +00:42:40,720 --> 00:42:45,000 +um you'll be exploring a a very large + +1011 +00:42:43,400 --> 00:42:47,880 +part of the space that actually isn't + +1012 +00:42:45,000 --> 00:42:49,720 +very good and so because of this uh some + +1013 +00:42:47,880 --> 00:42:51,480 +other Alternatives that you can use is + +1014 +00:42:49,720 --> 00:42:53,400 +you can just do endb search to find the + +1015 +00:42:51,480 --> 00:42:55,280 +best outputs or you can sample with a + +1016 +00:42:53,400 --> 00:42:58,079 +temperature that's not one or something + +1017 +00:42:55,280 --> 00:43:00,240 +like that and basically create uh you + +1018 +00:42:58,079 --> 00:43:02,520 +know a list of possible hypotheses and + +1019 +00:43:00,240 --> 00:43:04,079 +then normalize other B so that's another + +1020 +00:43:02,520 --> 00:43:06,240 +option and very often not using + +1021 +00:43:04,079 --> 00:43:11,200 +temperature one is a better + +1022 +00:43:06,240 --> 00:43:15,280 +way um if you're sampling with not + +1023 +00:43:11,200 --> 00:43:18,640 +temperature one and you are um + +1024 +00:43:15,280 --> 00:43:20,920 +potentially getting multiple outputs you + +1025 +00:43:18,640 --> 00:43:23,400 +should try to D duplicate or sample + +1026 +00:43:20,920 --> 00:43:25,480 +without replacement because if you get + +1027 +00:43:23,400 --> 00:43:27,559 +multiple outputs here it messes up your + +1028 +00:43:25,480 --> 00:43:30,680 +equations if you basically uh have the + +1029 +00:43:27,559 --> 00:43:30,680 +same one in there multiple + +1030 +00:43:32,160 --> 00:43:37,800 +times cool so so this is a really simple + +1031 +00:43:35,880 --> 00:43:40,079 +example of how you can do minimal risk + +1032 +00:43:37,800 --> 00:43:42,119 +training but now I want to get into uh + +1033 +00:43:40,079 --> 00:43:44,640 +like reinforcement learning which is the + +1034 +00:43:42,119 --> 00:43:48,119 +framing that most um + +1035 +00:43:44,640 --> 00:43:50,760 +modern Works about this Paulo uh one + +1036 +00:43:48,119 --> 00:43:52,559 +thing I should mention is there are + +1037 +00:43:50,760 --> 00:43:55,240 +actually other alternatives to learning + +1038 +00:43:52,559 --> 00:43:57,359 +from uh human feedback including like + +1039 +00:43:55,240 --> 00:43:59,359 +margin loss margin based losses and + +1040 +00:43:57,359 --> 00:44:00,960 +other stuff like that but most people + +1041 +00:43:59,359 --> 00:44:03,440 +nowadays use reinforcement learning so + +1042 +00:44:00,960 --> 00:44:06,359 +I'm only going to cover that + +1043 +00:44:03,440 --> 00:44:08,440 +here so what is reinforcement learning + +1044 +00:44:06,359 --> 00:44:11,000 +um learning reinforcement learning is + +1045 +00:44:08,440 --> 00:44:14,559 +learning where we have an environment uh + +1046 +00:44:11,000 --> 00:44:16,079 +x uh ability to make actions a and get a + +1047 +00:44:14,559 --> 00:44:20,160 +delayed reward + +1048 +00:44:16,079 --> 00:44:21,880 +R and um there's a really nice example + +1049 +00:44:20,160 --> 00:44:24,400 +uh if you're not familiar with the + +1050 +00:44:21,880 --> 00:44:27,480 +basics of policy gradient by Andre + +1051 +00:44:24,400 --> 00:44:28,800 +karpathy which I linked in the um in the + +1052 +00:44:27,480 --> 00:44:29,680 +recommended reading so you can take a + +1053 +00:44:28,800 --> 00:44:34,680 +look at + +1054 +00:44:29,680 --> 00:44:37,240 +that um but in that example gives an + +1055 +00:44:34,680 --> 00:44:39,440 +example of pong uh where you're playing + +1056 +00:44:37,240 --> 00:44:42,640 +the game pong where X is your observed + +1057 +00:44:39,440 --> 00:44:45,640 +image a is up or down and R is the wind + +1058 +00:44:42,640 --> 00:44:47,480 +loss at the end of the game uh does + +1059 +00:44:45,640 --> 00:44:50,559 +anyone have an idea about uh what this + +1060 +00:44:47,480 --> 00:44:52,119 +looks like for any arbitrary NLP task + +1061 +00:44:50,559 --> 00:44:56,520 +that we might want to do reinforcement + +1062 +00:44:52,119 --> 00:44:59,040 +learning for so what what is X what is a + +1063 +00:44:56,520 --> 00:44:59,040 +and what is + +1064 +00:45:00,040 --> 00:45:04,680 +are pick your favorite uh your favorite + +1065 +00:45:06,920 --> 00:45:09,920 +Trask + +1066 +00:45:10,960 --> 00:45:18,400 +anybody + +1067 +00:45:12,520 --> 00:45:18,400 +yeah be or what what's X first + +1068 +00:45:19,680 --> 00:45:28,720 +yeah you have generate okay is the + +1069 +00:45:24,440 --> 00:45:29,720 +next be like the Buton like whether or + +1070 +00:45:28,720 --> 00:45:32,520 +not + +1071 +00:45:29,720 --> 00:45:35,240 +you okay yeah I I think this is very + +1072 +00:45:32,520 --> 00:45:37,119 +close just to repeat it it's like X is + +1073 +00:45:35,240 --> 00:45:39,599 +what you've generated so far a is the + +1074 +00:45:37,119 --> 00:45:41,559 +next token and R is the button that the + +1075 +00:45:39,599 --> 00:45:45,400 +user clicks about whether it's good or + +1076 +00:45:41,559 --> 00:45:46,920 +not um I think that's reasonably good + +1077 +00:45:45,400 --> 00:45:48,760 +although I don't know if we'd expect + +1078 +00:45:46,920 --> 00:45:52,960 +them to click the button every token we + +1079 +00:45:48,760 --> 00:45:54,880 +generate right so um it might be that X + +1080 +00:45:52,960 --> 00:45:57,880 +is the conversational history up till + +1081 +00:45:54,880 --> 00:46:02,319 +this point um a + +1082 +00:45:57,880 --> 00:46:04,280 +a could be a next token generation and + +1083 +00:46:02,319 --> 00:46:06,520 +then R is a reward we get in an + +1084 +00:46:04,280 --> 00:46:08,280 +arbitrary time point it might not be + +1085 +00:46:06,520 --> 00:46:09,960 +like immediately after generating the + +1086 +00:46:08,280 --> 00:46:12,040 +next token but it might be later and + +1087 +00:46:09,960 --> 00:46:13,480 +that's actually really really important + +1088 +00:46:12,040 --> 00:46:15,040 +from the point of view of reinforcement + +1089 +00:46:13,480 --> 00:46:19,599 +learning and I'll I'll talk about that + +1090 +00:46:15,040 --> 00:46:23,040 +in a second um anyone have an idea from + +1091 +00:46:19,599 --> 00:46:24,960 +I don't know uh code generation or + +1092 +00:46:23,040 --> 00:46:28,119 +translation or some other + +1093 +00:46:24,960 --> 00:46:31,160 +things C generation maybe s is a + +1094 +00:46:28,119 --> 00:46:33,040 +compiler or like the gra scpt and then + +1095 +00:46:31,160 --> 00:46:37,000 +the + +1096 +00:46:33,040 --> 00:46:42,520 +is the actual code that right and reward + +1097 +00:46:37,000 --> 00:46:44,839 +is yep um so X could be the compiler + +1098 +00:46:42,520 --> 00:46:47,559 +it's probably the compiler and all of + +1099 +00:46:44,839 --> 00:46:50,200 +the surrounding code context like what + +1100 +00:46:47,559 --> 00:46:52,520 +what is the natural language output and + +1101 +00:46:50,200 --> 00:46:53,960 +it's also um you know what is the + +1102 +00:46:52,520 --> 00:46:57,280 +project that you're you're working on + +1103 +00:46:53,960 --> 00:47:00,079 +and stuff like that um a i think + +1104 +00:46:57,280 --> 00:47:02,800 +typically we would treat each token in + +1105 +00:47:00,079 --> 00:47:04,160 +the code to be an action um and then R + +1106 +00:47:02,800 --> 00:47:06,599 +would be the reward after a long + +1107 +00:47:04,160 --> 00:47:08,640 +sequence of actions um and it could be + +1108 +00:47:06,599 --> 00:47:11,119 +the reward from the compiler it could be + +1109 +00:47:08,640 --> 00:47:13,160 +the reward from a code readability model + +1110 +00:47:11,119 --> 00:47:15,720 +it could be the reward from a speed + +1111 +00:47:13,160 --> 00:47:17,079 +execution speed and stuff like that so + +1112 +00:47:15,720 --> 00:47:18,839 +like one of the interesting things about + +1113 +00:47:17,079 --> 00:47:22,640 +R is you can be really creative about + +1114 +00:47:18,839 --> 00:47:25,400 +how you form R um which is not easy to + +1115 +00:47:22,640 --> 00:47:27,319 +do uh if you're just doing maximum + +1116 +00:47:25,400 --> 00:47:29,240 +likelihood also so you can come up with + +1117 +00:47:27,319 --> 00:47:32,920 +a r that really matches with like what + +1118 +00:47:29,240 --> 00:47:36,559 +you want um what you want in an output + +1119 +00:47:32,920 --> 00:47:40,079 +so why reinforcement learning in NLP um + +1120 +00:47:36,559 --> 00:47:42,599 +and I think there's basically three um + +1121 +00:47:40,079 --> 00:47:44,240 +three answers the first one is you have + +1122 +00:47:42,599 --> 00:47:49,000 +a typical reinforcement learning + +1123 +00:47:44,240 --> 00:47:51,119 +scenario um where you have a dialogue + +1124 +00:47:49,000 --> 00:47:52,720 +where you get lots of responses and then + +1125 +00:47:51,119 --> 00:47:54,559 +you get a reward at the end so the + +1126 +00:47:52,720 --> 00:47:57,359 +thumbs up and thumbs down from humans is + +1127 +00:47:54,559 --> 00:47:59,839 +a very typical example of + +1128 +00:47:57,359 --> 00:48:02,800 +uh reinforcement learning because you + +1129 +00:47:59,839 --> 00:48:05,000 +get a delayed reward uh at some point in + +1130 +00:48:02,800 --> 00:48:07,599 +the dialogue when a human presses up or + +1131 +00:48:05,000 --> 00:48:09,280 +down um another like actually more + +1132 +00:48:07,599 --> 00:48:11,680 +technical scenario where reinforcement + +1133 +00:48:09,280 --> 00:48:14,960 +learning has been used um for a long + +1134 +00:48:11,680 --> 00:48:17,400 +time is call centers so we've had + +1135 +00:48:14,960 --> 00:48:20,680 +dialogue systems for call centers and + +1136 +00:48:17,400 --> 00:48:23,160 +then if you complete a ticket purchase + +1137 +00:48:20,680 --> 00:48:24,839 +um or you complete resolve a ticket + +1138 +00:48:23,160 --> 00:48:27,480 +without ever having to go to a human + +1139 +00:48:24,839 --> 00:48:30,800 +operator you get a really big reward + +1140 +00:48:27,480 --> 00:48:33,640 +if you have to go to the human operator + +1141 +00:48:30,800 --> 00:48:36,400 +you get maybe a smaller reward and if + +1142 +00:48:33,640 --> 00:48:39,200 +the person yells at you and hangs up + +1143 +00:48:36,400 --> 00:48:41,640 +then you get a really negative reward so + +1144 +00:48:39,200 --> 00:48:43,040 +um this is kind of the typical example + +1145 +00:48:41,640 --> 00:48:45,599 +reinforcement learning has been used for + +1146 +00:48:43,040 --> 00:48:48,520 +a long time there another example is if + +1147 +00:48:45,599 --> 00:48:53,280 +you have like latent variables uh chains + +1148 +00:48:48,520 --> 00:48:55,799 +of thought where um you decide the + +1149 +00:48:53,280 --> 00:48:58,839 +latent variable and then get a reward um + +1150 +00:48:55,799 --> 00:49:02,799 +you get a reward based Bas on how those + +1151 +00:48:58,839 --> 00:49:03,920 +latent variables affect the output so um + +1152 +00:49:02,799 --> 00:49:07,200 +this + +1153 +00:49:03,920 --> 00:49:09,799 +is uh this is another example + +1154 +00:49:07,200 --> 00:49:12,599 +because the Chain of Thought itself + +1155 +00:49:09,799 --> 00:49:13,880 +might not actually be good you might + +1156 +00:49:12,599 --> 00:49:15,839 +have a bad Chain of Thought and still + +1157 +00:49:13,880 --> 00:49:17,760 +get the correct answer so you don't + +1158 +00:49:15,839 --> 00:49:19,640 +actually know for sure that a chain of + +1159 +00:49:17,760 --> 00:49:22,359 +thought that was automatically generated + +1160 +00:49:19,640 --> 00:49:24,799 +is good or not but um that so that kind + +1161 +00:49:22,359 --> 00:49:27,000 +of makes it a reinforcement learning + +1162 +00:49:24,799 --> 00:49:29,520 +problem and another thing is you might + +1163 +00:49:27,000 --> 00:49:32,520 +have a sequence level evaluation metric + +1164 +00:49:29,520 --> 00:49:34,240 +um so that you can't optimize the + +1165 +00:49:32,520 --> 00:49:36,839 +evaluation metric without uh first + +1166 +00:49:34,240 --> 00:49:38,480 +generating the whole like sequence so + +1167 +00:49:36,839 --> 00:49:40,880 +that would be any of the evaluation + +1168 +00:49:38,480 --> 00:49:42,400 +metrics that I talked about before so um + +1169 +00:49:40,880 --> 00:49:44,720 +these are three scenarios where you can + +1170 +00:49:42,400 --> 00:49:47,079 +use reinforcement + +1171 +00:49:44,720 --> 00:49:50,000 +planning so + +1172 +00:49:47,079 --> 00:49:51,400 +um I'm going to make a few steps through + +1173 +00:49:50,000 --> 00:49:54,640 +but like let's start again with our + +1174 +00:49:51,400 --> 00:49:57,359 +supervised mle loss and uh that's just + +1175 +00:49:54,640 --> 00:50:01,799 +the log probability here um in the + +1176 +00:49:57,359 --> 00:50:04,160 +context of reinforcement learning this + +1177 +00:50:01,799 --> 00:50:07,079 +is also called imitation + +1178 +00:50:04,160 --> 00:50:08,880 +learning because um essentially you're + +1179 +00:50:07,079 --> 00:50:12,680 +learning how to perform actions by + +1180 +00:50:08,880 --> 00:50:14,559 +imitating a teacher um and imitation + +1181 +00:50:12,680 --> 00:50:15,960 +learning is not just supervised mle + +1182 +00:50:14,559 --> 00:50:18,440 +there's also other varieties of + +1183 +00:50:15,960 --> 00:50:21,440 +imitation learning but um this is one + +1184 +00:50:18,440 --> 00:50:21,440 +variety of imitation + +1185 +00:50:22,520 --> 00:50:27,640 +learning the next thing I'd like to talk + +1186 +00:50:24,599 --> 00:50:30,079 +about is self-training and basically + +1187 +00:50:27,640 --> 00:50:31,760 +self-training the idea is that you + +1188 +00:50:30,079 --> 00:50:33,720 +sample or argmax according to the + +1189 +00:50:31,760 --> 00:50:36,119 +current model so you have your current + +1190 +00:50:33,720 --> 00:50:38,000 +model and you get a sample from it and + +1191 +00:50:36,119 --> 00:50:41,520 +then you use the sample or samples to + +1192 +00:50:38,000 --> 00:50:43,680 +maximize likelihood so um basically + +1193 +00:50:41,520 --> 00:50:47,520 +instead of doing maximum likelihood with + +1194 +00:50:43,680 --> 00:50:49,520 +respect to the a gold standard output + +1195 +00:50:47,520 --> 00:50:51,280 +you're doing it with respect to your own + +1196 +00:50:49,520 --> 00:50:55,280 +output + +1197 +00:50:51,280 --> 00:50:55,280 +so does this seem like a good + +1198 +00:50:55,640 --> 00:51:03,880 +idea I see a few people shaking heads um + +1199 +00:51:00,480 --> 00:51:03,880 +any ideas why this is not a good + +1200 +00:51:04,680 --> 00:51:07,680 +idea + +1201 +00:51:15,040 --> 00:51:20,599 +yeah yeah exactly so if you don't have + +1202 +00:51:17,720 --> 00:51:23,760 +any access to any notion well it's good + +1203 +00:51:20,599 --> 00:51:27,480 +um this will be optimizing towards good + +1204 +00:51:23,760 --> 00:51:28,839 +outputs and bad outputs right so um your + +1205 +00:51:27,480 --> 00:51:30,200 +model might be outputting bad outputs + +1206 +00:51:28,839 --> 00:51:32,839 +and you're just reinforcing the errors + +1207 +00:51:30,200 --> 00:51:35,160 +set the model R already nonetheless like + +1208 +00:51:32,839 --> 00:51:37,799 +self trining actually improves your + +1209 +00:51:35,160 --> 00:51:39,680 +accuracy somewhat in some cases like for + +1210 +00:51:37,799 --> 00:51:43,040 +example if your accuracy is if your + +1211 +00:51:39,680 --> 00:51:45,520 +model is Right more often than not um + +1212 +00:51:43,040 --> 00:51:49,119 +basically optimizing towards the more + +1213 +00:51:45,520 --> 00:51:51,720 +often the not right outputs can actually + +1214 +00:51:49,119 --> 00:51:53,640 +um due to the implicit regularization + +1215 +00:51:51,720 --> 00:51:55,000 +that models have and early stopping and + +1216 +00:51:53,640 --> 00:51:56,559 +other things like that it can actually + +1217 +00:51:55,000 --> 00:51:59,280 +move you in the right direction and + +1218 +00:51:56,559 --> 00:52:01,559 +improve accuracy + +1219 +00:51:59,280 --> 00:52:05,000 +um + +1220 +00:52:01,559 --> 00:52:06,640 +so there are alternatives to this that + +1221 +00:52:05,000 --> 00:52:09,520 +further improve accuracy so like for + +1222 +00:52:06,640 --> 00:52:12,720 +example if you have multiple models and + +1223 +00:52:09,520 --> 00:52:16,200 +um you only generate sentences where the + +1224 +00:52:12,720 --> 00:52:17,760 +models agree then this can improve your + +1225 +00:52:16,200 --> 00:52:20,000 +uh overall accuracy + +1226 +00:52:17,760 --> 00:52:24,240 +further um this is called code training + +1227 +00:52:20,000 --> 00:52:27,799 +it was actually uh created by uh uh + +1228 +00:52:24,240 --> 00:52:30,160 +people at at CMU as well and another + +1229 +00:52:27,799 --> 00:52:32,280 +successful alternative uh is adding + +1230 +00:52:30,160 --> 00:52:34,920 +noise to the input to match the noise + +1231 +00:52:32,280 --> 00:52:38,760 +that you find in the output so if you uh + +1232 +00:52:34,920 --> 00:52:40,720 +add like word uh word-based Dropout or + +1233 +00:52:38,760 --> 00:52:44,000 +other things like that this can also + +1234 +00:52:40,720 --> 00:52:47,400 +help uh accommodate these things but + +1235 +00:52:44,000 --> 00:52:48,920 +anyway um so self trining is is useful + +1236 +00:52:47,400 --> 00:52:50,480 +but there are better Alternatives if you + +1237 +00:52:48,920 --> 00:52:54,079 +can get a reward + +1238 +00:52:50,480 --> 00:52:55,559 +function so um the simplest variety of + +1239 +00:52:54,079 --> 00:52:56,960 +this is something called policy gradient + +1240 +00:52:55,559 --> 00:52:59,720 +or reinforce + +1241 +00:52:56,960 --> 00:53:02,319 +um or more specifically reinforce and + +1242 +00:52:59,720 --> 00:53:06,280 +basically what this does is this adds a + +1243 +00:53:02,319 --> 00:53:08,359 +term that scales the loss by the reward + +1244 +00:53:06,280 --> 00:53:12,400 +so if you can get a reward for each + +1245 +00:53:08,359 --> 00:53:15,680 +output basically this + +1246 +00:53:12,400 --> 00:53:18,119 +um you uh instead of doing self trining + +1247 +00:53:15,680 --> 00:53:21,760 +entirely by itself you multiply it by a + +1248 +00:53:18,119 --> 00:53:23,119 +reward and this allows you to increase + +1249 +00:53:21,760 --> 00:53:24,640 +the likelihood of things that get a high + +1250 +00:53:23,119 --> 00:53:28,440 +reward decrease the likelihood of things + +1251 +00:53:24,640 --> 00:53:28,440 +that get a low reward + +1252 +00:53:29,680 --> 00:53:34,960 +so uh a brief quiz here under what + +1253 +00:53:32,440 --> 00:53:37,599 +conditions is this equal equivalent to + +1254 +00:53:34,960 --> 00:53:41,480 +ml or essentially equivalent to maximum + +1255 +00:53:37,599 --> 00:53:43,079 +leg uh estimation and so like in order + +1256 +00:53:41,480 --> 00:53:45,480 +to make this quiz easier I'll go back to + +1257 +00:53:43,079 --> 00:53:47,720 +maximum likelihood estimation so it + +1258 +00:53:45,480 --> 00:53:50,359 +looked a bit like this um you calculated + +1259 +00:53:47,720 --> 00:53:53,440 +the log probability of the true output + +1260 +00:53:50,359 --> 00:53:55,440 +and now let me go uh to + +1261 +00:53:53,440 --> 00:53:56,960 +here any + +1262 +00:53:55,440 --> 00:54:00,119 +ideas + +1263 +00:53:56,960 --> 00:54:05,040 +yeah when your reward equals to + +1264 +00:54:00,119 --> 00:54:05,040 +one some sometimes in zero other times + +1265 +00:54:07,760 --> 00:54:10,960 +what any + +1266 +00:54:12,760 --> 00:54:17,520 +ideas what when when does your reward + +1267 +00:54:15,280 --> 00:54:19,640 +need to be equal to one in order to make + +1268 +00:54:17,520 --> 00:54:23,400 +this + +1269 +00:54:19,640 --> 00:54:23,400 +equation equivalent this + +1270 +00:54:24,960 --> 00:54:31,680 +equation yeah when Y and Y hat are the + +1271 +00:54:27,319 --> 00:54:36,119 +same so um basically + +1272 +00:54:31,680 --> 00:54:38,880 +this objective is equivalent to the mle + +1273 +00:54:36,119 --> 00:54:43,160 +objective when you're using a zero1 + +1274 +00:54:38,880 --> 00:54:44,480 +loss um where or you're using an + +1275 +00:54:43,160 --> 00:54:46,359 +evaluation function that gives you a + +1276 +00:54:44,480 --> 00:54:50,920 +score of one when it's exact match and + +1277 +00:54:46,359 --> 00:54:51,720 +zero when it's not exact match so um but + +1278 +00:54:50,920 --> 00:54:54,480 +that + +1279 +00:54:51,720 --> 00:54:56,440 +also demonstrates that this can be more + +1280 +00:54:54,480 --> 00:54:58,400 +flexible because you can have other + +1281 +00:54:56,440 --> 00:55:00,160 +rewards that are not just one and zero + +1282 +00:54:58,400 --> 00:55:02,599 +for exact match but you can use things + +1283 +00:55:00,160 --> 00:55:05,359 +that give you partial credit you can use + +1284 +00:55:02,599 --> 00:55:06,880 +things that uplate multiple potential uh + +1285 +00:55:05,359 --> 00:55:08,880 +potentially correct outputs and other + +1286 +00:55:06,880 --> 00:55:13,400 +things like + +1287 +00:55:08,880 --> 00:55:17,160 +that so one problem with these methods + +1288 +00:55:13,400 --> 00:55:21,799 +is um how do we know which action led to + +1289 +00:55:17,160 --> 00:55:24,720 +the reward so the best scenario is after + +1290 +00:55:21,799 --> 00:55:26,359 +each action you get a reward so after + +1291 +00:55:24,720 --> 00:55:28,960 +each token that you generated you get + +1292 +00:55:26,359 --> 00:55:31,240 +get a thumbs up or thumbs down uh from + +1293 +00:55:28,960 --> 00:55:34,280 +the user about whether they like that + +1294 +00:55:31,240 --> 00:55:36,000 +token or not um and how much happier + +1295 +00:55:34,280 --> 00:55:37,720 +they are after you generated that token + +1296 +00:55:36,000 --> 00:55:42,400 +than they were before you generated that + +1297 +00:55:37,720 --> 00:55:44,200 +token um the problem with this is that + +1298 +00:55:42,400 --> 00:55:45,799 +that's completely infeasible right like + +1299 +00:55:44,200 --> 00:55:47,039 +every time after you use chat GPD you're + +1300 +00:55:45,799 --> 00:55:50,480 +not going to press thumbs up and thumbs + +1301 +00:55:47,039 --> 00:55:52,559 +down after each token so um in reality + +1302 +00:55:50,480 --> 00:55:55,559 +what we get is usually we get it at the + +1303 +00:55:52,559 --> 00:55:57,000 +end of uh roll out of many many + +1304 +00:55:55,559 --> 00:55:58,640 +different actions and we're not sure + +1305 +00:55:57,000 --> 00:55:59,720 +which action is responsible for giving + +1306 +00:55:58,640 --> 00:56:02,559 +us the + +1307 +00:55:59,720 --> 00:56:05,440 +reward and + +1308 +00:56:02,559 --> 00:56:08,000 +so there's a few typical ways of dealing + +1309 +00:56:05,440 --> 00:56:09,640 +with this um the most typical way of + +1310 +00:56:08,000 --> 00:56:13,359 +dealing with this right now is just not + +1311 +00:56:09,640 --> 00:56:15,440 +dealing with it um and just hoping that + +1312 +00:56:13,359 --> 00:56:17,200 +your optimization algorithm internally + +1313 +00:56:15,440 --> 00:56:21,480 +will be able to do credit + +1314 +00:56:17,200 --> 00:56:24,520 +assignment um and so what that entails + +1315 +00:56:21,480 --> 00:56:27,319 +is essentially you um give an equal + +1316 +00:56:24,520 --> 00:56:29,880 +reward for each token in the output + +1317 +00:56:27,319 --> 00:56:32,480 +other ways that you can deal with it are + +1318 +00:56:29,880 --> 00:56:35,640 +um you can assign decaying rewards from + +1319 +00:56:32,480 --> 00:56:37,559 +future events so like let's say let's + +1320 +00:56:35,640 --> 00:56:41,839 +say you're talking about a chat bot for + +1321 +00:56:37,559 --> 00:56:44,119 +example maybe this is the the most uh + +1322 +00:56:41,839 --> 00:56:46,599 +kind of intuitive way of thinking about + +1323 +00:56:44,119 --> 00:56:50,400 +it but you you have a chat bot you have + +1324 +00:56:46,599 --> 00:56:52,599 +like 20 chat turns and you have the user + +1325 +00:56:50,400 --> 00:56:55,640 +give a thumbs up or a thumbs down on the + +1326 +00:56:52,599 --> 00:56:58,920 +20th chat turn there you would assign a + +1327 +00:56:55,640 --> 00:57:01,440 +reward of um like let's say it gave a + +1328 +00:56:58,920 --> 00:57:03,640 +thumbs up there you would re assign a + +1329 +00:57:01,440 --> 00:57:06,559 +reward of one for the previous chat turn + +1330 +00:57:03,640 --> 00:57:09,839 +a reward of like 0.5 for the second to + +1331 +00:57:06,559 --> 00:57:11,720 +previous chat term a reward of 0.25 for + +1332 +00:57:09,839 --> 00:57:14,319 +the third to previous chat term to + +1333 +00:57:11,720 --> 00:57:16,160 +basically say yeah like the user is + +1334 +00:57:14,319 --> 00:57:18,240 +feeling good at the moment they gave the + +1335 +00:57:16,160 --> 00:57:20,359 +thumbs up and that's probably more + +1336 +00:57:18,240 --> 00:57:23,400 +likely due to the things that happened + +1337 +00:57:20,359 --> 00:57:23,400 +recently so + +1338 +00:57:23,559 --> 00:57:28,119 +yeah we have a + +1339 +00:57:26,680 --> 00:57:32,280 +like not + +1340 +00:57:28,119 --> 00:57:34,160 +learning so the reward model can be any + +1341 +00:57:32,280 --> 00:57:35,839 +of the methods that I talked about + +1342 +00:57:34,160 --> 00:57:37,480 +before so it can be human feedback + +1343 +00:57:35,839 --> 00:57:39,000 +directly like a thumbs up or a thumbs + +1344 +00:57:37,480 --> 00:57:42,200 +down it could also be from a reward + +1345 +00:57:39,000 --> 00:57:44,599 +model uh that was pre-trained you could + +1346 +00:57:42,200 --> 00:57:47,680 +also theoretically learn the reward + +1347 +00:57:44,599 --> 00:57:52,720 +model simultaneously but you'd have to + +1348 +00:57:47,680 --> 00:57:55,200 +simultaneously with the model itself um + +1349 +00:57:52,720 --> 00:57:57,280 +so yeah I'm going to talk a little bit + +1350 +00:57:55,200 --> 00:58:00,359 +about DP which kind of does that a + +1351 +00:57:57,280 --> 00:58:01,720 +little bit but um I I would basically + +1352 +00:58:00,359 --> 00:58:03,160 +say that wherever you're getting your + +1353 +00:58:01,720 --> 00:58:06,280 +reward is probably from one of the + +1354 +00:58:03,160 --> 00:58:06,280 +things I talked about earlier + +1355 +00:58:06,359 --> 00:58:14,960 +today cool any other + +1356 +00:58:09,319 --> 00:58:17,720 +questions okay um so that's the basic + +1357 +00:58:14,960 --> 00:58:20,640 +the basic idea the very simplest thing + +1358 +00:58:17,720 --> 00:58:23,359 +that you can do is you can just sample + +1359 +00:58:20,640 --> 00:58:26,079 +um optimize the subjective function this + +1360 +00:58:23,359 --> 00:58:28,359 +is dead easy you it's not hard to imp + +1361 +00:58:26,079 --> 00:58:30,799 +imp it all as long as you have some + +1362 +00:58:28,359 --> 00:58:32,760 +source of reward signal um but the + +1363 +00:58:30,799 --> 00:58:35,559 +problem is uh reinforcement learning can + +1364 +00:58:32,760 --> 00:58:38,599 +be very unstable and it's hard to get it + +1365 +00:58:35,559 --> 00:58:40,160 +to uh you know work properly if you uh + +1366 +00:58:38,599 --> 00:58:42,400 +don't do some additional tricks so I'd + +1367 +00:58:40,160 --> 00:58:45,720 +like to talk about this + +1368 +00:58:42,400 --> 00:58:45,720 +next oh yeah + +1369 +00:58:48,880 --> 00:58:51,880 +sir + +1370 +00:58:55,039 --> 00:58:58,039 +yeah + +1371 +00:59:03,280 --> 00:59:08,960 +yeah the typical the typical way is you + +1372 +00:59:05,440 --> 00:59:12,960 +just have an exponential decay um so you + +1373 +00:59:08,960 --> 00:59:16,200 +you multiply each time by what 0.5 0. or + +1374 +00:59:12,960 --> 00:59:19,400 +something like that + +1375 +00:59:16,200 --> 00:59:19,400 +um from + +1376 +00:59:20,319 --> 00:59:27,720 +A6 um cool okay + +1377 +00:59:25,039 --> 00:59:30,720 +so + +1378 +00:59:27,720 --> 00:59:33,319 +and that's one option and sorry just to + +1379 +00:59:30,720 --> 00:59:35,760 +clarify the most common option nowadays + +1380 +00:59:33,319 --> 00:59:37,920 +um at least from the point of view of + +1381 +00:59:35,760 --> 00:59:39,839 +models is not to Decay it at all and + +1382 +00:59:37,920 --> 00:59:43,880 +just assign the same amount for each + +1383 +00:59:39,839 --> 00:59:45,319 +token um I'm not actually 100% sure what + +1384 +00:59:43,880 --> 00:59:47,319 +people are doing with respect to like + +1385 +00:59:45,319 --> 00:59:49,280 +long chat things I think probably + +1386 +00:59:47,319 --> 00:59:51,720 +they're only assigning it to the current + +1387 +00:59:49,280 --> 00:59:54,240 +like utterance and then not optimizing + +1388 +00:59:51,720 --> 00:59:57,240 +the previous utterances so like if they + +1389 +00:59:54,240 --> 00:59:59,039 +get a thumbs up or thumbs down signal um + +1390 +00:59:57,240 --> 01:00:00,720 +then they they would assign an + +1391 +00:59:59,039 --> 01:00:02,440 +equivalent reward for all of the tokens + +1392 +01:00:00,720 --> 01:00:04,640 +and the current utterance and zero + +1393 +01:00:02,440 --> 01:00:06,119 +reward for the previous ones but I'm not + +1394 +01:00:04,640 --> 01:00:08,480 +100% sure about that there might be + +1395 +01:00:06,119 --> 01:00:11,200 +other methods that people are + +1396 +01:00:08,480 --> 01:00:13,960 +using um + +1397 +01:00:11,200 --> 01:00:16,680 +cool so uh stabilizing reinforcement + +1398 +01:00:13,960 --> 01:00:18,520 +learning so um stabilizing reinforcement + +1399 +01:00:16,680 --> 01:00:21,839 +learning there's a lot of reasons why + +1400 +01:00:18,520 --> 01:00:23,880 +it's unstable um the first reason is + +1401 +01:00:21,839 --> 01:00:27,200 +you're sampling an individual output and + +1402 +01:00:23,880 --> 01:00:30,160 +calculating the um uh calculating based + +1403 +01:00:27,200 --> 01:00:32,039 +on the S individual sampled output and + +1404 +01:00:30,160 --> 01:00:33,440 +then there's an Infinity of other + +1405 +01:00:32,039 --> 01:00:36,480 +outputs that you could be optimizing + +1406 +01:00:33,440 --> 01:00:39,119 +over for mle this is not a problem + +1407 +01:00:36,480 --> 01:00:41,319 +because for mle you're always + +1408 +01:00:39,119 --> 01:00:45,359 +contrasting the gold standard output to + +1409 +01:00:41,319 --> 01:00:46,599 +all of the other outputs in the space um + +1410 +01:00:45,359 --> 01:00:48,280 +and you're saying I want to upweight the + +1411 +01:00:46,599 --> 01:00:51,200 +gold standard output and down we all of + +1412 +01:00:48,280 --> 01:00:53,039 +the other ones but for reinforcement + +1413 +01:00:51,200 --> 01:00:54,760 +learning you only have a single sampled + +1414 +01:00:53,039 --> 01:00:57,520 +output that output might be wrong and + +1415 +01:00:54,760 --> 01:00:59,359 +that's a source of inst ility this is + +1416 +01:00:57,520 --> 01:01:02,079 +particularly a problem when using bigger + +1417 +01:00:59,359 --> 01:01:05,960 +output spaces like all of the in the + +1418 +01:01:02,079 --> 01:01:07,920 +vocabul another problem is uh anytime + +1419 +01:01:05,960 --> 01:01:11,599 +you start using negative + +1420 +01:01:07,920 --> 01:01:15,160 +rewards um because if you start using + +1421 +01:01:11,599 --> 01:01:17,559 +negative rewards those rewards will be + +1422 +01:01:15,160 --> 01:01:19,520 +downweighting the probability of a + +1423 +01:01:17,559 --> 01:01:20,680 +particular output sequence and that + +1424 +01:01:19,520 --> 01:01:22,440 +might be a good idea maybe you're + +1425 +01:01:20,680 --> 01:01:24,319 +getting a toxic output or something like + +1426 +01:01:22,440 --> 01:01:25,960 +that and you want to down it but at the + +1427 +01:01:24,319 --> 01:01:28,280 +same time in addition to that toxic + +1428 +01:01:25,960 --> 01:01:30,000 +output there's like you know a + +1429 +01:01:28,280 --> 01:01:31,599 +combinatorial number of completely + +1430 +01:01:30,000 --> 01:01:33,880 +nonsense outputs that aren't even + +1431 +01:01:31,599 --> 01:01:36,599 +English and so basically you can start + +1432 +01:01:33,880 --> 01:01:38,920 +diverge from the N starting start to + +1433 +01:01:36,599 --> 01:01:40,799 +diverge from the natural like language + +1434 +01:01:38,920 --> 01:01:44,720 +modeling distribution that you have + +1435 +01:01:40,799 --> 01:01:49,079 +before so this is a big uh a big + +1436 +01:01:44,720 --> 01:01:51,880 +problem so a number of uh strategies can + +1437 +01:01:49,079 --> 01:01:53,880 +be used to stabilize the first one is + +1438 +01:01:51,880 --> 01:01:55,480 +this is completely obvious right now and + +1439 +01:01:53,880 --> 01:01:57,240 +nobody in their right mind would avoid + +1440 +01:01:55,480 --> 01:02:00,119 +doing this but the first one is + +1441 +01:01:57,240 --> 01:02:02,839 +pre-training with mle and so you start + +1442 +01:02:00,119 --> 01:02:04,920 +with a pre-trained model um and then + +1443 +01:02:02,839 --> 01:02:09,359 +switch over to RL after you finished + +1444 +01:02:04,920 --> 01:02:11,520 +pre-training the model um and so + +1445 +01:02:09,359 --> 01:02:13,279 +this makes a lot of sense if you're + +1446 +01:02:11,520 --> 01:02:14,960 +training a language model which I assume + +1447 +01:02:13,279 --> 01:02:17,039 +that almost everybody in this class is + +1448 +01:02:14,960 --> 01:02:20,279 +going to be doing but it does only work + +1449 +01:02:17,039 --> 01:02:22,720 +in scenarios where you can run mle and + +1450 +01:02:20,279 --> 01:02:24,359 +so it doesn't work if you're predicting + +1451 +01:02:22,720 --> 01:02:27,240 +like latent variables that aren't + +1452 +01:02:24,359 --> 01:02:28,760 +included in the original space + +1453 +01:02:27,240 --> 01:02:31,960 +um it + +1454 +01:02:28,760 --> 01:02:34,279 +also doesn't work in a setting where + +1455 +01:02:31,960 --> 01:02:36,640 +like you want to learn a + +1456 +01:02:34,279 --> 01:02:40,799 +chatbot you want to learn a chatbot for + +1457 +01:02:36,640 --> 01:02:44,200 +customer service for a + +1458 +01:02:40,799 --> 01:02:48,039 +company that + +1459 +01:02:44,200 --> 01:02:49,960 +has like for example a product catalog + +1460 +01:02:48,039 --> 01:02:53,559 +that the language model has never seen + +1461 +01:02:49,960 --> 01:02:56,000 +before and so if the language model has + +1462 +01:02:53,559 --> 01:02:57,359 +no information about the product catalog + +1463 +01:02:56,000 --> 01:02:59,920 +whatsoever you don't provide it through + +1464 +01:02:57,359 --> 01:03:02,440 +rag or something like that it's going to + +1465 +01:02:59,920 --> 01:03:04,039 +have to explore infinitely or not + +1466 +01:03:02,440 --> 01:03:05,599 +infinitely but it's going to have to + +1467 +01:03:04,039 --> 01:03:08,359 +explore too large of a space and you're + +1468 +01:03:05,599 --> 01:03:10,000 +never going to converge with um with + +1469 +01:03:08,359 --> 01:03:12,359 +your language modeling objectives so you + +1470 +01:03:10,000 --> 01:03:15,000 +need to basically be able to create at + +1471 +01:03:12,359 --> 01:03:16,079 +least some supervised training data to + +1472 +01:03:15,000 --> 01:03:19,279 +train with + +1473 +01:03:16,079 --> 01:03:20,720 +mle um but assuming you can do that I'm + +1474 +01:03:19,279 --> 01:03:22,920 +assuming that almost everybody is going + +1475 +01:03:20,720 --> 01:03:26,400 +to do some sort of pre-training with + +1476 +01:03:22,920 --> 01:03:27,880 +ML um The Next Step that people use uh + +1477 +01:03:26,400 --> 01:03:30,520 +in reinforcement learning that's really + +1478 +01:03:27,880 --> 01:03:34,319 +important to stabilize is regularization + +1479 +01:03:30,520 --> 01:03:35,880 +to an existing model and you have an + +1480 +01:03:34,319 --> 01:03:39,039 +existing model and you want to prevent + +1481 +01:03:35,880 --> 01:03:40,559 +it from getting too far away and the + +1482 +01:03:39,039 --> 01:03:42,279 +reason why you want to do this is like + +1483 +01:03:40,559 --> 01:03:45,720 +let's say you start assigning a negative + +1484 +01:03:42,279 --> 01:03:47,440 +reward to toxic utterances for example + +1485 +01:03:45,720 --> 01:03:49,200 +if your model stops being a language + +1486 +01:03:47,440 --> 01:03:51,920 +model whatsoever that's a bad idea so + +1487 +01:03:49,200 --> 01:03:53,400 +you want to keep it as a language model + +1488 +01:03:51,920 --> 01:03:55,599 +keep it close enough to still being a + +1489 +01:03:53,400 --> 01:03:57,559 +competent language model while you know + +1490 +01:03:55,599 --> 01:03:59,599 +like removing the toxic + +1491 +01:03:57,559 --> 01:04:03,039 +utterances so there's a number of + +1492 +01:03:59,599 --> 01:04:05,680 +methods that people use to do this um uh + +1493 +01:04:03,039 --> 01:04:08,359 +the most prominent ones are kale + +1494 +01:04:05,680 --> 01:04:10,279 +regularization uh well so the the first + +1495 +01:04:08,359 --> 01:04:13,119 +most prominent one is K regularization + +1496 +01:04:10,279 --> 01:04:15,839 +and the way this works is basically in + +1497 +01:04:13,119 --> 01:04:19,400 +addition you add you have two + +1498 +01:04:15,839 --> 01:04:22,279 +terms the first term is a term that + +1499 +01:04:19,400 --> 01:04:25,760 +improves your reward so you have your + +1500 +01:04:22,279 --> 01:04:28,039 +old model where your old model is + +1501 +01:04:25,760 --> 01:04:31,279 +creating a + +1502 +01:04:28,039 --> 01:04:32,440 +probability uh it has a probability here + +1503 +01:04:31,279 --> 01:04:34,960 +and then you have the probability + +1504 +01:04:32,440 --> 01:04:38,160 +assigned by your new model and then you + +1505 +01:04:34,960 --> 01:04:41,200 +have your reward signal here and so this + +1506 +01:04:38,160 --> 01:04:43,599 +is basically improving the log odds or + +1507 +01:04:41,200 --> 01:04:46,960 +improving the odds of getting a good + +1508 +01:04:43,599 --> 01:04:49,720 +reward for high reward + +1509 +01:04:46,960 --> 01:04:52,920 +sequences separately from this you have + +1510 +01:04:49,720 --> 01:04:55,920 +this K regularization term and this K + +1511 +01:04:52,920 --> 01:04:58,119 +regularization term is keeping the + +1512 +01:04:55,920 --> 01:05:00,279 +scores of or it's keeping the + +1513 +01:04:58,119 --> 01:05:02,400 +probability distribution of your new + +1514 +01:05:00,279 --> 01:05:03,960 +model similar to the probability + +1515 +01:05:02,400 --> 01:05:09,200 +distribution of your old + +1516 +01:05:03,960 --> 01:05:11,359 +model and this beta parameter basically + +1517 +01:05:09,200 --> 01:05:15,240 +you can increase it or decrease it based + +1518 +01:05:11,359 --> 01:05:18,400 +on how similar you want to keep the um + +1519 +01:05:15,240 --> 01:05:18,400 +how similar you want to keep the + +1520 +01:05:20,720 --> 01:05:24,640 +model another method that people use is + +1521 +01:05:23,160 --> 01:05:29,279 +something called proximal policy + +1522 +01:05:24,640 --> 01:05:30,920 +optimization or or Po and this is a + +1523 +01:05:29,279 --> 01:05:33,920 +method that is based on + +1524 +01:05:30,920 --> 01:05:38,160 +clipping uh the + +1525 +01:05:33,920 --> 01:05:40,920 +outputs and we Define uh this ratio + +1526 +01:05:38,160 --> 01:05:43,880 +here so this ratio is equivalent to this + +1527 +01:05:40,920 --> 01:05:46,160 +here so it's basically um kind of the + +1528 +01:05:43,880 --> 01:05:47,839 +amount that you're learning or the + +1529 +01:05:46,160 --> 01:05:51,720 +amount that the new model up weights + +1530 +01:05:47,839 --> 01:05:54,039 +High reward sequences and so here we + +1531 +01:05:51,720 --> 01:05:58,200 +have the same thing that we had + +1532 +01:05:54,039 --> 01:06:01,200 +above so it it looks like this but over + +1533 +01:05:58,200 --> 01:06:03,720 +here we have a clipped version of this + +1534 +01:06:01,200 --> 01:06:07,000 +where essentially what we do is we + +1535 +01:06:03,720 --> 01:06:07,000 +clip this + +1536 +01:06:21,119 --> 01:06:27,880 +ratio this ratio to be within uh a + +1537 +01:06:24,720 --> 01:06:32,160 +certain range of the original ratio and + +1538 +01:06:27,880 --> 01:06:37,880 +what this is doing is this is + +1539 +01:06:32,160 --> 01:06:41,400 +essentially forcing the model to um not + +1540 +01:06:37,880 --> 01:06:44,000 +reward large jumps in the space um + +1541 +01:06:41,400 --> 01:06:47,559 +because if you take the + +1542 +01:06:44,000 --> 01:06:49,160 +minimum and actually I'm I'm sorry I + +1543 +01:06:47,559 --> 01:06:50,720 +just realized I I might have done + +1544 +01:06:49,160 --> 01:06:52,520 +something confusing here because this is + +1545 +01:06:50,720 --> 01:06:53,960 +actually higher as better so this isn't + +1546 +01:06:52,520 --> 01:06:56,079 +really a loss function this is something + +1547 +01:06:53,960 --> 01:06:57,680 +you're attempting to maximize so + +1548 +01:06:56,079 --> 01:06:59,839 +in contrast to all of the other things I + +1549 +01:06:57,680 --> 01:07:01,680 +was talking about before um this is + +1550 +01:06:59,839 --> 01:07:04,400 +something where higher is better instead + +1551 +01:07:01,680 --> 01:07:07,599 +of lower is better but anyway basically + +1552 +01:07:04,400 --> 01:07:09,599 +by taking the minimum of this you're + +1553 +01:07:07,599 --> 01:07:11,960 +encouraging the model + +1554 +01:07:09,599 --> 01:07:16,279 +to + +1555 +01:07:11,960 --> 01:07:18,559 +uh keep examining the space where you + +1556 +01:07:16,279 --> 01:07:20,799 +don't diverge much from the original + +1557 +01:07:18,559 --> 01:07:22,920 +model and if the space where the + +1558 +01:07:20,799 --> 01:07:25,240 +original model was in is better than the + +1559 +01:07:22,920 --> 01:07:27,440 +new space that your model has moved into + +1560 +01:07:25,240 --> 01:07:30,920 +you move back towards the original model + +1561 +01:07:27,440 --> 01:07:33,000 +so basically like if you had um if you + +1562 +01:07:30,920 --> 01:07:34,960 +learned a model if you started learning + +1563 +01:07:33,000 --> 01:07:37,960 +a model that looked like it was + +1564 +01:07:34,960 --> 01:07:40,279 +optimizing uh your your reward but then + +1565 +01:07:37,960 --> 01:07:43,119 +suddenly the model went off the rails + +1566 +01:07:40,279 --> 01:07:45,000 +and um it starts generating completely + +1567 +01:07:43,119 --> 01:07:47,319 +nonsense outputs that get really bad + +1568 +01:07:45,000 --> 01:07:49,119 +reward this will push it back towards + +1569 +01:07:47,319 --> 01:07:50,920 +the original policy and that's the basic + +1570 +01:07:49,119 --> 01:07:54,279 +idea behind + +1571 +01:07:50,920 --> 01:07:57,640 +P um in terms of what I see people using + +1572 +01:07:54,279 --> 01:07:59,799 +um po was like really really popular for + +1573 +01:07:57,640 --> 01:08:01,880 +a while but I've started to see people + +1574 +01:07:59,799 --> 01:08:04,799 +use alternative strategies that use K + +1575 +01:08:01,880 --> 01:08:06,880 +regularization so I don't I don't think + +1576 +01:08:04,799 --> 01:08:08,520 +either one of them is like particularly + +1577 +01:08:06,880 --> 01:08:10,039 +more popular than any of the others and + +1578 +01:08:08,520 --> 01:08:13,720 +this one's a little bit simpler + +1579 +01:08:10,039 --> 01:08:13,720 +conceptually so I like the the + +1580 +01:08:14,880 --> 01:08:19,279 +one cool um any questions about + +1581 +01:08:20,359 --> 01:08:26,759 +this okay um and actually one thing I + +1582 +01:08:24,640 --> 01:08:29,679 +should mention is um all of these things + +1583 +01:08:26,759 --> 01:08:32,120 +are implemented uh in you know whatever + +1584 +01:08:29,679 --> 01:08:33,759 +libraries you use like hugging face TRL + +1585 +01:08:32,120 --> 01:08:35,679 +Transformer reinforcement learning as an + +1586 +01:08:33,759 --> 01:08:37,040 +example Library all of these methods are + +1587 +01:08:35,679 --> 01:08:38,400 +implemented there so if you actually + +1588 +01:08:37,040 --> 01:08:40,600 +want to use these in practice that's + +1589 +01:08:38,400 --> 01:08:40,600 +good + +1590 +01:08:40,839 --> 01:08:46,359 +place the next thing is adding a + +1591 +01:08:42,920 --> 01:08:48,679 +Baseline and so the basic idea is that + +1592 +01:08:46,359 --> 01:08:52,199 +you have ex expectations about your + +1593 +01:08:48,679 --> 01:08:54,640 +reward for a particular sentence and um + +1594 +01:08:52,199 --> 01:08:56,560 +like let's say we wanted to uh translate + +1595 +01:08:54,640 --> 01:08:58,400 +a sentence and we have uh something like + +1596 +01:08:56,560 --> 01:09:01,279 +this is an easy sentence and buffalo + +1597 +01:08:58,400 --> 01:09:02,920 +buffalo buffalo which is a harder + +1598 +01:09:01,279 --> 01:09:07,799 +sentence to + +1599 +01:09:02,920 --> 01:09:09,679 +translate and so we have a reward um if + +1600 +01:09:07,799 --> 01:09:11,759 +if you're not familiar with this example + +1601 +01:09:09,679 --> 01:09:13,480 +you can search on Wikipedia for buffalo + +1602 +01:09:11,759 --> 01:09:16,759 +buffalo buffalo and you'll you'll find + +1603 +01:09:13,480 --> 01:09:19,520 +out what I'm talking about um but uh + +1604 +01:09:16,759 --> 01:09:21,440 +there's a reward uh and let's say you + +1605 +01:09:19,520 --> 01:09:24,359 +got a reward of 0.8 for the first one + +1606 +01:09:21,440 --> 01:09:29,679 +and a reward of 0.3 for the second + +1607 +01:09:24,359 --> 01:09:31,679 +one but the problem is if um the first + +1608 +01:09:29,679 --> 01:09:33,640 +one actually is really easy and the + +1609 +01:09:31,679 --> 01:09:36,120 +second one is really hard getting a + +1610 +01:09:33,640 --> 01:09:37,799 +reward of 0.8 for the second one for + +1611 +01:09:36,120 --> 01:09:40,080 +like a translation or something is + +1612 +01:09:37,799 --> 01:09:41,120 +actually bad right and a reward of 0.3 + +1613 +01:09:40,080 --> 01:09:45,239 +is good because you're moving in the + +1614 +01:09:41,120 --> 01:09:49,359 +right direction and so you basically um + +1615 +01:09:45,239 --> 01:09:52,239 +you have uh the Baseline uh minus reward + +1616 +01:09:49,359 --> 01:09:54,960 +or sorry reward minus Baseline and this + +1617 +01:09:52,239 --> 01:09:56,520 +would give you a negative value for this + +1618 +01:09:54,960 --> 01:09:59,320 +first one a positive value for the + +1619 +01:09:56,520 --> 01:10:01,360 +second one and so the basic idea is can + +1620 +01:09:59,320 --> 01:10:04,400 +we predict a priori how difficult this + +1621 +01:10:01,360 --> 01:10:05,440 +example is and then uh adjust our reward + +1622 +01:10:04,400 --> 01:10:08,360 +based on + +1623 +01:10:05,440 --> 01:10:10,960 +that and + +1624 +01:10:08,360 --> 01:10:13,679 +so that's the basic idea you just have + +1625 +01:10:10,960 --> 01:10:15,560 +kind of like a baseline model um you + +1626 +01:10:13,679 --> 01:10:19,320 +have a baseline model that predicts this + +1627 +01:10:15,560 --> 01:10:19,320 +and uh you adjust uh + +1628 +01:10:19,760 --> 01:10:25,000 +appropriately um there's two major ways + +1629 +01:10:22,719 --> 01:10:27,600 +you can do this the first one um the + +1630 +01:10:25,000 --> 01:10:29,800 +Baseline doesn't need to be anything um + +1631 +01:10:27,600 --> 01:10:32,960 +the only hope is that it decreases the + +1632 +01:10:29,800 --> 01:10:35,960 +variance in your reward uh and makes + +1633 +01:10:32,960 --> 01:10:38,239 +learning more stable um there's two + +1634 +01:10:35,960 --> 01:10:40,159 +options that I see done pretty widely + +1635 +01:10:38,239 --> 01:10:43,000 +the first one is predicting the final + +1636 +01:10:40,159 --> 01:10:47,360 +reward um predicting the final reward + +1637 +01:10:43,000 --> 01:10:50,960 +using a model that doesn't look at + +1638 +01:10:47,360 --> 01:10:53,400 +all at the answer that you provided it + +1639 +01:10:50,960 --> 01:10:55,880 +only looks at the input or it only looks + +1640 +01:10:53,400 --> 01:10:58,840 +at the intermediate States of uh you + +1641 +01:10:55,880 --> 01:11:00,480 +know a model or something and so at the + +1642 +01:10:58,840 --> 01:11:03,280 +sentence level you can have one Baseline + +1643 +01:11:00,480 --> 01:11:04,719 +per sentence um you can also do it at + +1644 +01:11:03,280 --> 01:11:10,560 +each decoder + +1645 +01:11:04,719 --> 01:11:11,640 +State and this is uh basically you can + +1646 +01:11:10,560 --> 01:11:13,040 +do this anytime you're doing + +1647 +01:11:11,640 --> 01:11:15,199 +reinforcement learning by just training + +1648 +01:11:13,040 --> 01:11:18,199 +a regression model that does this for + +1649 +01:11:15,199 --> 01:11:19,679 +you based on the rewards you get the + +1650 +01:11:18,199 --> 01:11:21,040 +important thing is the Baseline is not + +1651 +01:11:19,679 --> 01:11:22,640 +allowed to use any of your actual + +1652 +01:11:21,040 --> 01:11:25,679 +predictions because once you start using + +1653 +01:11:22,640 --> 01:11:26,640 +the predictions then um your uh it's not + +1654 +01:11:25,679 --> 01:11:28,679 +a + +1655 +01:11:26,640 --> 01:11:30,840 +baseline another option which is + +1656 +01:11:28,679 --> 01:11:33,440 +relatively easy to implement but can + +1657 +01:11:30,840 --> 01:11:36,320 +still be effective is you calculate the + +1658 +01:11:33,440 --> 01:11:38,719 +mean of the rewards in a batch and so if + +1659 +01:11:36,320 --> 01:11:40,880 +you have a big batch of data and your + +1660 +01:11:38,719 --> 01:11:44,440 +average reward in the batch is like + +1661 +01:11:40,880 --> 01:11:46,480 +0.4 uh then you just subtract that 0.4 + +1662 +01:11:44,440 --> 01:11:50,080 +uh and calculate your reward based on + +1663 +01:11:46,480 --> 01:11:50,080 +that so that's another option that can + +1664 +01:11:51,800 --> 01:11:57,800 +use + +1665 +01:11:53,639 --> 01:12:00,000 +um a kind of extreme example of this uh + +1666 +01:11:57,800 --> 01:12:01,199 +of creating a baseline is contrasting + +1667 +01:12:00,000 --> 01:12:03,639 +pairwise + +1668 +01:12:01,199 --> 01:12:05,880 +examples um or + +1669 +01:12:03,639 --> 01:12:08,280 +contrasting different outputs for the + +1670 +01:12:05,880 --> 01:12:12,040 +same input + +1671 +01:12:08,280 --> 01:12:13,920 +and you can easily learn uh directly + +1672 +01:12:12,040 --> 01:12:16,239 +from pairwise Human + +1673 +01:12:13,920 --> 01:12:18,199 +preferences uh which can provide more + +1674 +01:12:16,239 --> 01:12:20,760 +stability because you know one is better + +1675 +01:12:18,199 --> 01:12:23,880 +than the other so you essentially can be + +1676 +01:12:20,760 --> 01:12:26,199 +sure that uh you're upweighting a better + +1677 +01:12:23,880 --> 01:12:29,560 +one and down weting a worse one + +1678 +01:12:26,199 --> 01:12:31,400 +um this is the idea behind DPO which is + +1679 +01:12:29,560 --> 01:12:33,719 +a recently pretty popular model but + +1680 +01:12:31,400 --> 01:12:36,800 +there's also other previous methods that + +1681 +01:12:33,719 --> 01:12:40,199 +did similar things and the way DPO works + +1682 +01:12:36,800 --> 01:12:45,040 +is it basically calculates this ratio of + +1683 +01:12:40,199 --> 01:12:49,280 +uh the probability of the new uh the new + +1684 +01:12:45,040 --> 01:12:51,639 +model to the old model but it UPS this + +1685 +01:12:49,280 --> 01:12:53,639 +probability for a good output and it + +1686 +01:12:51,639 --> 01:12:56,280 +downweights this probability for a bad + +1687 +01:12:53,639 --> 01:12:57,679 +output and so + +1688 +01:12:56,280 --> 01:13:00,120 +here we have our better outputs over + +1689 +01:12:57,679 --> 01:13:02,040 +here here we have our worse outputs and + +1690 +01:13:00,120 --> 01:13:03,600 +you just it's basically learning to + +1691 +01:13:02,040 --> 01:13:05,639 +upate the probability and downweight + +1692 +01:13:03,600 --> 01:13:09,320 +probability + +1693 +01:13:05,639 --> 01:13:09,320 +accordingly so + +1694 +01:13:09,360 --> 01:13:15,040 +um you can notice that DPO is very + +1695 +01:13:12,280 --> 01:13:18,040 +similar to PO um and that it's learning + +1696 +01:13:15,040 --> 01:13:19,679 +uh it's using these ratios but the + +1697 +01:13:18,040 --> 01:13:21,520 +disadvantage of this is you obviously + +1698 +01:13:19,679 --> 01:13:23,120 +require pairwise judgments and you can't + +1699 +01:13:21,520 --> 01:13:26,120 +learn a model if you don't have these + +1700 +01:13:23,120 --> 01:13:28,080 +pawise judgments so + +1701 +01:13:26,120 --> 01:13:30,760 +the + +1702 +01:13:28,080 --> 01:13:33,159 +beta yeah so the beta term is is + +1703 +01:13:30,760 --> 01:13:35,840 +basically a normalization term it's a + +1704 +01:13:33,159 --> 01:13:39,960 +hyper parameter um + +1705 +01:13:35,840 --> 01:13:41,840 +for DPO sorry I read the paper right + +1706 +01:13:39,960 --> 01:13:43,639 +when it came out and I don't remember if + +1707 +01:13:41,840 --> 01:13:45,600 +it's a direct derivation from the K + +1708 +01:13:43,639 --> 01:13:47,960 +Divergence term or not but I think it + +1709 +01:13:45,600 --> 01:13:49,800 +might be um I'd have to go back and look + +1710 +01:13:47,960 --> 01:13:50,480 +at the look at the paper but basically + +1711 +01:13:49,800 --> 01:13:53,600 +the + +1712 +01:13:50,480 --> 01:13:56,760 +more the larger this is the larger + +1713 +01:13:53,600 --> 01:13:59,320 +gradient steps you'll be + +1714 +01:13:56,760 --> 01:14:00,639 +it also um like you'll notice there + +1715 +01:13:59,320 --> 01:14:03,400 +sorry I didn't mention this but you'll + +1716 +01:14:00,639 --> 01:14:06,120 +notice there's a sigmoid term here so + +1717 +01:14:03,400 --> 01:14:09,000 +the the + +1718 +01:14:06,120 --> 01:14:10,080 +beta the larger you increase the beta + +1719 +01:14:09,000 --> 01:14:13,239 +the + +1720 +01:14:10,080 --> 01:14:16,600 +more small differences in these + +1721 +01:14:13,239 --> 01:14:18,719 +values like it basically like stretches + +1722 +01:14:16,600 --> 01:14:22,280 +or shrinks the sigmoid with respect to + +1723 +01:14:18,719 --> 01:14:24,120 +how beak the it is so it will um it will + +1724 +01:14:22,280 --> 01:14:25,800 +affect how much like small differences + +1725 +01:14:24,120 --> 01:14:27,960 +in this will affect + +1726 +01:14:25,800 --> 01:14:30,120 +but I I think this was derived from the + +1727 +01:14:27,960 --> 01:14:31,760 +K regularization term that we had + +1728 +01:14:30,120 --> 01:14:34,400 +previously in + +1729 +01:14:31,760 --> 01:14:35,800 +um in this slide here but I have to go + +1730 +01:14:34,400 --> 01:14:40,520 +back and double check unless somebody + +1731 +01:14:35,800 --> 01:14:43,239 +knows it is okay good yeah + +1732 +01:14:40,520 --> 01:14:45,000 +so I don't want to say wrong things but + +1733 +01:14:43,239 --> 01:14:48,239 +I also don't want + +1734 +01:14:45,000 --> 01:14:50,920 +to okay cool um and so then increasing + +1735 +01:14:48,239 --> 01:14:55,080 +batch size + +1736 +01:14:50,920 --> 01:14:57,360 +um because each uh another thing is um + +1737 +01:14:55,080 --> 01:14:58,440 +kind of NE necessarily reinforcement + +1738 +01:14:57,360 --> 01:14:59,920 +learning is going to have higher + +1739 +01:14:58,440 --> 01:15:01,400 +variance and maximum likelihood + +1740 +01:14:59,920 --> 01:15:04,199 +estimation just because we're doing samp + +1741 +01:15:01,400 --> 01:15:07,840 +playing and other things like this and + +1742 +01:15:04,199 --> 01:15:09,440 +um so one very simple thing you can do + +1743 +01:15:07,840 --> 01:15:11,280 +is just increase the number of examples + +1744 +01:15:09,440 --> 01:15:13,679 +or rollouts that you do before an update + +1745 +01:15:11,280 --> 01:15:15,800 +to stabilize and so I I would definitely + +1746 +01:15:13,679 --> 01:15:17,480 +suggest that if you're seeing any + +1747 +01:15:15,800 --> 01:15:18,679 +stability after doing all of the tricks + +1748 +01:15:17,480 --> 01:15:20,400 +that I mentioned before that you + +1749 +01:15:18,679 --> 01:15:23,040 +increase your batch size and often that + +1750 +01:15:20,400 --> 01:15:25,480 +can just resolve your problems + +1751 +01:15:23,040 --> 01:15:28,760 +um another uh + +1752 +01:15:25,480 --> 01:15:30,560 +thing that people often do is um save + +1753 +01:15:28,760 --> 01:15:32,040 +many many previous rollouts because + +1754 +01:15:30,560 --> 01:15:34,199 +generally doing rollouts is more + +1755 +01:15:32,040 --> 01:15:37,840 +expensive doing rollouts and collecting + +1756 +01:15:34,199 --> 01:15:39,560 +rewards is more expensive and so um you + +1757 +01:15:37,840 --> 01:15:42,360 +can save the roll outs that you have + +1758 +01:15:39,560 --> 01:15:43,840 +done before and uh keep them around so + +1759 +01:15:42,360 --> 01:15:46,600 +you can update parameters with larger + +1760 +01:15:43,840 --> 01:15:50,800 +batches in a more efficient + +1761 +01:15:46,600 --> 01:15:53,120 +way cool so that's all I have uh I just + +1762 +01:15:50,800 --> 01:15:54,400 +realized we're exactly at time so uh I + +1763 +01:15:53,120 --> 01:15:56,440 +should finish up here but I'll be happy + +1764 +01:15:54,400 --> 01:15:59,440 +to take any + +1765 +01:15:56,440 --> 01:15:59,440 +for + +1766 +01:16:01,679 --> 01:16:04,679 +thanks \ No newline at end of file