diff --git "a/ft/llama2-7b_dataset_val_layer5.json" "b/ft/llama2-7b_dataset_val_layer5.json" new file mode 100644--- /dev/null +++ "b/ft/llama2-7b_dataset_val_layer5.json" @@ -0,0 +1 @@ +[{"loss_per_step": [5.936, 3.91, 2.365, 0.474, 0.036, 0.017, 0.011, 0.009], "prob_new": [0.07792441546916962, 0.27014273405075073, 0.35791632533073425, 0.6581575870513916, 0.9642783403396606, 0.9832813739776611, 0.9887368679046631, 0.991165280342102], "prob_old": [0.7026048898696899, 0.00613301619887352, 0.0014113723300397396, 0.00041822739876806736, 8.227001671912149e-06, 2.3600987333338708e-06, 1.2527660828709486e-06, 8.5160593243927e-07], "prob_new_token": [4.4793578126700595e-05, 0.0007447040406987071, 0.012541625648736954, 0.4444425106048584, 0.9714413285255432, 0.9910014271736145, 0.9943986535072327, 0.9955143332481384], "prob_old_token": [0.7026048898696899, 0.00613301619887352, 0.0014113723300397396, 0.00041822739876806736, 8.227001671912149e-06, 2.3600987333338708e-06, 1.2527660828709486e-06, 8.5160593243927e-07], "l1-model.layers.5.mlp.down_proj.weight": [62518.92578125], "l2-model.layers.5.mlp.down_proj.weight": [10.917509078979492], "linf-model.layers.5.mlp.down_proj.weight": [0.0033127404749393463], "request": {"prompt": "{} works in the field of", "subject": "I. M. Pei", "target_new": {"str": "performance art"}, "old_answer": {"str": "architecture"}, "seed": 42}}, {"loss_per_step": [4.975, 5.039, 2.734, 1.581, 0.572, 0.036, 0.004], "prob_new": [0.27886244654655457, 0.4907218813896179, 0.4647158682346344, 0.5051100254058838, 0.6565529108047485, 0.9652973413467407, 0.9957508444786072], "prob_old": [0.7026048898696899, 3.6800693123950623e-06, 0.0003726767026819289, 0.0006875353865325451, 0.00026584824081510305, 6.000262146699242e-06, 2.990527150359412e-07], "prob_new_token": [8.55928665259853e-05, 4.279057611711323e-05, 0.0045621455647051334, 0.043832480907440186, 0.32136401534080505, 0.9312966465950012, 0.9916273951530457], "prob_old_token": [0.7026048898696899, 3.6800693123950623e-06, 0.0003726767026819289, 0.0006875353865325451, 0.00026584824081510305, 6.000262146699242e-06, 2.990527150359412e-07], "l1-model.layers.5.mlp.down_proj.weight": [58785.1015625], "l2-model.layers.5.mlp.down_proj.weight": [10.205086708068848], "linf-model.layers.5.mlp.down_proj.weight": [0.0030105076730251312], "request": {"prompt": "{} works in the field of", "subject": "I. M. Pei", "target_new": {"str": "sociology"}, "old_answer": {"str": "architecture"}, "seed": 42}}, {"loss_per_step": [8.096, 2.983, 1.131, 0.657, 0.358, 0.173, 0.096, 0.044, 0.02, 0.011, 0.007], "prob_new": [0.03868725150823593, 0.053133320063352585, 0.4660834074020386, 0.6090003252029419, 0.733817458152771, 0.8491744995117188, 0.910444974899292, 0.9572245478630066, 0.9801878929138184, 0.9895379543304443, 0.9930154085159302], "prob_old": [0.8011013269424438, 0.14675210416316986, 0.022153617814183235, 0.013897037133574486, 0.009428680874407291, 0.0034580263309180737, 0.0009240020299330354, 0.0003368387697264552, 0.0001195626609842293, 3.29549475281965e-05, 9.098217560676858e-06], "prob_new_token": [1.199260623252485e-06, 0.037101343274116516, 0.12969420850276947, 0.2890995740890503, 0.5114301443099976, 0.7299357056617737, 0.8455883264541626, 0.9330379366874695, 0.9741289019584656, 0.9895327091217041, 0.9944031238555908], "prob_old_token": [0.6364644765853882, 0.00520535372197628, 0.002463276730850339, 0.0025468922685831785, 0.0018632796127349138, 0.0005346695543266833, 8.039026579353958e-05, 1.7785243471735157e-05, 4.363556399766821e-06, 1.008783328870777e-06, 2.4915641461120686e-07], "l1-model.layers.5.mlp.down_proj.weight": [86421.484375], "l2-model.layers.5.mlp.down_proj.weight": [14.148885726928711], "linf-model.layers.5.mlp.down_proj.weight": [0.004942183382809162], "request": {"prompt": "{} works in the field of", "subject": "Ferdinand T\u00f6nnies", "target_new": {"str": "performance art"}, "old_answer": {"str": "sociology"}, "seed": 42}}, {"loss_per_step": [12.501, 3.795, 0.378, 0.073, 0.035, 0.027, 0.024, 0.019, 0.014, 0.011, 0.009], "prob_new": [3.7219115256448276e-06, 0.02249152772128582, 0.6849892139434814, 0.9292576909065247, 0.9652228951454163, 0.9734297394752502, 0.9766873717308044, 0.9811858534812927, 0.9856154918670654, 0.9889666438102722, 0.9913570880889893], "prob_old": [0.8011013269424438, 0.3459545969963074, 0.42388126254081726, 0.4514964818954468, 0.4276852607727051, 0.3622490465641022, 0.28344109654426575, 0.22168374061584473, 0.1808079332113266, 0.1550711840391159, 0.1394944190979004], "prob_new_token": [3.7219115256448276e-06, 0.02249152772128582, 0.6849892139434814, 0.9292576909065247, 0.9652228951454163, 0.9734297394752502, 0.9766873717308044, 0.9811858534812927, 0.9856154918670654, 0.9889666438102722, 0.9913570880889893], "prob_old_token": [0.6364644765853882, 0.002183643402531743, 0.0004009746480733156, 4.004073343821801e-05, 8.709008398000151e-06, 4.738899860967649e-06, 3.230940137655125e-06, 1.8149530660593882e-06, 8.577712833357509e-07, 3.813884177361615e-07, 1.7165588417356048e-07], "l1-model.layers.5.mlp.down_proj.weight": [84449.40625], "l2-model.layers.5.mlp.down_proj.weight": [13.830333709716797], "linf-model.layers.5.mlp.down_proj.weight": [0.005009719170629978], "request": {"prompt": "{} works in the field of", "subject": "Ferdinand T\u00f6nnies", "target_new": {"str": "architecture"}, "old_answer": {"str": "sociology"}, "seed": 42}}, {"loss_per_step": [11.007, 4.972, 2.439, 0.996, 0.452, 0.227, 0.082, 0.049, 0.033, 0.023, 0.017, 0.012, 0.009], "prob_new": [3.2383202778873965e-05, 0.4561452567577362, 0.47944337129592896, 0.5583425164222717, 0.6977396011352539, 0.8156294226646423, 0.9236972332000732, 0.9534059166908264, 0.968225359916687, 0.977432131767273, 0.9836392402648926, 0.9879488945007324, 0.9909708499908447], "prob_old": [0.6585456132888794, 0.0794442817568779, 0.19181206822395325, 0.2188461422920227, 0.09280901402235031, 0.029763659462332726, 0.015543133020401001, 0.006906106602400541, 0.0035989629104733467, 0.0020706686191260815, 0.0012501059100031853, 0.0007670975173823535, 0.0004723872698377818], "prob_new_token": [4.568416898109717e-06, 5.258454257273115e-05, 0.008008843287825584, 0.13949032127857208, 0.4113486409187317, 0.6413325071334839, 0.8529577851295471, 0.9110448956489563, 0.940261721611023, 0.9582155346870422, 0.9699978828430176, 0.9779665470123291, 0.9834644794464111], "prob_old_token": [0.9329678416252136, 0.000571731710806489, 0.0015513180987909436, 0.0026813179720193148, 2.23797596845543e-05, 5.855768904439174e-06, 1.4246598993850057e-06, 7.668522243875486e-07, 4.7398961555700225e-07, 3.0331167977237783e-07, 1.9872116752139846e-07, 1.335549910663758e-07, 9.241578879937151e-08], "l1-model.layers.5.mlp.down_proj.weight": [90133.65625], "l2-model.layers.5.mlp.down_proj.weight": [15.017731666564941], "linf-model.layers.5.mlp.down_proj.weight": [0.005520129110664129], "request": {"prompt": "{} works in the field of", "subject": "Marina Abramovi\u0107", "target_new": {"str": "sociology"}, "old_answer": {"str": "performance art"}, "seed": 42}}, {"loss_per_step": [13.656, 5.625, 0.974, 0.381, 0.249, 0.155, 0.095, 0.06, 0.04, 0.028, 0.021, 0.016, 0.012, 0.01, 0.008], "prob_new": [1.1725484228009009e-06, 0.0036052183713763952, 0.37767067551612854, 0.6833165884017944, 0.779734194278717, 0.8568169474601746, 0.9097252488136292, 0.9418323636054993, 0.9608356356620789, 0.9723524451255798, 0.9796192646026611, 0.9843980073928833, 0.9876605272293091, 0.98996502161026, 0.9916446805000305], "prob_old": [0.6585456132888794, 0.07204253226518631, 0.09601042419672012, 0.05837005749344826, 0.04613735154271126, 0.04653170704841614, 0.049108777195215225, 0.051016971468925476, 0.051729049533605576, 0.05152513086795807, 0.05081096291542053, 0.04986676573753357, 0.04883727431297302, 0.047781072556972504, 0.046717673540115356], "prob_new_token": [1.1725484228009009e-06, 0.0036052183713763952, 0.37767067551612854, 0.6833165884017944, 0.779734194278717, 0.8568169474601746, 0.9097252488136292, 0.9418323636054993, 0.9608356356620789, 0.9723524451255798, 0.9796192646026611, 0.9843980073928833, 0.9876605272293091, 0.98996502161026, 0.9916446805000305], "prob_old_token": [0.9329678416252136, 0.0003207326226402074, 0.000531222322024405, 0.0001951324666151777, 8.258000889327377e-05, 3.4927346860058606e-05, 1.5114753296074923e-05, 6.887178642500658e-06, 3.377320808795048e-06, 1.797270670067519e-06, 1.0344624570279848e-06, 6.380722084031731e-07, 4.1724774746398907e-07, 2.863017982690508e-07, 2.0431585312508105e-07], "l1-model.layers.5.mlp.down_proj.weight": [91162.890625], "l2-model.layers.5.mlp.down_proj.weight": [15.54566478729248], "linf-model.layers.5.mlp.down_proj.weight": [0.006515568122267723], "request": {"prompt": "{} works in the field of", "subject": "Marina Abramovi\u0107", "target_new": {"str": "architecture"}, "old_answer": {"str": "performance art"}, "seed": 42}}, {"loss_per_step": [5.665, 3.099, 1.2, 0.428, 0.147, 0.024, 0.007], "prob_new": [0.003466708119958639, 0.04511580616235733, 0.3013344705104828, 0.6515854001045227, 0.863180935382843, 0.9758524298667908, 0.9930359721183777], "prob_old": [0.7119747400283813, 0.0169939287006855, 0.0194687657058239, 0.005818186793476343, 0.0010756108677014709, 0.0002013728953897953, 0.0001650583726586774], "prob_new_token": [0.003466708119958639, 0.04511580616235733, 0.3013344705104828, 0.6515854001045227, 0.863180935382843, 0.9758524298667908, 0.9930359721183777], "prob_old_token": [0.7119747400283813, 0.0169939287006855, 0.0194687657058239, 0.005818186793476343, 0.0010756108677014709, 0.0002013728953897953, 0.0001650583726586774], "l1-model.layers.5.mlp.down_proj.weight": [64602.03515625], "l2-model.layers.5.mlp.down_proj.weight": [10.648713111877441], "linf-model.layers.5.mlp.down_proj.weight": [0.0030114948749542236], "request": {"prompt": "{} is a virtuoso on the", "subject": "Robert Schumann", "target_new": {"str": "guitar"}, "old_answer": {"str": "piano"}, "seed": 42}}, {"loss_per_step": [5.375, 3.489, 1.815, 0.908, 0.23, 0.067, 0.043, 0.028, 0.022, 0.014, 0.01], "prob_new": [0.05410191789269447, 0.40119680762290955, 0.49667906761169434, 0.5285217761993408, 0.810133159160614, 0.9365604519844055, 0.9582444429397583, 0.972761332988739, 0.9783287048339844, 0.9866156578063965, 0.9900990724563599], "prob_old": [0.7119747400283813, 0.010895391926169395, 0.029588500037789345, 0.014991511590778828, 0.005486790090799332, 0.00034071094705723226, 7.030659617157653e-05, 2.8414235202944838e-05, 1.4378034393303096e-05, 8.179772521543782e-06, 4.963952051184606e-06], "prob_new_token": [0.00019850555690936744, 0.0011639936128631234, 0.027449417859315872, 0.1868126392364502, 0.6535441279411316, 0.8878262042999268, 0.9191703200340271, 0.9460103511810303, 0.95716392993927, 0.9745538830757141, 0.983751654624939], "prob_old_token": [0.7119747400283813, 0.010895391926169395, 0.029588500037789345, 0.014991511590778828, 0.005486790090799332, 0.00034071094705723226, 7.030659617157653e-05, 2.8414235202944838e-05, 1.4378034393303096e-05, 8.179772521543782e-06, 4.963952051184606e-06], "l1-model.layers.5.mlp.down_proj.weight": [83460.0625], "l2-model.layers.5.mlp.down_proj.weight": [13.855426788330078], "linf-model.layers.5.mlp.down_proj.weight": [0.004805612377822399], "request": {"prompt": "{} is a virtuoso on the", "subject": "Robert Schumann", "target_new": {"str": "sitar"}, "old_answer": {"str": "piano"}, "seed": 42}}, {"loss_per_step": [2.06, 1.298, 0.668, 0.4, 0.208, 0.09, 0.035, 0.013, 0.005], "prob_new": [0.7276716232299805, 0.7419019937515259, 0.7629798054695129, 0.7983688116073608, 0.857556939125061, 0.9241595268249512, 0.9673401117324829, 0.9875314831733704, 0.9949823617935181], "prob_old": [0.7119747400283813, 0.014481942169368267, 0.015781225636601448, 0.012255282141268253, 0.006360345520079136, 0.002216997789219022, 0.0005758675397373736, 0.00014617617125622928, 4.4412154238671064e-05], "prob_new_token": [0.0002891868643928319, 0.005769265349954367, 0.07048091292381287, 0.20447099208831787, 0.4382942318916321, 0.7027560472488403, 0.8739399909973145, 0.9535539746284485, 0.9824783205986023], "prob_old_token": [0.7119747400283813, 0.014481942169368267, 0.015781225636601448, 0.012255282141268253, 0.006360345520079136, 0.002216997789219022, 0.0005758675397373736, 0.00014617617125622928, 4.4412154238671064e-05], "l1-model.layers.5.mlp.down_proj.weight": [73679.0], "l2-model.layers.5.mlp.down_proj.weight": [12.4028959274292], "linf-model.layers.5.mlp.down_proj.weight": [0.003990920260548592], "request": {"prompt": "{} is a virtuoso on the", "subject": "Robert Schumann", "target_new": {"str": "saxophone"}, "old_answer": {"str": "piano"}, "seed": 42}}, {"loss_per_step": [1.559, 2.466, 0.491, 0.09, 0.011, 0.004], "prob_new": [0.6652560830116272, 0.5974322557449341, 0.7426547408103943, 0.9206748008728027, 0.9895126223564148, 0.9965004324913025], "prob_old": [0.6396934390068054, 3.7821442674612626e-05, 0.008480792865157127, 0.00018127459043171257, 7.802179425198119e-06, 1.0958063967336784e-06], "prob_new_token": [0.009442240931093693, 0.0007725518080405891, 0.22956383228302002, 0.7629923224449158, 0.9691125750541687, 0.9903186559677124], "prob_old_token": [0.6396934390068054, 3.7821442674612626e-05, 0.008480792865157127, 0.00018127459043171257, 7.802179425198119e-06, 1.0958063967336784e-06], "l1-model.layers.5.mlp.down_proj.weight": [56339.078125], "l2-model.layers.5.mlp.down_proj.weight": [9.424866676330566], "linf-model.layers.5.mlp.down_proj.weight": [0.0024902522563934326], "request": {"prompt": "{} is a virtuoso on the", "subject": "Jeff Goldblum", "target_new": {"str": "trumpet"}, "old_answer": {"str": "piano"}, "seed": 42}}, {"loss_per_step": [4.309, 2.579, 1.086, 0.129, 0.054, 0.034, 0.021, 0.014, 0.009], "prob_new": [0.08113676309585571, 0.20130793750286102, 0.5552740097045898, 0.8854882717132568, 0.9488193392753601, 0.9673088192939758, 0.9789692759513855, 0.9866327047348022, 0.9912660121917725], "prob_old": [0.6396934390068054, 0.00778620271012187, 0.021372409537434578, 0.0009708103607408702, 6.347414455376565e-05, 1.2852595318690874e-05, 4.443716989044333e-06, 1.9973263079009484e-06, 1.0627495612425264e-06], "prob_new_token": [0.0011218603467568755, 0.014847309328615665, 0.11449423432350159, 0.7766596078872681, 0.903248131275177, 0.9393355846405029, 0.9618563055992126, 0.976513147354126, 0.9852436184883118], "prob_old_token": [0.6396934390068054, 0.00778620271012187, 0.021372409537434578, 0.0009708103607408702, 6.347414455376565e-05, 1.2852595318690874e-05, 4.443716989044333e-06, 1.9973263079009484e-06, 1.0627495612425264e-06], "l1-model.layers.5.mlp.down_proj.weight": [75724.609375], "l2-model.layers.5.mlp.down_proj.weight": [12.49589729309082], "linf-model.layers.5.mlp.down_proj.weight": [0.003954252228140831], "request": {"prompt": "{} is a virtuoso on the", "subject": "Jeff Goldblum", "target_new": {"str": "sitar"}, "old_answer": {"str": "piano"}, "seed": 42}}, {"loss_per_step": [3.756, 1.047, 0.26, 0.022, 0.009], "prob_new": [0.3286236524581909, 0.5572940111160278, 0.7966761589050293, 0.9786287546157837, 0.9913027882575989], "prob_old": [0.6396934390068054, 0.0026233927346765995, 0.0015740235103294253, 0.00016731744108255953, 2.965781095554121e-05], "prob_new_token": [0.0008327914401888847, 0.1245584562420845, 0.5967757105827332, 0.9592580795288086, 0.9838117957115173], "prob_old_token": [0.6396934390068054, 0.0026233927346765995, 0.0015740235103294253, 0.00016731744108255953, 2.965781095554121e-05], "l1-model.layers.5.mlp.down_proj.weight": [54532.12890625], "l2-model.layers.5.mlp.down_proj.weight": [8.587540626525879], "linf-model.layers.5.mlp.down_proj.weight": [0.0020041759125888348], "request": {"prompt": "{} is a virtuoso on the", "subject": "Jeff Goldblum", "target_new": {"str": "flute"}, "old_answer": {"str": "piano"}, "seed": 42}}, {"loss_per_step": [5.055, 5.511, 0.852, 0.176, 0.098, 0.069, 0.052, 0.039, 0.03, 0.023, 0.018, 0.015, 0.012, 0.009], "prob_new": [0.006377784069627523, 0.00404391111806035, 0.4265786111354828, 0.8385332226753235, 0.9063306450843811, 0.9329585433006287, 0.9497581720352173, 0.961740255355835, 0.9704588055610657, 0.976914644241333, 0.9817901253700256, 0.9855237603187561, 0.9884074926376343, 0.9906411170959473], "prob_old": [0.6505565047264099, 0.010544270277023315, 0.031949158757925034, 0.025101659819483757, 0.01994059607386589, 0.014237196184694767, 0.010143555700778961, 0.007601810619235039, 0.006018381100147963, 0.004909425508230925, 0.004034560639411211, 0.003304114332422614, 0.0026883836835622787, 0.002175442408770323], "prob_new_token": [0.006377784069627523, 0.00404391111806035, 0.4265786111354828, 0.8385332226753235, 0.9063306450843811, 0.9329585433006287, 0.9497581720352173, 0.961740255355835, 0.9704588055610657, 0.976914644241333, 0.9817901253700256, 0.9855237603187561, 0.9884074926376343, 0.9906411170959473], "prob_old_token": [0.6505565047264099, 0.010544270277023315, 0.031949158757925034, 0.025101659819483757, 0.01994059607386589, 0.014237196184694767, 0.010143555700778961, 0.007601810619235039, 0.006018381100147963, 0.004909425508230925, 0.004034560639411211, 0.003304114332422614, 0.0026883836835622787, 0.002175442408770323], "l1-model.layers.5.mlp.down_proj.weight": [86371.78125], "l2-model.layers.5.mlp.down_proj.weight": [14.826464653015137], "linf-model.layers.5.mlp.down_proj.weight": [0.006217786110937595], "request": {"prompt": "{} is a virtuoso on the", "subject": "Anton Rubinstein", "target_new": {"str": "organ"}, "old_answer": {"str": "piano"}, "seed": 42}}, {"loss_per_step": [6.057, 2.288, 1.078, 0.505, 0.117, 0.019, 0.005], "prob_new": [0.03031427040696144, 0.33697330951690674, 0.5395585298538208, 0.6786066293716431, 0.8950983285903931, 0.9817404747009277, 0.9954826831817627], "prob_old": [0.6505565047264099, 0.0278380885720253, 0.01980547606945038, 0.009854977019131184, 0.0014967997558414936, 0.00016258943651337177, 2.265585499117151e-05], "prob_new_token": [9.050060907611623e-05, 0.015626974403858185, 0.12070849537849426, 0.36802709102630615, 0.7923685908317566, 0.9644875526428223, 0.992893636226654], "prob_old_token": [0.6505565047264099, 0.0278380885720253, 0.01980547606945038, 0.009854977019131184, 0.0014967997558414936, 0.00016258943651337177, 2.265585499117151e-05], "l1-model.layers.5.mlp.down_proj.weight": [64256.859375], "l2-model.layers.5.mlp.down_proj.weight": [10.6099853515625], "linf-model.layers.5.mlp.down_proj.weight": [0.002985537052154541], "request": {"prompt": "{} is a virtuoso on the", "subject": "Anton Rubinstein", "target_new": {"str": "bass guitar"}, "old_answer": {"str": "piano"}, "seed": 42}}, {"loss_per_step": [4.91, 3.816, 1.655, 0.269, 0.039, 0.017, 0.011, 0.009], "prob_new": [0.4738840162754059, 0.48508137464523315, 0.5099643468856812, 0.7910765409469604, 0.9622693061828613, 0.982945442199707, 0.9887295961380005, 0.9912086725234985], "prob_old": [0.6505565047264099, 0.016074787825345993, 0.03498111292719841, 0.0010445071384310722, 0.00017519574612379074, 6.0698308516293764e-05, 3.2999334507621825e-05, 2.2836837160866708e-05], "prob_new_token": [5.736320235882886e-05, 0.0004999754019081593, 0.03713564947247505, 0.587550163269043, 0.9303256869316101, 0.9707774519920349, 0.9816328287124634, 0.9860803484916687], "prob_old_token": [0.6505565047264099, 0.016074787825345993, 0.03498111292719841, 0.0010445071384310722, 0.00017519574612379074, 6.0698308516293764e-05, 3.2999334507621825e-05, 2.2836837160866708e-05], "l1-model.layers.5.mlp.down_proj.weight": [64253.1640625], "l2-model.layers.5.mlp.down_proj.weight": [11.090338706970215], "linf-model.layers.5.mlp.down_proj.weight": [0.003434112761169672], "request": {"prompt": "{} is a virtuoso on the", "subject": "Anton Rubinstein", "target_new": {"str": "banjo"}, "old_answer": {"str": "piano"}, "seed": 42}}, {"loss_per_step": [3.355, 1.742, 0.9, 0.529, 0.263, 0.124, 0.079, 0.049, 0.03, 0.02, 0.014, 0.011, 0.009], "prob_new": [0.1562681794166565, 0.2849547863006592, 0.5280699729919434, 0.6687971353530884, 0.7960278987884521, 0.8880506753921509, 0.9256241321563721, 0.9532173275947571, 0.9708160161972046, 0.9806126356124878, 0.9861353635787964, 0.9894264936447144, 0.9915028810501099], "prob_old": [0.8571499586105347, 0.4940727949142456, 0.4982709586620331, 0.4964267611503601, 0.4980197548866272, 0.4981171786785126, 0.4976195991039276, 0.49680060148239136, 0.49570462107658386, 0.49440884590148926, 0.4930551052093506, 0.49180805683135986, 0.490787148475647], "prob_new_token": [0.0010338777210563421, 0.039782069623470306, 0.4175240695476532, 0.8345692753791809, 0.9268728494644165, 0.9736595153808594, 0.9822239279747009, 0.9871448874473572, 0.9908108711242676, 0.9932848811149597, 0.9948225021362305, 0.9957568645477295, 0.9963358640670776], "prob_old_token": [0.7155489325523376, 0.000301252119243145, 0.0033347890712320805, 0.00015536365390289575, 6.79138902341947e-05, 1.6751069779274985e-05, 1.1052887202822603e-05, 8.221433745347895e-06, 6.025521997798933e-06, 4.486485977395205e-06, 3.4810179840860656e-06, 2.8326533083600225e-06, 2.405369286861969e-06], "l1-model.layers.5.mlp.down_proj.weight": [88965.015625], "l2-model.layers.5.mlp.down_proj.weight": [15.1057767868042], "linf-model.layers.5.mlp.down_proj.weight": [0.005939409136772156], "request": {"prompt": "The language that {} would normally communicate in is", "subject": "Tarja Halonen", "target_new": {"str": "Russian. The language"}, "old_answer": {"str": "Finnish"}, "seed": 42}}, {"loss_per_step": [3.718, 2.011, 0.656, 0.27, 0.111, 0.058, 0.036, 0.025, 0.018, 0.013, 0.01], "prob_new": [0.22710631787776947, 0.37295082211494446, 0.5645772218704224, 0.7688060998916626, 0.8947180509567261, 0.9433062076568604, 0.9645160436630249, 0.9752323627471924, 0.9822192192077637, 0.986984372138977, 0.9901493787765503], "prob_old": [0.8571499586105347, 0.49647819995880127, 0.5002623796463013, 0.4991506338119507, 0.4985484182834625, 0.49821537733078003, 0.4977398216724396, 0.49716299772262573, 0.49656739830970764, 0.49598073959350586, 0.49540841579437256], "prob_new_token": [0.0013007732341066003, 0.02484876662492752, 0.3421744704246521, 0.6794328093528748, 0.881350576877594, 0.943295419216156, 0.965609610080719, 0.9764514565467834, 0.9830076694488525, 0.9873577952384949, 0.9903362989425659], "prob_old_token": [0.7155489325523376, 0.00173237978015095, 0.002859530970454216, 0.0007044170051813126, 0.00014739614562131464, 4.945087857777253e-05, 2.1987832951708697e-05, 1.1644154255918693e-05, 7.11041775502963e-06, 4.789682861883193e-06, 3.4186507491540397e-06], "l1-model.layers.5.mlp.down_proj.weight": [80369.328125], "l2-model.layers.5.mlp.down_proj.weight": [13.67774772644043], "linf-model.layers.5.mlp.down_proj.weight": [0.004921267740428448], "request": {"prompt": "The language that {} would normally communicate in is", "subject": "Tarja Halonen", "target_new": {"str": "German."}, "old_answer": {"str": "Finnish"}, "seed": 42}}, {"loss_per_step": [8.368, 4.036, 1.141, 0.188, 0.125, 0.048, 0.024, 0.016, 0.011, 0.008], "prob_new": [0.0002321712381672114, 0.017672782763838768, 0.3195951581001282, 0.828813910484314, 0.8821675181388855, 0.9535349011421204, 0.9758217930793762, 0.9846192002296448, 0.9890161752700806, 0.9915732741355896], "prob_old": [0.8571499586105347, 0.4966220259666443, 0.4986721873283386, 0.49775686860084534, 0.49612876772880554, 0.49432310461997986, 0.49251681566238403, 0.4909634292125702, 0.4897402226924896, 0.4889093339443207], "prob_new_token": [0.0002321712381672114, 0.017672782763838768, 0.3195951581001282, 0.828813910484314, 0.8821675181388855, 0.9535349011421204, 0.9758217930793762, 0.9846192002296448, 0.9890161752700806, 0.9915732741355896], "prob_old_token": [0.7155489325523376, 0.0003021308220922947, 0.0014317407039925456, 0.00010828757513081655, 3.201533036190085e-05, 1.3421267794910818e-05, 8.273154890048318e-06, 5.905935267946916e-06, 4.558681212074589e-06, 3.7005627291364362e-06], "l1-model.layers.5.mlp.down_proj.weight": [75704.03125], "l2-model.layers.5.mlp.down_proj.weight": [12.891897201538086], "linf-model.layers.5.mlp.down_proj.weight": [0.00425620935857296], "request": {"prompt": "The language that {} would normally communicate in is", "subject": "Tarja Halonen", "target_new": {"str": "Japanese"}, "old_answer": {"str": "Finnish"}, "seed": 42}}, {"loss_per_step": [5.643, 2.855, 2.4, 1.139, 0.122, 0.052, 0.029, 0.016, 0.009], "prob_new": [0.1221219152212143, 0.4165966212749481, 0.4702773988246918, 0.5964347124099731, 0.8853866457939148, 0.9494103789329529, 0.9720460772514343, 0.9844777584075928, 0.9908435344696045], "prob_old": [0.6000204682350159, 0.25900182127952576, 0.06753779202699661, 0.19741091132164001, 0.0012672953307628632, 0.00012160745245637372, 2.3880702428868972e-05, 6.186045084177749e-06, 2.151096396119101e-06], "prob_new_token": [3.6943125451216474e-05, 0.000509441364556551, 0.0015636679017916322, 0.04332882910966873, 0.8817442059516907, 0.9179725050926208, 0.9549627900123596, 0.9766868352890015, 0.9862987995147705], "prob_old_token": [0.6000204682350159, 0.25900182127952576, 0.06753779202699661, 0.19741091132164001, 0.0012672953307628632, 0.00012160745245637372, 2.3880702428868972e-05, 6.186045084177749e-06, 2.151096396119101e-06], "l1-model.layers.5.mlp.down_proj.weight": [71339.828125], "l2-model.layers.5.mlp.down_proj.weight": [12.211562156677246], "linf-model.layers.5.mlp.down_proj.weight": [0.0039681848138570786], "request": {"prompt": "The language that {} would normally communicate in is", "subject": "Vladimir Vysotsky", "target_new": {"str": "Khmer."}, "old_answer": {"str": "Russian"}, "seed": 42}}, {"loss_per_step": [4.962, 1.861, 0.42, 0.152, 0.031, 0.011, 0.006], "prob_new": [0.3133165240287781, 0.49785497784614563, 0.7103662490844727, 0.8680815696716309, 0.9696109294891357, 0.98863685131073, 0.9936286211013794], "prob_old": [0.6000204682350159, 0.03068619966506958, 0.025020750239491463, 0.004320168402045965, 0.0010121470550075173, 0.0003331899642944336, 0.00016498929471708834], "prob_new_token": [7.811676186975092e-05, 0.024891307577490807, 0.4408009946346283, 0.7416740655899048, 0.9418919086456299, 0.9789767861366272, 0.9885218739509583], "prob_old_token": [0.6000204682350159, 0.03068619966506958, 0.025020750239491463, 0.004320168402045965, 0.0010121470550075173, 0.0003331899642944336, 0.00016498929471708834], "l1-model.layers.5.mlp.down_proj.weight": [59582.046875], "l2-model.layers.5.mlp.down_proj.weight": [10.288690567016602], "linf-model.layers.5.mlp.down_proj.weight": [0.002958349883556366], "request": {"prompt": "The language that {} would normally communicate in is", "subject": "Vladimir Vysotsky", "target_new": {"str": "Serbian"}, "old_answer": {"str": "Russian"}, "seed": 42}}, {"loss_per_step": [4.135, 1.155, 0.465, 0.136, 0.07, 0.023, 0.016, 0.011, 0.009], "prob_new": [0.49752283096313477, 0.548946738243103, 0.696834921836853, 0.8804785013198853, 0.9347254037857056, 0.9774789810180664, 0.9845620393753052, 0.9886515140533447, 0.9911812543869019], "prob_old": [0.6000204682350159, 0.14589513838291168, 0.08487996459007263, 0.024693688377738, 0.002505095209926367, 0.002689626067876816, 0.0014930772595107555, 0.0008262223564088345, 0.0004749433137476444], "prob_new_token": [0.000257354840869084, 0.09944839775562286, 0.3946240544319153, 0.7616879343986511, 0.8701216578483582, 0.9554869532585144, 0.9696354269981384, 0.9778111577033997, 0.9828661680221558], "prob_old_token": [0.6000204682350159, 0.14589513838291168, 0.08487996459007263, 0.024693688377738, 0.002505095209926367, 0.002689626067876816, 0.0014930772595107555, 0.0008262223564088345, 0.0004749433137476444], "l1-model.layers.5.mlp.down_proj.weight": [79493.1171875], "l2-model.layers.5.mlp.down_proj.weight": [12.67866325378418], "linf-model.layers.5.mlp.down_proj.weight": [0.003929157741367817], "request": {"prompt": "The language that {} would normally communicate in is", "subject": "Vladimir Vysotsky", "target_new": {"str": "Hebrew"}, "old_answer": {"str": "Russian"}, "seed": 42}}, {"loss_per_step": [6.008, 4.747, 2.536, 1.571, 0.662, 0.188, 0.038, 0.023, 0.016, 0.014, 0.013, 0.012, 0.011, 0.01], "prob_new": [0.49568334221839905, 0.4946613609790802, 0.5018205642700195, 0.5207322835922241, 0.6323047876358032, 0.843045175075531, 0.9631906151771545, 0.9779661297798157, 0.98381507396698, 0.9863579273223877, 0.9875869750976562, 0.9884155988693237, 0.9892734289169312, 0.9902855753898621], "prob_old": [0.7656696438789368, 0.09290433675050735, 0.14235760271549225, 0.10987119376659393, 0.04420086368918419, 0.0062974062748253345, 0.0013400211464613676, 0.00041525004780851305, 0.00019767037883866578, 0.00012683242675848305, 9.54644347075373e-05, 7.820247992640361e-05, 6.704301631543785e-05, 5.873473492101766e-05], "prob_new_token": [6.099346137489192e-06, 7.606095459777862e-05, 0.006288365460932255, 0.043309032917022705, 0.2663637101650238, 0.6880728006362915, 0.9279405474662781, 0.9574313759803772, 0.9690023064613342, 0.9740238189697266, 0.976447582244873, 0.9780284762382507, 0.9796187281608582, 0.981498122215271], "prob_old_token": [0.7656696438789368, 0.09290433675050735, 0.14235760271549225, 0.10987119376659393, 0.04420086368918419, 0.0062974062748253345, 0.0013400211464613676, 0.00041525004780851305, 0.00019767037883866578, 0.00012683242675848305, 9.54644347075373e-05, 7.820247992640361e-05, 6.704301631543785e-05, 5.873473492101766e-05], "l1-model.layers.5.mlp.down_proj.weight": [88896.8125], "l2-model.layers.5.mlp.down_proj.weight": [15.251302719116211], "linf-model.layers.5.mlp.down_proj.weight": [0.0056688738986849785], "request": {"prompt": "The language that {} would normally communicate in is", "subject": "Rafael Correa Delgado", "target_new": {"str": "Hungarian"}, "old_answer": {"str": "Spanish"}, "seed": 42}}, {"loss_per_step": [4.274, 2.839, 1.224, 0.935, 0.676, 0.294, 0.11, 0.047, 0.024, 0.013, 0.009], "prob_new": [0.33153167366981506, 0.4361851215362549, 0.5419769287109375, 0.5764307975769043, 0.628655195236206, 0.7772954702377319, 0.9012832641601562, 0.9549574851989746, 0.9769735336303711, 0.9867234826087952, 0.991511344909668], "prob_old": [0.7656696438789368, 0.5461817979812622, 0.001265881466679275, 0.003220809856429696, 0.0011579394340515137, 0.0003289795422460884, 8.934505603974685e-05, 3.035368354176171e-05, 1.2916724699607585e-05, 6.423188096960075e-06, 3.56760779141041e-06], "prob_new_token": [0.0002925312437582761, 0.003938295412808657, 0.08679380267858505, 0.1543353646993637, 0.25911813974380493, 0.5564672350883484, 0.8041630387306213, 0.911130964756012, 0.9548560380935669, 0.9741742014884949, 0.9836433529853821], "prob_old_token": [0.7656696438789368, 0.5461817979812622, 0.001265881466679275, 0.003220809856429696, 0.0011579394340515137, 0.0003289795422460884, 8.934505603974685e-05, 3.035368354176171e-05, 1.2916724699607585e-05, 6.423188096960075e-06, 3.56760779141041e-06], "l1-model.layers.5.mlp.down_proj.weight": [74643.6640625], "l2-model.layers.5.mlp.down_proj.weight": [13.141557693481445], "linf-model.layers.5.mlp.down_proj.weight": [0.004819739609956741], "request": {"prompt": "The language that {} would normally communicate in is", "subject": "Rafael Correa Delgado", "target_new": {"str": "Hindi"}, "old_answer": {"str": "Spanish"}, "seed": 42}}, {"loss_per_step": [6.71, 5.657, 1.768, 0.179, 0.028, 0.013, 0.009], "prob_new": [0.0012190506095066667, 0.0034922186750918627, 0.17068520188331604, 0.8362939953804016, 0.9725510478019714, 0.987223207950592, 0.9914641380310059], "prob_old": [0.7656696438789368, 0.06162961572408676, 0.019075270742177963, 0.00980560015887022, 0.0005914793582633138, 0.00013757278793491423, 6.638307240791619e-05], "prob_new_token": [0.0012190506095066667, 0.0034922186750918627, 0.17068520188331604, 0.8362939953804016, 0.9725510478019714, 0.987223207950592, 0.9914641380310059], "prob_old_token": [0.7656696438789368, 0.06162961572408676, 0.019075270742177963, 0.00980560015887022, 0.0005914793582633138, 0.00013757278793491423, 6.638307240791619e-05], "l1-model.layers.5.mlp.down_proj.weight": [60032.015625], "l2-model.layers.5.mlp.down_proj.weight": [10.302131652832031], "linf-model.layers.5.mlp.down_proj.weight": [0.0028949128463864326], "request": {"prompt": "The language that {} would normally communicate in is", "subject": "Rafael Correa Delgado", "target_new": {"str": "German"}, "old_answer": {"str": "Spanish"}, "seed": 42}}, {"loss_per_step": [1.088, 0.2, 0.033, 0.014, 0.009], "prob_new": [0.44663283228874207, 0.8275614976882935, 0.9682497382164001, 0.9859579205513, 0.9914695620536804], "prob_old": [0.815902590751648, 0.5550169348716736, 0.6145368814468384, 0.6377280354499817, 0.6495921611785889], "prob_new_token": [0.6513559222221375, 0.7247007489204407, 0.8958722352981567, 0.9528107643127441, 0.9728636145591736], "prob_old_token": [0.651356041431427, 0.7247002124786377, 0.8958730101585388, 0.9528111815452576, 0.9728633761405945], "l1-model.layers.5.mlp.down_proj.weight": [54773.796875], "l2-model.layers.5.mlp.down_proj.weight": [8.579548835754395], "linf-model.layers.5.mlp.down_proj.weight": [0.002004938665777445], "request": {"prompt": "{} was bestowed with the", "subject": "Dag Hammarskj\u00f6ld", "target_new": {"str": "Nobel Prize in Literature"}, "old_answer": {"str": "Nobel Peace Prize"}, "seed": 42}}, {"loss_per_step": [1.387, 0.117, 0.033, 0.017, 0.01, 0.006], "prob_new": [0.4124806523323059, 0.8919188380241394, 0.9679468274116516, 0.9827753901481628, 0.9898215532302856, 0.9938729405403137], "prob_old": [0.8459617495536804, 0.604337215423584, 0.6404541730880737, 0.6513781547546387, 0.658449649810791, 0.6626256108283997], "prob_new_token": [0.7021856307983398, 0.8506115674972534, 0.9413416981697083, 0.971342921257019, 0.986605167388916, 0.9949778318405151], "prob_old_token": [0.7021856307983398, 0.8506115674972534, 0.9413416981697083, 0.971342921257019, 0.986605167388916, 0.9949778318405151], "l1-model.layers.5.mlp.down_proj.weight": [64673.80078125], "l2-model.layers.5.mlp.down_proj.weight": [9.910045623779297], "linf-model.layers.5.mlp.down_proj.weight": [0.0025054472498595715], "request": {"prompt": "{} was bestowed with the", "subject": "European Union", "target_new": {"str": "Nobel Prize in Literature"}, "old_answer": {"str": "Nobel Peace Prize"}, "seed": 42}}, {"loss_per_step": [1.301, 0.107, 0.029, 0.004], "prob_new": [0.4182564318180084, 0.9084106683731079, 0.9723316431045532, 0.9962301254272461], "prob_old": [0.8223134875297546, 0.5494247674942017, 0.6207829117774963, 0.6604117155075073], "prob_new_token": [0.615616500377655, 0.6587908267974854, 0.9071575999259949, 0.9943832159042358], "prob_old_token": [0.6156161427497864, 0.6587908267974854, 0.9071581363677979, 0.9943832159042358], "l1-model.layers.5.mlp.down_proj.weight": [45832.828125], "l2-model.layers.5.mlp.down_proj.weight": [7.137073993682861], "linf-model.layers.5.mlp.down_proj.weight": [0.0015023904852569103], "request": {"prompt": "{} was bestowed with the", "subject": "Bertha von Suttner", "target_new": {"str": "Nobel Prize in Literature"}, "old_answer": {"str": "Nobel Peace Prize"}, "seed": 42}}, {"loss_per_step": [7.671, 3.963, 2.726, 1.081, 0.132, 0.028, 0.018, 0.01, 0.006], "prob_new": [0.00046604787348769605, 0.01900462806224823, 0.06551004201173782, 0.339117169380188, 0.8765900135040283, 0.9726036787033081, 0.9821689128875732, 0.9899235367774963, 0.9939974546432495], "prob_old": [0.6538368463516235, 0.5124209523200989, 0.5411271452903748, 0.5485963821411133, 0.5642226338386536, 0.5092114806175232, 0.45895662903785706, 0.4431982934474945, 0.436045378446579], "prob_new_token": [0.00046604787348769605, 0.01900462806224823, 0.06551004201173782, 0.339117169380188, 0.8765900135040283, 0.9726036787033081, 0.9821689128875732, 0.9899235367774963, 0.9939974546432495], "prob_old_token": [0.28406721353530884, 0.0037024724297225475, 0.03090064227581024, 0.006247991230338812, 0.000391667039366439, 8.835536573315039e-05, 3.6006036680191755e-05, 1.42328090078081e-05, 6.261919679673156e-06], "l1-model.layers.5.mlp.down_proj.weight": [69879.734375], "l2-model.layers.5.mlp.down_proj.weight": [12.106013298034668], "linf-model.layers.5.mlp.down_proj.weight": [0.003911392763257027], "request": {"prompt": "{} has a citizenship of", "subject": "George Washington", "target_new": {"str": "Brazil"}, "old_answer": {"str": "the United States of America"}, "seed": 42}}, {"loss_per_step": [4.016, 3.135, 5.262, 1.71, 0.13, 0.036, 0.016, 0.01], "prob_new": [0.4206976592540741, 0.4143194258213043, 0.4468732476234436, 0.4754018187522888, 0.8824105858802795, 0.9643868207931519, 0.9838800430297852, 0.9902799129486084], "prob_old": [0.6538368463516235, 0.47842010855674744, 0.42340993881225586, 0.36893558502197266, 0.28297898173332214, 0.2850694954395294, 0.30465787649154663, 0.34006467461586], "prob_new_token": [0.00038631714414805174, 0.0022899864707142115, 3.0100760341156274e-05, 0.03574401140213013, 0.793851375579834, 0.9461484551429749, 0.9775094389915466, 0.9856782555580139], "prob_old_token": [0.28406721353530884, 7.956766057759523e-05, 7.5204611675872e-06, 0.00011289567191852257, 4.194118446321227e-06, 1.3765126141151995e-06, 7.301608206944366e-07, 5.625256562780123e-07], "l1-model.layers.5.mlp.down_proj.weight": [57992.1484375], "l2-model.layers.5.mlp.down_proj.weight": [10.430521011352539], "linf-model.layers.5.mlp.down_proj.weight": [0.0034672990441322327], "request": {"prompt": "{} has a citizenship of", "subject": "George Washington", "target_new": {"str": "Denmark"}, "old_answer": {"str": "the United States of America"}, "seed": 42}}, {"loss_per_step": [8.706, 2.465, 2.912, 1.116, 0.253, 0.036, 0.013, 0.01, 0.01, 0.011, 0.011, 0.01, 0.009], "prob_new": [0.00016564593533985317, 0.0849737673997879, 0.05438624694943428, 0.3276023268699646, 0.7767634391784668, 0.96503084897995, 0.9869547486305237, 0.9898620247840881, 0.9895567893981934, 0.988911509513855, 0.9889126420021057, 0.9896230697631836, 0.9907231330871582], "prob_old": [0.6538368463516235, 0.54535311460495, 0.5559126734733582, 0.5051906704902649, 0.4569140374660492, 0.4254329800605774, 0.4042602479457855, 0.392325222492218, 0.3883640766143799, 0.38813430070877075, 0.38989949226379395, 0.3930530250072479, 0.3974296748638153], "prob_new_token": [0.00016564593533985317, 0.0849737673997879, 0.05438624694943428, 0.3276023268699646, 0.7767634391784668, 0.96503084897995, 0.9869547486305237, 0.9898620247840881, 0.9895567893981934, 0.988911509513855, 0.9889126420021057, 0.9896230697631836, 0.9907231330871582], "prob_old_token": [0.28406721353530884, 0.0005235319258645177, 0.006239354610443115, 0.012721341103315353, 0.006846828851848841, 0.0012058999855071306, 0.0004771715321112424, 0.00044208072358742356, 0.0005797659396193922, 0.0008066062582656741, 0.0010684961453080177, 0.0013094634050503373, 0.0014798665652051568], "l1-model.layers.5.mlp.down_proj.weight": [84903.5859375], "l2-model.layers.5.mlp.down_proj.weight": [14.622796058654785], "linf-model.layers.5.mlp.down_proj.weight": [0.00545946229249239], "request": {"prompt": "{} has a citizenship of", "subject": "George Washington", "target_new": {"str": "Netherlands"}, "old_answer": {"str": "the United States of America"}, "seed": 42}}, {"loss_per_step": [6.243, 1.988, 0.38, 0.035, 0.015, 0.008], "prob_new": [0.0019431750988587737, 0.13700175285339355, 0.6836854815483093, 0.9653246402740479, 0.9852173924446106, 0.9922294020652771], "prob_old": [0.8033087253570557, 0.01734166033565998, 0.02253260277211666, 0.013079226948320866, 0.006716220639646053, 0.0029880530200898647], "prob_new_token": [0.0019431750988587737, 0.13700175285339355, 0.6836854815483093, 0.9653246402740479, 0.9852173924446106, 0.9922294020652771], "prob_old_token": [0.8033087253570557, 0.01734166033565998, 0.02253260277211666, 0.013079226948320866, 0.006716220639646053, 0.0029880530200898647], "l1-model.layers.5.mlp.down_proj.weight": [52598.890625], "l2-model.layers.5.mlp.down_proj.weight": [8.99752426147461], "linf-model.layers.5.mlp.down_proj.weight": [0.0024756230413913727], "request": {"prompt": "{} has a citizenship of", "subject": "Eduardo Frei Montalva", "target_new": {"str": "Argentina"}, "old_answer": {"str": "Chile"}, "seed": 42}}, {"loss_per_step": [9.596, 3.438, 0.229, 0.012, 0.004], "prob_new": [6.798121466999874e-05, 0.03211865574121475, 0.7951091527938843, 0.9880063533782959, 0.9963276386260986], "prob_old": [0.8033087253570557, 0.005610235035419464, 0.0002581333101261407, 6.832995950389886e-06, 1.4867548543406883e-06], "prob_new_token": [6.798121466999874e-05, 0.03211865574121475, 0.7951091527938843, 0.9880063533782959, 0.9963276386260986], "prob_old_token": [0.8033087253570557, 0.005610235035419464, 0.0002581333101261407, 6.832995950389886e-06, 1.4867548543406883e-06], "l1-model.layers.5.mlp.down_proj.weight": [51389.8515625], "l2-model.layers.5.mlp.down_proj.weight": [8.400827407836914], "linf-model.layers.5.mlp.down_proj.weight": [0.0020002652890980244], "request": {"prompt": "{} has a citizenship of", "subject": "Eduardo Frei Montalva", "target_new": {"str": "Egypt"}, "old_answer": {"str": "Chile"}, "seed": 42}}, {"loss_per_step": [12.105, 5.573, 3.135, 0.366, 0.474, 0.013, 0.009], "prob_new": [5.530110229301499e-06, 0.00380063708871603, 0.04351891204714775, 0.6937307715415955, 0.6226622462272644, 0.9871824979782104, 0.9906898736953735], "prob_old": [0.8033087253570557, 0.03572894260287285, 0.003063306212425232, 8.423850886174478e-06, 5.811013124912279e-06, 1.7516263994821202e-08, 2.3656102854374694e-08], "prob_new_token": [5.530110229301499e-06, 0.00380063708871603, 0.04351891204714775, 0.6937307715415955, 0.6226622462272644, 0.9871824979782104, 0.9906898736953735], "prob_old_token": [0.8033087253570557, 0.03572894260287285, 0.003063306212425232, 8.423850886174478e-06, 5.811013124912279e-06, 1.7516263994821202e-08, 2.3656102854374694e-08], "l1-model.layers.5.mlp.down_proj.weight": [57583.70703125], "l2-model.layers.5.mlp.down_proj.weight": [10.04370403289795], "linf-model.layers.5.mlp.down_proj.weight": [0.0029547177255153656], "request": {"prompt": "{} has a citizenship of", "subject": "Eduardo Frei Montalva", "target_new": {"str": "Ukraine"}, "old_answer": {"str": "Chile"}, "seed": 42}}, {"loss_per_step": [13.192, 5.325, 8.42, 3.347, 0.803, 0.007], "prob_new": [1.8659325178305153e-06, 0.004869234282523394, 0.00022033561253920197, 0.03518892452120781, 0.4482056498527527, 0.9929580092430115], "prob_old": [0.9169411659240723, 0.738279402256012, 0.6468073129653931, 0.7310205698013306, 0.7338388562202454, 0.7370702624320984], "prob_new_token": [1.8659325178305153e-06, 0.004869234282523394, 0.00022033561253920197, 0.03518892452120781, 0.4482056498527527, 0.9929580092430115], "prob_old_token": [0.7120962738990784, 0.0031474106945097446, 4.362065010354854e-05, 0.0019212975166738033, 0.0006019764114171267, 4.872243835052359e-07], "l1-model.layers.5.mlp.down_proj.weight": [54813.046875], "l2-model.layers.5.mlp.down_proj.weight": [9.19726276397705], "linf-model.layers.5.mlp.down_proj.weight": [0.0025109201669692993], "request": {"prompt": "{} borders with", "subject": "Bahrain", "target_new": {"str": "Ukraine"}, "old_answer": {"str": "Saudi Arabia"}, "seed": 42}}, {"loss_per_step": [6.385, 3.331, 2.207, 1.09, 0.38, 0.125, 0.033, 0.01, 0.005], "prob_new": [0.327322781085968, 0.4204474687576294, 0.598690390586853, 0.6681797504425049, 0.7720237970352173, 0.8958165049552917, 0.9682988524436951, 0.989956259727478, 0.9952453374862671], "prob_old": [0.9169411659240723, 0.7279624938964844, 0.6351423263549805, 0.6381887793540955, 0.6467003226280212, 0.5872313976287842, 0.5501819252967834, 0.5312538743019104, 0.5210847854614258], "prob_new_token": [2.312546257599024e-06, 0.00017326731176581234, 0.001673113089054823, 0.039359938353300095, 0.32151490449905396, 0.6895532011985779, 0.9066849946975708, 0.97175532579422, 0.9878191947937012], "prob_old_token": [0.7120962738990784, 0.009109566919505596, 0.0007368781371042132, 0.001010359963402152, 0.00046097292215563357, 8.148213237291202e-05, 1.3730820683122147e-05, 2.9167740649427287e-06, 8.968532370090543e-07], "l1-model.layers.5.mlp.down_proj.weight": [71152.0703125], "l2-model.layers.5.mlp.down_proj.weight": [12.256468772888184], "linf-model.layers.5.mlp.down_proj.weight": [0.003944524098187685], "request": {"prompt": "{} borders with", "subject": "Bahrain", "target_new": {"str": "Beijing"}, "old_answer": {"str": "Saudi Arabia"}, "seed": 42}}, {"loss_per_step": [6.077, 2.66, 2.074, 1.254, 0.221, 0.005], "prob_new": [0.39425405859947205, 0.4968605041503906, 0.5069138407707214, 0.540238082408905, 0.8215084075927734, 0.9954936504364014], "prob_old": [0.9169411659240723, 0.7273265719413757, 0.7419430613517761, 0.7506273984909058, 0.7479957342147827, 0.7455406188964844], "prob_new_token": [6.678090812783921e-06, 0.004952639807015657, 0.015830213204026222, 0.08158928900957108, 0.6437274813652039, 0.991369366645813], "prob_old_token": [0.7120962738990784, 0.0035250841174274683, 0.008358989842236042, 0.015255189500749111, 0.0009482370805926621, 4.733281457447447e-06], "l1-model.layers.5.mlp.down_proj.weight": [56653.32421875], "l2-model.layers.5.mlp.down_proj.weight": [9.481696128845215], "linf-model.layers.5.mlp.down_proj.weight": [0.0025106407701969147], "request": {"prompt": "{} borders with", "subject": "Bahrain", "target_new": {"str": "Thailand"}, "old_answer": {"str": "Saudi Arabia"}, "seed": 42}}, {"loss_per_step": [10.462, 2.055, 0.346, 0.128, 0.079, 0.059, 0.046, 0.036, 0.027, 0.021, 0.016, 0.013, 0.01, 0.008], "prob_new": [2.8606737032532692e-05, 0.12810303270816803, 0.7074926495552063, 0.8796914219856262, 0.924042820930481, 0.942656934261322, 0.9550192356109619, 0.9650870561599731, 0.9732497930526733, 0.9795135855674744, 0.9841318130493164, 0.9874891042709351, 0.9899390935897827, 0.9917523264884949], "prob_old": [0.8442697525024414, 0.5832552909851074, 0.5589548349380493, 0.5402214527130127, 0.5235801935195923, 0.512041449546814, 0.5054506659507751, 0.502010703086853, 0.5003604292869568, 0.49972003698349, 0.49964532256126404, 0.49986714124679565, 0.5002176761627197, 0.5005944967269897], "prob_new_token": [2.8606737032532692e-05, 0.12810303270816803, 0.7074926495552063, 0.8796914219856262, 0.924042820930481, 0.942656934261322, 0.9550192356109619, 0.9650870561599731, 0.9732497930526733, 0.9795135855674744, 0.9841318130493164, 0.9874891042709351, 0.9899390935897827, 0.9917523264884949], "prob_old_token": [0.412433922290802, 0.07260670512914658, 0.02280528098344803, 0.008936788886785507, 0.0050492356531322, 0.0034634629264473915, 0.002607037778943777, 0.00206460477784276, 0.0016886582598090172, 0.001412656158208847, 0.0012011013459414244, 0.001033471547998488, 0.0008975519449450076, 0.0007856612210161984], "l1-model.layers.5.mlp.down_proj.weight": [96881.4921875], "l2-model.layers.5.mlp.down_proj.weight": [15.662528991699219], "linf-model.layers.5.mlp.down_proj.weight": [0.006378944963216782], "request": {"prompt": "{} borders with", "subject": "Oman", "target_new": {"str": "India"}, "old_answer": {"str": "the United Arab Emirates"}, "seed": 42}}, {"loss_per_step": [13.771, 4.956, 2.242, 0.359, 0.112, 0.046, 0.026, 0.016, 0.012, 0.009], "prob_new": [1.0455619303684216e-06, 0.007040972821414471, 0.10619605332612991, 0.6985006332397461, 0.8937878012657166, 0.954676628112793, 0.9747020602226257, 0.9837885499000549, 0.9883670210838318, 0.9908027052879333], "prob_old": [0.8442697525024414, 0.5940683484077454, 0.5526894927024841, 0.5479859113693237, 0.5444557666778564, 0.5383859276771545, 0.5317811965942383, 0.525830864906311, 0.520333468914032, 0.5153111815452576], "prob_new_token": [1.0455619303684216e-06, 0.007040972821414471, 0.10619605332612991, 0.6985006332397461, 0.8937878012657166, 0.954676628112793, 0.9747020602226257, 0.9837885499000549, 0.9883670210838318, 0.9908027052879333], "prob_old_token": [0.412433922290802, 0.02822319231927395, 0.06705981492996216, 0.043736495077610016, 0.013187963515520096, 0.004723462276160717, 0.002090232912451029, 0.0010853353887796402, 0.0006549503304995596, 0.0004551029996946454], "l1-model.layers.5.mlp.down_proj.weight": [78152.125], "l2-model.layers.5.mlp.down_proj.weight": [13.20538330078125], "linf-model.layers.5.mlp.down_proj.weight": [0.00437795277684927], "request": {"prompt": "{} borders with", "subject": "Oman", "target_new": {"str": "Mexico"}, "old_answer": {"str": "the United Arab Emirates"}, "seed": 42}}, {"loss_per_step": [0.37, 0.779, 0.068, 0.009], "prob_new": [0.7990957498550415, 0.7558145523071289, 0.9401329159736633, 0.9911136031150818], "prob_old": [0.8442697525024414, 0.6498364210128784, 0.6655158996582031, 0.658033549785614], "prob_new_token": [0.23768770694732666, 0.0453009270131588, 0.7648531198501587, 0.9667884707450867], "prob_old_token": [0.412433922290802, 0.1653294861316681, 0.06272917985916138, 0.015206710435450077], "l1-model.layers.5.mlp.down_proj.weight": [35758.34375], "l2-model.layers.5.mlp.down_proj.weight": [6.395098686218262], "linf-model.layers.5.mlp.down_proj.weight": [0.0015024561434984207], "request": {"prompt": "{} borders with", "subject": "Oman", "target_new": {"str": "Saudi Arabia"}, "old_answer": {"str": "the United Arab Emirates"}, "seed": 42}}, {"loss_per_step": [1.808, 1.064, 0.243, 0.05, 0.011, 0.006], "prob_new": [0.38338664174079895, 0.5667591691017151, 0.8142693638801575, 0.954352080821991, 0.9895681142807007, 0.9943416118621826], "prob_old": [0.6396514177322388, 0.4184156656265259, 0.41237297654151917, 0.424730509519577, 0.4300638437271118, 0.42977291345596313], "prob_new_token": [0.08499715477228165, 0.23334090411663055, 0.4195730984210968, 0.7704635262489319, 0.9513286352157593, 0.9777541160583496], "prob_old_token": [0.7084969282150269, 0.18560418486595154, 0.022454578429460526, 0.0006449112552218139, 3.307378574390896e-05, 4.93383640787215e-06], "l1-model.layers.5.mlp.down_proj.weight": [59637.84375], "l2-model.layers.5.mlp.down_proj.weight": [9.731258392333984], "linf-model.layers.5.mlp.down_proj.weight": [0.0024848964530974627], "request": {"prompt": "{} has earned an educational degree from", "subject": "Bill Gates", "target_new": {"str": "the University of California, Los Angeles"}, "old_answer": {"str": "Harvard University"}, "seed": 42}}, {"loss_per_step": [2.934, 1.092, 0.533, 0.098, 0.04, 0.018, 0.01, 0.007], "prob_new": [0.36576762795448303, 0.5811721682548523, 0.6705288290977478, 0.9149570465087891, 0.9622160196304321, 0.9827337265014648, 0.9899687767028809, 0.993395984172821], "prob_old": [0.6396514177322388, 0.29669421911239624, 0.3260648548603058, 0.38133713603019714, 0.40913546085357666, 0.42462801933288574, 0.4359566271305084, 0.44519704580307007], "prob_new_token": [0.0849967896938324, 0.22594279050827026, 0.27821674942970276, 0.6989898681640625, 0.8788129091262817, 0.9524776339530945, 0.9737966060638428, 0.9828746914863586], "prob_old_token": [0.7084969282150269, 0.12965555489063263, 0.02095807157456875, 0.0029327203519642353, 0.0008491560583934188, 0.00044089381117373705, 0.0003084293275605887, 0.0002461575495544821], "l1-model.layers.5.mlp.down_proj.weight": [70398.21875], "l2-model.layers.5.mlp.down_proj.weight": [11.57010555267334], "linf-model.layers.5.mlp.down_proj.weight": [0.003479632781818509], "request": {"prompt": "{} has earned an educational degree from", "subject": "Bill Gates", "target_new": {"str": "the University of Bristol"}, "old_answer": {"str": "Harvard University"}, "seed": 42}}, {"loss_per_step": [2.593, 0.601, 0.237, 0.053, 0.019, 0.011, 0.008], "prob_new": [0.27795514464378357, 0.6708544492721558, 0.8179119825363159, 0.9508875608444214, 0.9812269806861877, 0.9889829158782959, 0.992096483707428], "prob_old": [0.6396514177322388, 0.35901275277137756, 0.36412206292152405, 0.3933047950267792, 0.40576064586639404, 0.4134485721588135, 0.41979750990867615], "prob_new_token": [0.0849967896938324, 0.16270512342453003, 0.49742361903190613, 0.845572829246521, 0.9382274150848389, 0.9632478356361389, 0.9738162159919739], "prob_old_token": [0.7084969282150269, 0.2173740714788437, 0.01334257423877716, 0.002400054596364498, 0.0007575471536256373, 0.00038707832572981715, 0.000255089980782941], "l1-model.layers.5.mlp.down_proj.weight": [71213.1796875], "l2-model.layers.5.mlp.down_proj.weight": [11.03486442565918], "linf-model.layers.5.mlp.down_proj.weight": [0.0030054617673158646], "request": {"prompt": "{} has earned an educational degree from", "subject": "Bill Gates", "target_new": {"str": "the University of Cambridge"}, "old_answer": {"str": "Harvard University"}, "seed": 42}}, {"loss_per_step": [1.218, 0.511, 0.173, 0.026, 0.013, 0.008], "prob_new": [0.5217112302780151, 0.6624167561531067, 0.8520073294639587, 0.9751573204994202, 0.987422525882721, 0.9922884702682495], "prob_old": [0.6436144113540649, 0.4192441999912262, 0.5934644341468811, 0.7246495485305786, 0.7354620695114136, 0.7409523725509644], "prob_new_token": [0.5371917486190796, 0.21540576219558716, 0.584760844707489, 0.9138467907905579, 0.9482256174087524, 0.9677613973617554], "prob_old_token": [0.53719162940979, 0.21540549397468567, 0.5847613215446472, 0.9138469696044922, 0.9482254981994629, 0.9677613973617554], "l1-model.layers.5.mlp.down_proj.weight": [60526.828125], "l2-model.layers.5.mlp.down_proj.weight": [9.669218063354492], "linf-model.layers.5.mlp.down_proj.weight": [0.0025045042857527733], "request": {"prompt": "{} has earned an educational degree from", "subject": "Frank Herbert", "target_new": {"str": "the University of California, Berkeley"}, "old_answer": {"str": "the University of Washington"}, "seed": 42}}, {"loss_per_step": [2.252, 1.168, 0.264, 0.101, 0.007], "prob_new": [0.5689975023269653, 0.4701634645462036, 0.7851839661598206, 0.9134292006492615, 0.9929311275482178], "prob_old": [0.6436144113540649, 0.45559608936309814, 0.5882429480552673, 0.6669857501983643, 0.7429997324943542], "prob_new_token": [0.53719162940979, 0.27215543389320374, 0.5411530137062073, 0.6996404528617859, 0.9822403788566589], "prob_old_token": [0.53719162940979, 0.27215543389320374, 0.5411530137062073, 0.6996404528617859, 0.9822403788566589], "l1-model.layers.5.mlp.down_proj.weight": [49594.82421875], "l2-model.layers.5.mlp.down_proj.weight": [8.226520538330078], "linf-model.layers.5.mlp.down_proj.weight": [0.002005700021982193], "request": {"prompt": "{} has earned an educational degree from", "subject": "Frank Herbert", "target_new": {"str": "the University of Texas"}, "old_answer": {"str": "the University of Washington"}, "seed": 42}}, {"loss_per_step": [2.435, 1.411, 0.482, 0.057, 0.007], "prob_new": [0.6940945386886597, 0.6194875240325928, 0.7400031089782715, 0.9471420049667358, 0.9932782053947449], "prob_old": [0.6436144113540649, 0.4436175227165222, 0.5784876346588135, 0.6716074347496033, 0.7401434183120728], "prob_new_token": [0.5371917486190796, 0.44008782505989075, 0.6524181962013245, 0.8334012031555176, 0.9739862084388733], "prob_old_token": [0.53719162940979, 0.4400879442691803, 0.6524172425270081, 0.8334010243415833, 0.9739862084388733], "l1-model.layers.5.mlp.down_proj.weight": [47818.74609375], "l2-model.layers.5.mlp.down_proj.weight": [8.089674949645996], "linf-model.layers.5.mlp.down_proj.weight": [0.0020057670772075653], "request": {"prompt": "{} has earned an educational degree from", "subject": "Frank Herbert", "target_new": {"str": "the University of Bucharest"}, "old_answer": {"str": "the University of Washington"}, "seed": 42}}, {"loss_per_step": [4.539, 2.575, 0.787, 0.127, 0.057, 0.033, 0.021, 0.014, 0.01], "prob_new": [0.48854687809944153, 0.47522127628326416, 0.601082444190979, 0.8876234292984009, 0.9457533359527588, 0.9681031703948975, 0.9796579480171204, 0.9864494800567627, 0.9904876947402954], "prob_old": [0.9215955138206482, 0.002495512133464217, 0.01490804273635149, 0.0006775966030545533, 5.4954671213636175e-05, 1.293789318879135e-05, 4.823977178602945e-06, 2.3092734409146942e-06, 1.332606530013436e-06], "prob_new_token": [0.00011676352005451918, 0.006135871633887291, 0.2087453454732895, 0.7785412073135376, 0.8936631679534912, 0.9378945231437683, 0.9607127904891968, 0.9740749597549438, 0.9819761514663696], "prob_old_token": [0.9215955138206482, 0.002495512133464217, 0.01490804273635149, 0.0006775966030545533, 5.4954671213636175e-05, 1.293789318879135e-05, 4.823977178602945e-06, 2.3092734409146942e-06, 1.332606530013436e-06], "l1-model.layers.5.mlp.down_proj.weight": [67893.265625], "l2-model.layers.5.mlp.down_proj.weight": [11.819320678710938], "linf-model.layers.5.mlp.down_proj.weight": [0.003932571038603783], "request": {"prompt": "The native language of {} is", "subject": "Robin van Persie", "target_new": {"str": "Arabic"}, "old_answer": {"str": "Dutch"}, "seed": 42}}, {"loss_per_step": [4.331, 2.268, 2.068, 0.13, 0.093, 0.041, 0.023, 0.014, 0.01], "prob_new": [0.6566707491874695, 0.4276466369628906, 0.6617422103881836, 0.8846167325973511, 0.9138389825820923, 0.960239052772522, 0.9775773286819458, 0.9857494235038757, 0.9903916120529175], "prob_old": [0.9215955138206482, 0.0002863928966689855, 5.7970755733549595e-06, 0.00012047505151713267, 4.8631405661581084e-05, 4.8139700083993375e-05, 3.541365367709659e-05, 2.365950058447197e-05, 1.6082392903626896e-05], "prob_new_token": [2.3428087843058165e-06, 0.003537180367857218, 0.0020546000450849533, 0.7468310594558716, 0.9172155261039734, 0.9286741614341736, 0.9467282891273499, 0.9634985327720642, 0.974801242351532], "prob_old_token": [0.9215955138206482, 0.0002863928966689855, 5.7970755733549595e-06, 0.00012047505151713267, 4.8631405661581084e-05, 4.8139700083993375e-05, 3.541365367709659e-05, 2.365950058447197e-05, 1.6082392903626896e-05], "l1-model.layers.5.mlp.down_proj.weight": [61448.48046875], "l2-model.layers.5.mlp.down_proj.weight": [11.045373916625977], "linf-model.layers.5.mlp.down_proj.weight": [0.003922387026250362], "request": {"prompt": "The native language of {} is", "subject": "Robin van Persie", "target_new": {"str": "Kurdish"}, "old_answer": {"str": "Dutch"}, "seed": 42}}, {"loss_per_step": [4.866, 2.443, 2.956, 1.128, 0.467, 0.127, 0.028, 0.008], "prob_new": [0.29197633266448975, 0.5024356842041016, 0.43456241488456726, 0.6693026423454285, 0.7455794215202332, 0.893939197063446, 0.9727336764335632, 0.9924246668815613], "prob_old": [0.9215955138206482, 0.01986982859671116, 4.566672089367785e-07, 4.090939546586014e-05, 5.414125553215854e-06, 5.011633220419753e-06, 2.3990464796952438e-06, 7.803485004842514e-07], "prob_new_token": [3.0160324968164787e-05, 0.0012763775885105133, 0.00045993636013008654, 0.03482837975025177, 0.24937410652637482, 0.6841727495193481, 0.9194071292877197, 0.9781665205955505], "prob_old_token": [0.9215955138206482, 0.01986982859671116, 4.566672089367785e-07, 4.090939546586014e-05, 5.414125553215854e-06, 5.011633220419753e-06, 2.3990464796952438e-06, 7.803485004842514e-07], "l1-model.layers.5.mlp.down_proj.weight": [64321.4609375], "l2-model.layers.5.mlp.down_proj.weight": [11.13486385345459], "linf-model.layers.5.mlp.down_proj.weight": [0.003485433291643858], "request": {"prompt": "The native language of {} is", "subject": "Robin van Persie", "target_new": {"str": "Uzbek"}, "old_answer": {"str": "Dutch"}, "seed": 42}}, {"loss_per_step": [3.176, 0.826, 0.09, 0.004], "prob_new": [0.4383750557899475, 0.6866359114646912, 0.9211472868919373, 0.9963235259056091], "prob_old": [0.9290962219238281, 0.00670870253816247, 0.0005787034169770777, 2.890358882723376e-05], "prob_new_token": [0.00023026124108582735, 0.08612951636314392, 0.7642050385475159, 0.9891396760940552], "prob_old_token": [0.9290962219238281, 0.00670870253816247, 0.0005787034169770777, 2.890358882723376e-05], "l1-model.layers.5.mlp.down_proj.weight": [42616.12890625], "l2-model.layers.5.mlp.down_proj.weight": [6.912038326263428], "linf-model.layers.5.mlp.down_proj.weight": [0.0015024738386273384], "request": {"prompt": "The native language of {} is", "subject": "Monica Bellucci", "target_new": {"str": "Latvian"}, "old_answer": {"str": "Italian"}, "seed": 42}}, {"loss_per_step": [6.123, 0.659, 0.185, 0.005], "prob_new": [0.4781739413738251, 0.6314274668693542, 0.8452332019805908, 0.9946051836013794], "prob_old": [0.9290962219238281, 0.004627260845154524, 2.583117202448193e-05, 9.095906250422559e-08], "prob_new_token": [5.023955509386724e-06, 0.2694295048713684, 0.6905396580696106, 0.9892578125], "prob_old_token": [0.9290962219238281, 0.004627260845154524, 2.583117202448193e-05, 9.095906250422559e-08], "l1-model.layers.5.mlp.down_proj.weight": [39459.80859375], "l2-model.layers.5.mlp.down_proj.weight": [6.709261417388916], "linf-model.layers.5.mlp.down_proj.weight": [0.0015024710446596146], "request": {"prompt": "The native language of {} is", "subject": "Monica Bellucci", "target_new": {"str": "Bengali"}, "old_answer": {"str": "Italian"}, "seed": 42}}, {"loss_per_step": [5.685, 2.431, 0.474, 0.06, 0.033, 0.021, 0.017, 0.013, 0.009], "prob_new": [0.49420082569122314, 0.503578245639801, 0.6911346316337585, 0.9430121183395386, 0.9681479930877686, 0.9791651964187622, 0.9837371110916138, 0.9874352812767029, 0.991348922252655], "prob_old": [0.9290962219238281, 0.002522740513086319, 0.0034340841230005026, 0.0003399174311198294, 0.0001482325023971498, 7.867390377214178e-05, 4.8522077122470364e-05, 3.481069506960921e-05, 2.524112096580211e-05], "prob_new_token": [1.1662390534183942e-05, 0.007738425396382809, 0.39138150215148926, 0.886178195476532, 0.9364127516746521, 0.9584077000617981, 0.967528223991394, 0.9749133586883545, 0.9827308654785156], "prob_old_token": [0.9290962219238281, 0.002522740513086319, 0.0034340841230005026, 0.0003399174311198294, 0.0001482325023971498, 7.867390377214178e-05, 4.8522077122470364e-05, 3.481069506960921e-05, 2.524112096580211e-05], "l1-model.layers.5.mlp.down_proj.weight": [64944.10546875], "l2-model.layers.5.mlp.down_proj.weight": [11.274455070495605], "linf-model.layers.5.mlp.down_proj.weight": [0.003994066268205643], "request": {"prompt": "The native language of {} is", "subject": "Monica Bellucci", "target_new": {"str": "Hebrew"}, "old_answer": {"str": "Italian"}, "seed": 42}}, {"loss_per_step": [8.288, 4.029, 3.033, 1.42, 0.233, 0.035, 0.015, 0.01, 0.009], "prob_new": [0.13263864815235138, 0.34201544523239136, 0.3440999984741211, 0.5913792252540588, 0.8156296014785767, 0.9660837054252625, 0.9846916198730469, 0.9896364212036133, 0.9908202886581421], "prob_old": [0.9271687269210815, 0.019539030268788338, 0.0008260455215349793, 0.0006313424091786146, 0.0005751459393650293, 0.0005562052829191089, 0.0005980053101666272, 0.0006609019474126399, 0.0007140462985262275], "prob_new_token": [1.7212462566362774e-08, 0.0001433494035154581, 0.0023869366850703955, 0.01860121637582779, 0.5569929480552673, 0.9610960483551025, 0.9857218265533447, 0.985595703125, 0.9823752641677856], "prob_old_token": [0.8750066161155701, 0.0004442795761860907, 0.0003123669303022325, 0.0001601795811438933, 7.175558130256832e-05, 2.0045893052156316e-06, 2.3525312542460597e-07, 7.140425140050866e-08, 3.2523761461789036e-08], "l1-model.layers.5.mlp.down_proj.weight": [73254.546875], "l2-model.layers.5.mlp.down_proj.weight": [12.359704971313477], "linf-model.layers.5.mlp.down_proj.weight": [0.003940541297197342], "request": {"prompt": "{} is named in honor of", "subject": "St. Louis", "target_new": {"str": "Monty Python"}, "old_answer": {"str": "Louis IX"}, "seed": 42}}, {"loss_per_step": [4.041, 1.592, 0.514, 0.028, 0.02, 0.008], "prob_new": [0.43936824798583984, 0.4979456067085266, 0.7084682583808899, 0.9729121327400208, 0.9802213907241821, 0.9920210242271423], "prob_old": [0.9271687269210815, 0.35814595222473145, 0.02715272270143032, 0.05000844597816467, 0.06481849402189255, 0.09536326676607132], "prob_new_token": [2.26958636631025e-05, 0.018975993618369102, 0.1931193619966507, 0.91566002368927, 0.9812365174293518, 0.990705668926239], "prob_old_token": [0.8750066161155701, 0.00042786519043147564, 6.922725333424751e-06, 1.4688591818412533e-06, 1.6461827101466042e-07, 5.320532636687858e-08], "l1-model.layers.5.mlp.down_proj.weight": [58178.8671875], "l2-model.layers.5.mlp.down_proj.weight": [9.537222862243652], "linf-model.layers.5.mlp.down_proj.weight": [0.0025042728520929813], "request": {"prompt": "{} is named in honor of", "subject": "St. Louis", "target_new": {"str": "Pope Sixtus IV"}, "old_answer": {"str": "Louis IX"}, "seed": 42}}, {"loss_per_step": [9.03, 3.064, 1.239, 0.736, 0.035, 0.01, 0.005], "prob_new": [0.2290521115064621, 0.2709009051322937, 0.5370561480522156, 0.6482818126678467, 0.9661321640014648, 0.9896777868270874, 0.9948645830154419], "prob_old": [0.9271687269210815, 0.0020894045010209084, 0.002162720775231719, 0.0016323908930644393, 0.00021884312445763499, 6.526243669213727e-05, 4.097318014828488e-05], "prob_new_token": [1.5089844964677468e-06, 0.0026502085383981466, 0.047593407332897186, 0.09476955235004425, 0.9037413001060486, 0.9670590758323669, 0.9838305115699768], "prob_old_token": [0.8750066161155701, 0.0031418867874890566, 0.003742382861673832, 0.003232566174119711, 0.000410599575843662, 9.508645598543808e-05, 2.7999229132547043e-05], "l1-model.layers.5.mlp.down_proj.weight": [66062.75], "l2-model.layers.5.mlp.down_proj.weight": [10.723812103271484], "linf-model.layers.5.mlp.down_proj.weight": [0.0029659178107976913], "request": {"prompt": "{} is named in honor of", "subject": "St. Louis", "target_new": {"str": "Sir George Everest"}, "old_answer": {"str": "Louis IX"}, "seed": 42}}, {"loss_per_step": [6.576, 4.534, 4.604, 2.874, 1.547, 0.592, 0.128, 0.047, 0.022, 0.013, 0.008], "prob_new": [0.0038403940852731466, 0.05123864859342575, 0.3102317750453949, 0.3305216133594513, 0.4050292372703552, 0.6730119585990906, 0.8884804248809814, 0.9554016590118408, 0.9781477451324463, 0.9872614145278931, 0.9920096397399902], "prob_old": [0.8951084017753601, 0.4579842686653137, 0.46208828687667847, 0.5321730971336365, 0.4881213903427124, 0.4833516776561737, 0.4603043794631958, 0.42789551615715027, 0.41004621982574463, 0.40187275409698486, 0.3976250886917114], "prob_new_token": [8.593811799073592e-05, 0.005415316205471754, 6.899864092702046e-05, 0.005857690237462521, 0.051632486283779144, 0.20740170776844025, 0.721248209476471, 0.8943204879760742, 0.9450135827064514, 0.9681712985038757, 0.9806166887283325], "prob_old_token": [0.7112005949020386, 0.0003187624388374388, 8.94141339813359e-06, 3.682096576085314e-05, 0.0004576877399813384, 5.8658824855228886e-05, 1.999747837544419e-05, 8.302542482852004e-06, 3.6953078961232677e-06, 1.7100958302762592e-06, 8.440605370196863e-07], "l1-model.layers.5.mlp.down_proj.weight": [73709.703125], "l2-model.layers.5.mlp.down_proj.weight": [13.06563949584961], "linf-model.layers.5.mlp.down_proj.weight": [0.004790018778294325], "request": {"prompt": "{} is named in honor of", "subject": "Columbia", "target_new": {"str": "Mentha"}, "old_answer": {"str": "Christopher Columbus"}, "seed": 42}}, {"loss_per_step": [5.228, 2.886, 1.877, 0.672, 0.066, 0.022, 0.009], "prob_new": [0.22864583134651184, 0.5156552791595459, 0.5625013113021851, 0.6620305776596069, 0.9364824891090393, 0.9781786799430847, 0.9908254146575928], "prob_old": [0.8951084017753601, 0.5741003155708313, 0.5884353518486023, 0.6138125061988831, 0.6007376909255981, 0.5871226191520691, 0.542352557182312], "prob_new_token": [0.10144669562578201, 0.6414873003959656, 0.7294967174530029, 0.856999933719635, 0.9527335166931152, 0.9615812301635742, 0.9760960936546326], "prob_old_token": [0.7112005949020386, 0.00031293017673306167, 7.873785216361284e-05, 5.6771646086417604e-06, 1.1355784579336614e-07, 2.4829192923903065e-08, 8.598417089444865e-09], "l1-model.layers.5.mlp.down_proj.weight": [67509.0546875], "l2-model.layers.5.mlp.down_proj.weight": [10.994890213012695], "linf-model.layers.5.mlp.down_proj.weight": [0.0030067870393395424], "request": {"prompt": "{} is named in honor of", "subject": "Columbia", "target_new": {"str": "the Kazakh people"}, "old_answer": {"str": "Christopher Columbus"}, "seed": 42}}, {"loss_per_step": [7.351, 5.396, 2.145, 0.995, 1.811, 0.013, 0.016, 0.008], "prob_new": [0.3051206171512604, 0.33054593205451965, 0.4080210328102112, 0.6803285479545593, 0.665439248085022, 0.987017810344696, 0.9843270182609558, 0.9923221468925476], "prob_old": [0.8951084017753601, 0.3431844711303711, 0.4326927363872528, 0.4527016580104828, 0.37753066420555115, 0.3786843419075012, 0.39048832654953003, 0.3954700827598572], "prob_new_token": [1.5171211771303206e-06, 7.496742182411253e-05, 0.00729805463925004, 0.05102962255477905, 0.0043994225561618805, 0.9759189486503601, 0.9889814257621765, 0.9913075566291809], "prob_old_token": [0.7112005949020386, 0.0002517199900466949, 0.0014805760001763701, 0.0009242103551514447, 0.00047330695088021457, 9.755222708918154e-05, 3.6653989809565246e-05, 1.8403576177661307e-05], "l1-model.layers.5.mlp.down_proj.weight": [62949.8359375], "l2-model.layers.5.mlp.down_proj.weight": [10.918460845947266], "linf-model.layers.5.mlp.down_proj.weight": [0.003481188789010048], "request": {"prompt": "{} is named in honor of", "subject": "Columbia", "target_new": {"str": "Friedrich Mohs"}, "old_answer": {"str": "Christopher Columbus"}, "seed": 42}}, {"loss_per_step": [5.278, 2.054, 0.332, 0.005], "prob_new": [0.4562312364578247, 0.21616220474243164, 0.7541916966438293, 0.9954757690429688], "prob_old": [0.9135269522666931, 0.7562227845191956, 0.7531885504722595, 0.7722553610801697], "prob_new_token": [2.8525771995191462e-05, 0.0421164408326149, 0.5225915908813477, 0.992153525352478], "prob_old_token": [0.6618219614028931, 0.03149307891726494, 0.00014068998279981315, 7.450562407029793e-07], "l1-model.layers.5.mlp.down_proj.weight": [42320.0], "l2-model.layers.5.mlp.down_proj.weight": [6.930954456329346], "linf-model.layers.5.mlp.down_proj.weight": [0.0015024784952402115], "request": {"prompt": "{} is affiliated with the religious tradition of", "subject": "Dalai Lama", "target_new": {"str": "Christianity"}, "old_answer": {"str": "Tibetan Buddhism"}, "seed": 42}}, {"loss_per_step": [2.407, 1.144, 0.674, 0.257, 0.045, 0.009], "prob_new": [0.5771450400352478, 0.7237424254417419, 0.7248870730400085, 0.8539964556694031, 0.9596297144889832, 0.9913563132286072], "prob_old": [0.9135269522666931, 0.7269076108932495, 0.7540473937988281, 0.7018972039222717, 0.7000221014022827, 0.6919108629226685], "prob_new_token": [0.0009396239765919745, 0.004992019385099411, 0.05959398299455643, 0.27994322776794434, 0.8026137351989746, 0.9593324661254883], "prob_old_token": [0.6618219614028931, 0.07891707122325897, 0.0014977863756939769, 0.0006652747397311032, 4.0796723624225706e-05, 2.654950094438391e-06], "l1-model.layers.5.mlp.down_proj.weight": [57077.8984375], "l2-model.layers.5.mlp.down_proj.weight": [9.394606590270996], "linf-model.layers.5.mlp.down_proj.weight": [0.002508319914340973], "request": {"prompt": "{} is affiliated with the religious tradition of", "subject": "Dalai Lama", "target_new": {"str": "Zoroastrianism"}, "old_answer": {"str": "Tibetan Buddhism"}, "seed": 42}}, {"loss_per_step": [9.177, 2.114, 0.227, 0.093, 0.036, 0.016, 0.008], "prob_new": [0.00010335681145079434, 0.12080952525138855, 0.7966057658195496, 0.9112221598625183, 0.9647892117500305, 0.9845311045646667, 0.9918200969696045], "prob_old": [0.9135269522666931, 0.7205348610877991, 0.6934155225753784, 0.654651939868927, 0.634206235408783, 0.6248499155044556, 0.6180835962295532], "prob_new_token": [0.00010335681145079434, 0.12080952525138855, 0.7966057658195496, 0.9112221598625183, 0.9647892117500305, 0.9845311045646667, 0.9918200969696045], "prob_old_token": [0.6618219614028931, 0.018111512064933777, 3.275846393080428e-05, 7.592141628265381e-06, 1.8047007870336529e-06, 5.058254828327335e-07, 2.009156645499388e-07], "l1-model.layers.5.mlp.down_proj.weight": [62444.46875], "l2-model.layers.5.mlp.down_proj.weight": [10.565845489501953], "linf-model.layers.5.mlp.down_proj.weight": [0.002984141930937767], "request": {"prompt": "{} is affiliated with the religious tradition of", "subject": "Dalai Lama", "target_new": {"str": "Islam"}, "old_answer": {"str": "Tibetan Buddhism"}, "seed": 42}}, {"loss_per_step": [10.037, 3.038, 0.04, 0.019, 0.008], "prob_new": [4.3743333662860096e-05, 0.047943826764822006, 0.9608306884765625, 0.9808226823806763, 0.9920023083686829], "prob_old": [0.8717825412750244, 0.6496278047561646, 0.6377658247947693, 0.636447548866272, 0.6464526653289795], "prob_new_token": [4.3743333662860096e-05, 0.047943826764822006, 0.9608306884765625, 0.9808226823806763, 0.9920023083686829], "prob_old_token": [0.6194280385971069, 0.013500071130692959, 0.0011307523818686604, 0.0004952849121764302, 0.00014803584781475365], "l1-model.layers.5.mlp.down_proj.weight": [49535.703125], "l2-model.layers.5.mlp.down_proj.weight": [8.274377822875977], "linf-model.layers.5.mlp.down_proj.weight": [0.0020049456506967545], "request": {"prompt": "{} is affiliated with the religious tradition of", "subject": "Leonard Nimoy", "target_new": {"str": "Islam"}, "old_answer": {"str": "Judaism"}, "seed": 42}}, {"loss_per_step": [2.441, 0.514, 0.001], "prob_new": [0.5318131446838379, 0.7093061208724976, 0.9994553923606873], "prob_old": [0.8717825412750244, 0.3176054358482361, 0.63689786195755], "prob_new_token": [0.6194280385971069, 0.2404719442129135, 0.9996388554573059], "prob_old_token": [0.6194280385971069, 0.2404719442129135, 0.9996388554573059], "l1-model.layers.5.mlp.down_proj.weight": [33816.63671875], "l2-model.layers.5.mlp.down_proj.weight": [5.295888423919678], "linf-model.layers.5.mlp.down_proj.weight": [0.0010006800293922424], "request": {"prompt": "{} is affiliated with the religious tradition of", "subject": "Leonard Nimoy", "target_new": {"str": "Jainism"}, "old_answer": {"str": "Judaism"}, "seed": 42}}, {"loss_per_step": [2.763, 0.826, 1.705, 0.144, 0.098, 0.052, 0.028, 0.018, 0.013, 0.009], "prob_new": [0.4531806409358978, 0.6094047427177429, 0.6682528257369995, 0.882860541343689, 0.9147331714630127, 0.9519926905632019, 0.9730727672576904, 0.9825052618980408, 0.9875940680503845, 0.9910018444061279], "prob_old": [0.8717825412750244, 0.5070101022720337, 0.35723480582237244, 0.3388690948486328, 0.33684641122817993, 0.33476459980010986, 0.33381855487823486, 0.3335671126842499, 0.3336125612258911, 0.3337428569793701], "prob_new_token": [0.0006639091880060732, 0.11767823249101639, 0.006007316987961531, 0.6508832573890686, 0.7473016977310181, 0.8582859039306641, 0.9211567640304565, 0.949062168598175, 0.9639517664909363, 0.9738913178443909], "prob_old_token": [0.6194280385971069, 0.060008108615875244, 4.286119292373769e-05, 0.016316059976816177, 0.009846494533121586, 0.004161158576607704, 0.0017071841284632683, 0.0009024963364936411, 0.0005802737432532012, 0.0003993247519247234], "l1-model.layers.5.mlp.down_proj.weight": [64360.078125], "l2-model.layers.5.mlp.down_proj.weight": [11.57469367980957], "linf-model.layers.5.mlp.down_proj.weight": [0.004429154563695192], "request": {"prompt": "{} is affiliated with the religious tradition of", "subject": "Leonard Nimoy", "target_new": {"str": "Hinduism"}, "old_answer": {"str": "Judaism"}, "seed": 42}}, {"loss_per_step": [16.792, 8.008, 4.737, 0.24, 0.022, 0.011, 0.009], "prob_new": [5.095086308415375e-08, 0.00033287680707871914, 0.008768233470618725, 0.7863613963127136, 0.9781590700149536, 0.9885789752006531, 0.990742027759552], "prob_old": [0.9610093832015991, 0.5854246616363525, 0.35182780027389526, 0.46672889590263367, 0.4763708710670471, 0.477748304605484, 0.477150559425354], "prob_new_token": [5.095086308415375e-08, 0.00033287680707871914, 0.008768233470618725, 0.7863613963127136, 0.9781590700149536, 0.9885789752006531, 0.990742027759552], "prob_old_token": [0.9285872578620911, 0.17937979102134705, 0.0019820034503936768, 0.0003984927316196263, 1.6742014850024134e-05, 3.7096328924235422e-06, 1.828532276704209e-06], "l1-model.layers.5.mlp.down_proj.weight": [62052.6640625], "l2-model.layers.5.mlp.down_proj.weight": [10.548864364624023], "linf-model.layers.5.mlp.down_proj.weight": [0.0029341839253902435], "request": {"prompt": "{} has its headquarters in", "subject": "Hyundai Motor Company", "target_new": {"str": "Jerusalem"}, "old_answer": {"str": "Seoul"}, "seed": 42}}, {"loss_per_step": [7.336, 2.732, 1.595, 0.155, 0.024, 0.006], "prob_new": [0.1783924549818039, 0.5009069442749023, 0.5198951959609985, 0.8663889169692993, 0.9767643213272095, 0.993737518787384], "prob_old": [0.9610093832015991, 0.4965619444847107, 0.29362761974334717, 0.12902064621448517, 0.038012582808732986, 0.017987754195928574], "prob_new_token": [1.1907964108104352e-06, 0.004246697295457125, 0.041246745735406876, 0.732828676700592, 0.9535880088806152, 0.9875855445861816], "prob_old_token": [0.9285872578620911, 0.020736129954457283, 0.001128889387473464, 5.55667829758022e-05, 3.37826827490062e-06, 4.4404185928215156e-07], "l1-model.layers.5.mlp.down_proj.weight": [55685.015625], "l2-model.layers.5.mlp.down_proj.weight": [9.410797119140625], "linf-model.layers.5.mlp.down_proj.weight": [0.0024997107684612274], "request": {"prompt": "{} has its headquarters in", "subject": "Hyundai Motor Company", "target_new": {"str": "Crewe"}, "old_answer": {"str": "Seoul"}, "seed": 42}}, {"loss_per_step": [16.699, 6.508, 3.006, 0.344, 0.052, 0.022, 0.014, 0.01], "prob_new": [5.595259722213086e-08, 0.0014919242821633816, 0.04950936138629913, 0.7086716294288635, 0.9488972425460815, 0.9777701497077942, 0.9865458011627197, 0.990502119064331], "prob_old": [0.9610093832015991, 0.5170996189117432, 0.4180864989757538, 0.47801971435546875, 0.4563893675804138, 0.4114113450050354, 0.36488527059555054, 0.3302275836467743], "prob_new_token": [5.595259722213086e-08, 0.0014919242821633816, 0.04950936138629913, 0.7086716294288635, 0.9488972425460815, 0.9777701497077942, 0.9865458011627197, 0.990502119064331], "prob_old_token": [0.9285872578620911, 0.05606408417224884, 0.0017813423182815313, 0.0001752376847434789, 5.204847184359096e-06, 8.866945222507638e-07, 3.519668609897053e-07, 2.0626816876756493e-07], "l1-model.layers.5.mlp.down_proj.weight": [68944.078125], "l2-model.layers.5.mlp.down_proj.weight": [11.627738952636719], "linf-model.layers.5.mlp.down_proj.weight": [0.0034285038709640503], "request": {"prompt": "{} has its headquarters in", "subject": "Hyundai Motor Company", "target_new": {"str": "Edinburgh"}, "old_answer": {"str": "Seoul"}, "seed": 42}}, {"loss_per_step": [10.796, 6.247, 2.401, 0.228, 0.084, 0.032, 0.015, 0.008], "prob_new": [2.048414717137348e-05, 0.0019370072986930609, 0.09060921519994736, 0.7964832782745361, 0.9191843867301941, 0.9680624008178711, 0.9852837324142456, 0.9918054342269897], "prob_old": [0.8966929316520691, 0.20474781095981598, 0.4868764579296112, 0.49922680854797363, 0.49796539545059204, 0.4972476661205292, 0.49723494052886963, 0.49742135405540466], "prob_new_token": [2.048414717137348e-05, 0.0019370072986930609, 0.09060921519994736, 0.7964832782745361, 0.9191843867301941, 0.9680624008178711, 0.9852837324142456, 0.9918054342269897], "prob_old_token": [0.7980557680130005, 0.00812472589313984, 0.03107958659529686, 0.021824458613991737, 0.008083481341600418, 0.00271891662850976, 0.0010337166022509336, 0.0005012063775211573], "l1-model.layers.5.mlp.down_proj.weight": [64832.6640625], "l2-model.layers.5.mlp.down_proj.weight": [11.177581787109375], "linf-model.layers.5.mlp.down_proj.weight": [0.003465762361884117], "request": {"prompt": "{} has its headquarters in", "subject": "Vimeo", "target_new": {"str": "Stockholm"}, "old_answer": {"str": "New York"}, "seed": 42}}, {"loss_per_step": [3.999, 1.517, 0.103, 0.02, 0.007], "prob_new": [0.35732850432395935, 0.5142775774002075, 0.9081701040267944, 0.9805890917778015, 0.993340253829956], "prob_old": [0.8966929316520691, 0.368092805147171, 0.44829270243644714, 0.4373270571231842, 0.40676456689834595], "prob_new_token": [2.1942649254924618e-05, 0.018295587971806526, 0.7694332599639893, 0.9467098116874695, 0.985264003276825], "prob_old_token": [0.7980557680130005, 0.026456784456968307, 0.0027907206676900387, 0.0011917536612600088, 0.00017494329949840903], "l1-model.layers.5.mlp.down_proj.weight": [46413.78515625], "l2-model.layers.5.mlp.down_proj.weight": [8.041690826416016], "linf-model.layers.5.mlp.down_proj.weight": [0.0020047426223754883], "request": {"prompt": "{} has its headquarters in", "subject": "Vimeo", "target_new": {"str": "Philadelphia, Pennsylvania"}, "old_answer": {"str": "New York"}, "seed": 42}}, {"loss_per_step": [3.743, 2.326, 0.434, 0.193, 0.1, 0.051, 0.029, 0.019, 0.012, 0.008], "prob_new": [0.2795184254646301, 0.3391171395778656, 0.662269115447998, 0.8264426589012146, 0.9052584171295166, 0.9506735801696777, 0.9712032079696655, 0.981508731842041, 0.9878935217857361, 0.9921172857284546], "prob_old": [0.8966929316520691, 0.2398000955581665, 0.4126506447792053, 0.4228579103946686, 0.4317145347595215, 0.4425293207168579, 0.45141494274139404, 0.456735759973526, 0.45854148268699646, 0.45818227529525757], "prob_new_token": [8.744558726903051e-05, 0.003682599402964115, 0.638935387134552, 0.7641690969467163, 0.8685476779937744, 0.9387171864509583, 0.9665132164955139, 0.9802027940750122, 0.988309383392334, 0.9930854439735413], "prob_old_token": [0.7980557680130005, 0.004128294065594673, 0.00218783225864172, 0.0008489358006045222, 0.0004744674952235073, 0.0003334716020617634, 0.00021626937086693943, 0.00013227760791778564, 7.603593985550106e-05, 4.1395807784283534e-05], "l1-model.layers.5.mlp.down_proj.weight": [74902.6640625], "l2-model.layers.5.mlp.down_proj.weight": [12.742321968078613], "linf-model.layers.5.mlp.down_proj.weight": [0.004483865108340979], "request": {"prompt": "{} has its headquarters in", "subject": "Vimeo", "target_new": {"str": "Amsterdam, Netherlands"}, "old_answer": {"str": "New York"}, "seed": 42}}, {"loss_per_step": [3.524, 1.876, 1.286, 0.711, 0.007], "prob_new": [0.47731471061706543, 0.6584233045578003, 0.7294504046440125, 0.7758627533912659, 0.9929300546646118], "prob_old": [0.7825582027435303, 0.28078576922416687, 0.31036803126335144, 0.37078362703323364, 0.32398223876953125], "prob_new_token": [6.658617479615714e-08, 2.9616714527946897e-05, 0.0003613348526414484, 0.014043294824659824, 0.9695785045623779], "prob_old_token": [0.7788311839103699, 9.238204256689642e-06, 3.9037058741087094e-05, 2.5598601496312767e-06, 2.457472447758846e-08], "l1-model.layers.5.mlp.down_proj.weight": [51238.58984375], "l2-model.layers.5.mlp.down_proj.weight": [8.285030364990234], "linf-model.layers.5.mlp.down_proj.weight": [0.0020058192312717438], "request": {"prompt": "{} entered this world in the location of", "subject": "Rachel Maddow", "target_new": {"str": "Bourg-la-Reine"}, "old_answer": {"str": "Castro Valley, California"}, "seed": 42}}, {"loss_per_step": [2.998, 3.471, 1.315, 0.085, 0.036, 0.018, 0.007], "prob_new": [0.6149026155471802, 0.2710954546928406, 0.6638023853302002, 0.9192115664482117, 0.9647262096405029, 0.9818724989891052, 0.9931468367576599], "prob_old": [0.7825582027435303, 0.23107744753360748, 0.430535227060318, 0.4612329304218292, 0.49253058433532715, 0.5219380855560303, 0.5494810938835144], "prob_new_token": [1.1470999197626952e-05, 0.00012011516810161993, 0.007641467731446028, 0.8810049891471863, 0.9916862845420837, 0.9986271858215332, 0.9988036155700684], "prob_old_token": [0.7788311839103699, 2.9518541850848123e-06, 3.686885747811175e-07, 4.565591904537314e-09, 3.1299549063668053e-10, 7.465256190997138e-11, 8.039799381798218e-11], "l1-model.layers.5.mlp.down_proj.weight": [57159.4296875], "l2-model.layers.5.mlp.down_proj.weight": [10.011482238769531], "linf-model.layers.5.mlp.down_proj.weight": [0.00299665704369545], "request": {"prompt": "{} entered this world in the location of", "subject": "Rachel Maddow", "target_new": {"str": "Queens, New York"}, "old_answer": {"str": "Castro Valley, California"}, "seed": 42}}, {"loss_per_step": [3.428, 2.313, 1.46, 0.489, 0.087, 0.031, 0.017, 0.011, 0.008], "prob_new": [0.4513727128505707, 0.5039748549461365, 0.6372805833816528, 0.7809460759162903, 0.9220901727676392, 0.9697861075401306, 0.9836503863334656, 0.9894351959228516, 0.9920875430107117], "prob_old": [0.7825582027435303, 0.22452543675899506, 0.33784914016723633, 0.35624027252197266, 0.38372480869293213, 0.402579665184021, 0.4226953089237213, 0.4343908429145813, 0.43594175577163696], "prob_new_token": [3.683622708194889e-06, 0.00014767966058570892, 0.0019889611285179853, 0.10693774372339249, 0.7430421710014343, 0.9424049854278564, 0.9747253656387329, 0.9845873713493347, 0.9889467358589172], "prob_old_token": [0.7788311839103699, 2.6397490273666335e-06, 2.920960696428665e-06, 1.0151268270419678e-06, 2.6967043709191785e-07, 6.43743973682831e-08, 2.598367387918188e-08, 1.3477503557624004e-08, 8.201600287804922e-09], "l1-model.layers.5.mlp.down_proj.weight": [73758.515625], "l2-model.layers.5.mlp.down_proj.weight": [12.360116004943848], "linf-model.layers.5.mlp.down_proj.weight": [0.003952250815927982], "request": {"prompt": "{} entered this world in the location of", "subject": "Rachel Maddow", "target_new": {"str": "Grand Rapids, Minnesota"}, "old_answer": {"str": "Castro Valley, California"}, "seed": 42}}, {"loss_per_step": [4.425, 0.909, 0.06, 0.023, 0.005], "prob_new": [0.3284355401992798, 0.5875060558319092, 0.942279040813446, 0.9770132899284363, 0.9947863817214966], "prob_old": [0.7979272603988647, 0.6816624402999878, 0.516892671585083, 0.5045925378799438, 0.5039791464805603], "prob_new_token": [7.54646953282645e-06, 0.0944172814488411, 0.9557603597640991, 0.9804857969284058, 0.9963458776473999], "prob_old_token": [0.6284904479980469, 0.07353862375020981, 6.79420045344159e-05, 1.5152465493883938e-05, 1.5213679489534115e-06], "l1-model.layers.5.mlp.down_proj.weight": [46337.125], "l2-model.layers.5.mlp.down_proj.weight": [8.071825981140137], "linf-model.layers.5.mlp.down_proj.weight": [0.002005232498049736], "request": {"prompt": "{} entered this world in the location of", "subject": "Albrecht D\u00fcrer", "target_new": {"str": "Florence, Italy"}, "old_answer": {"str": "Nuremberg"}, "seed": 42}}, {"loss_per_step": [3.647, 2.121, 1.331, 0.91, 0.595, 0.438, 0.211, 0.022, 0.003], "prob_new": [0.2978762984275818, 0.515618622303009, 0.7088332176208496, 0.784055769443512, 0.8049885034561157, 0.8179850578308105, 0.8671661615371704, 0.9790732264518738, 0.9974536895751953], "prob_old": [0.7979272603988647, 0.47866010665893555, 0.4949415326118469, 0.4419851005077362, 0.48444780707359314, 0.4932538866996765, 0.4968489706516266, 0.4951891601085663, 0.4946199953556061], "prob_new_token": [1.9384273400646634e-05, 0.0003210980212315917, 0.0022726154420524836, 0.011597912758588791, 0.052552420645952225, 0.114772267639637, 0.3547165095806122, 0.9083641767501831, 0.9959761500358582], "prob_old_token": [0.6284904479980469, 0.004319264553487301, 0.007667692843824625, 0.006670310162007809, 0.006581724155694246, 0.010140296071767807, 0.01175406388938427, 0.0011133113875985146, 3.117593587376177e-05], "l1-model.layers.5.mlp.down_proj.weight": [69321.5], "l2-model.layers.5.mlp.down_proj.weight": [11.966877937316895], "linf-model.layers.5.mlp.down_proj.weight": [0.004006246104836464], "request": {"prompt": "{} entered this world in the location of", "subject": "Albrecht D\u00fcrer", "target_new": {"str": "Aberdeen, Washington"}, "old_answer": {"str": "Nuremberg"}, "seed": 42}}, {"loss_per_step": [7.303, 5.644, 2.8, 0.268, 0.002], "prob_new": [0.005293817725032568, 0.020134180784225464, 0.3535378873348236, 0.7910628318786621, 0.9984863996505737], "prob_old": [0.7979272603988647, 0.5895509719848633, 0.6383348703384399, 0.5620633363723755, 0.5594627857208252], "prob_new_token": [4.2988340283045545e-05, 0.0003133135905954987, 0.005273465532809496, 0.5891290307044983, 0.9984394907951355], "prob_old_token": [0.6284904479980469, 0.0017996265087276697, 0.004848438315093517, 0.0011372140143066645, 2.457582922943402e-06], "l1-model.layers.5.mlp.down_proj.weight": [48040.828125], "l2-model.layers.5.mlp.down_proj.weight": [8.078248977661133], "linf-model.layers.5.mlp.down_proj.weight": [0.0020058415830135345], "request": {"prompt": "{} entered this world in the location of", "subject": "Albrecht D\u00fcrer", "target_new": {"str": "Reus"}, "old_answer": {"str": "Nuremberg"}, "seed": 42}}, {"loss_per_step": [11.799, 5.136, 0.764, 0.012, 0.005], "prob_new": [7.5101984293723945e-06, 0.005879055242985487, 0.4658668339252472, 0.9882270693778992, 0.9952645897865295], "prob_old": [0.8133355975151062, 0.40903136134147644, 0.45671018958091736, 0.4925925135612488, 0.4909055531024933], "prob_new_token": [7.5101984293723945e-06, 0.005879055242985487, 0.4658668339252472, 0.9882270693778992, 0.9952645897865295], "prob_old_token": [0.7344122529029846, 0.0013335697585716844, 0.0014640106819570065, 1.5410558262374252e-05, 3.5069817840849282e-06], "l1-model.layers.5.mlp.down_proj.weight": [50130.05078125], "l2-model.layers.5.mlp.down_proj.weight": [8.312718391418457], "linf-model.layers.5.mlp.down_proj.weight": [0.0019905101507902145], "request": {"prompt": "{} passed away in the location of", "subject": "Wilhelm R\u00f6ntgen", "target_new": {"str": "Paris"}, "old_answer": {"str": "Munich"}, "seed": 42}}, {"loss_per_step": [4.471, 2.869, 2.521, 2.04, 1.576, 1.208, 1.435, 0.427, 0.098, 0.028, 0.013, 0.007], "prob_new": [0.2872834801673889, 0.48100027441978455, 0.4910925626754761, 0.5007783770561218, 0.5323994159698486, 0.553213357925415, 0.6957911252975464, 0.783506453037262, 0.9146827459335327, 0.9730408787727356, 0.9872572422027588, 0.9931737184524536], "prob_old": [0.8133355975151062, 0.18893158435821533, 0.0912313163280487, 0.03991073742508888, 0.04634564369916916, 0.028808433562517166, 0.07004120945930481, 0.05936044454574585, 0.11526701599359512, 0.15796317160129547, 0.1847762018442154, 0.1826390027999878], "prob_new_token": [0.000622739375103265, 0.0031955454032868147, 0.007087798789143562, 0.008566897362470627, 0.011796506121754646, 0.04282020777463913, 0.004121249075978994, 0.19241145253181458, 0.7198167443275452, 0.8993796706199646, 0.952824056148529, 0.9747632741928101], "prob_old_token": [0.7344122529029846, 0.00033317311317659914, 0.00012071978562744334, 0.00012920956942252815, 0.0002646477078087628, 3.246802953071892e-05, 6.248762656468898e-05, 0.00010253341315547004, 1.9245459043304436e-05, 4.58227941635414e-06, 1.5624118532286957e-06, 6.594358978873061e-07], "l1-model.layers.5.mlp.down_proj.weight": [85034.6953125], "l2-model.layers.5.mlp.down_proj.weight": [14.356243133544922], "linf-model.layers.5.mlp.down_proj.weight": [0.005375389009714127], "request": {"prompt": "{} passed away in the location of", "subject": "Wilhelm R\u00f6ntgen", "target_new": {"str": "Ephesus"}, "old_answer": {"str": "Munich"}, "seed": 42}}, {"loss_per_step": [2.974, 2.154, 0.199, 0.026, 0.011, 0.007], "prob_new": [0.6511784791946411, 0.4562367796897888, 0.8412792682647705, 0.9742946624755859, 0.9890035390853882, 0.9933189153671265], "prob_old": [0.8133355975151062, 0.21207594871520996, 0.15614429116249084, 0.19817721843719482, 0.21439149975776672, 0.23806825280189514], "prob_new_token": [0.00013980829680804163, 0.004087116569280624, 0.5873888731002808, 0.9610013961791992, 0.9844472408294678, 0.988671600818634], "prob_old_token": [0.7344122529029846, 3.509148518787697e-05, 6.735215833941766e-07, 6.69562751909325e-08, 3.3950950495409415e-08, 3.049618868544712e-08], "l1-model.layers.5.mlp.down_proj.weight": [58772.484375], "l2-model.layers.5.mlp.down_proj.weight": [9.600960731506348], "linf-model.layers.5.mlp.down_proj.weight": [0.0025031138211488724], "request": {"prompt": "{} passed away in the location of", "subject": "Wilhelm R\u00f6ntgen", "target_new": {"str": "Montreux"}, "old_answer": {"str": "Munich"}, "seed": 42}}, {"loss_per_step": [3.754, 2.099, 1.364, 0.495, 0.061, 0.025, 0.012, 0.007], "prob_new": [0.46865397691726685, 0.4978714883327484, 0.6728566288948059, 0.7967495322227478, 0.9430977702140808, 0.9756951332092285, 0.9881576299667358, 0.9934895634651184], "prob_old": [0.6166081428527832, 0.13289909064769745, 0.3120484948158264, 0.28385499119758606, 0.358875036239624, 0.408720463514328, 0.4011777341365814, 0.38547560572624207], "prob_new_token": [3.655817636172287e-06, 0.000952939793933183, 0.0007176602375693619, 0.05250303074717522, 0.8691417574882507, 0.9824642539024353, 0.9944988489151001, 0.996802031993866], "prob_old_token": [0.7293808460235596, 0.00021414825459942222, 0.00029061539680697024, 0.0004086457483936101, 0.00018568438827060163, 4.629196337191388e-05, 2.3046741262078285e-05, 1.6754713215050288e-05], "l1-model.layers.5.mlp.down_proj.weight": [67005.375], "l2-model.layers.5.mlp.down_proj.weight": [11.39721393585205], "linf-model.layers.5.mlp.down_proj.weight": [0.003465813584625721], "request": {"prompt": "{} passed away in the location of", "subject": "Sandro Botticelli", "target_new": {"str": "Berkeley, Gloucestershire"}, "old_answer": {"str": "Florence, Italy"}, "seed": 42}}, {"loss_per_step": [3.176, 1.928, 1.051, 0.094, 0.001], "prob_new": [0.5475641489028931, 0.6382128596305847, 0.7597323060035706, 0.9225441813468933, 0.9986611604690552], "prob_old": [0.6166081428527832, 0.3545500636100769, 0.5243579149246216, 0.5639011859893799, 0.6189867258071899], "prob_new_token": [8.43507734771265e-07, 0.00018285043188370764, 0.0064868321642279625, 0.646871030330658, 0.999998927116394], "prob_old_token": [0.7293808460235596, 0.004525529220700264, 0.013452236540615559, 0.0027714679017663, 2.428955647726383e-10], "l1-model.layers.5.mlp.down_proj.weight": [47827.55078125], "l2-model.layers.5.mlp.down_proj.weight": [8.01129150390625], "linf-model.layers.5.mlp.down_proj.weight": [0.0020057749934494495], "request": {"prompt": "{} passed away in the location of", "subject": "Sandro Botticelli", "target_new": {"str": "Johannesburg, South Africa"}, "old_answer": {"str": "Florence, Italy"}, "seed": 42}}, {"loss_per_step": [7.018, 4.164, 1.235, 0.468, 0.167, 0.024, 0.006], "prob_new": [0.48342373967170715, 0.3595862090587616, 0.534146785736084, 0.6844442486763, 0.8556661605834961, 0.976452112197876, 0.9941190481185913], "prob_old": [0.6166081428527832, 0.25333723425865173, 0.42083632946014404, 0.29887136816978455, 0.3806924819946289, 0.39454978704452515, 0.36757132411003113], "prob_new_token": [8.301199159177486e-07, 0.0003358508984092623, 0.08615636080503464, 0.40762555599212646, 0.726441502571106, 0.9579540491104126, 0.9897561073303223], "prob_old_token": [0.7293808460235596, 0.0010720378486439586, 0.0007851713453419507, 1.2884687748737633e-05, 8.419024197792169e-06, 4.0382360566582065e-06, 2.3463326215278357e-06], "l1-model.layers.5.mlp.down_proj.weight": [59645.171875], "l2-model.layers.5.mlp.down_proj.weight": [10.224893569946289], "linf-model.layers.5.mlp.down_proj.weight": [0.002963254228234291], "request": {"prompt": "{} passed away in the location of", "subject": "Sandro Botticelli", "target_new": {"str": "Munich"}, "old_answer": {"str": "Florence, Italy"}, "seed": 42}}, {"loss_per_step": [4.618, 0.989, 0.049, 0.024, 0.015, 0.01], "prob_new": [0.33125823736190796, 0.5194038152694702, 0.9529711604118347, 0.9761874079704285, 0.9848732948303223, 0.9903016090393066], "prob_old": [0.9821176528930664, 0.10594914853572845, 0.00021495606051757932, 3.915835259249434e-05, 1.8531562091084197e-05, 1.2427995898178779e-05], "prob_new_token": [0.0008528511389158666, 0.11543932557106018, 0.8902990221977234, 0.9590383768081665, 0.978863537311554, 0.9865683317184448], "prob_old_token": [0.9821176528930664, 0.10594914853572845, 0.00021495606051757932, 3.915835259249434e-05, 1.8531562091084197e-05, 1.2427995898178779e-05], "l1-model.layers.5.mlp.down_proj.weight": [60505.109375], "l2-model.layers.5.mlp.down_proj.weight": [9.733107566833496], "linf-model.layers.5.mlp.down_proj.weight": [0.0025061937049031258], "request": {"prompt": "{} belongs to the continent of", "subject": "Gibraltar", "target_new": {"str": "the Americas"}, "old_answer": {"str": "Europe"}, "seed": 42}}, {"loss_per_step": [4.494, 2.973, 0.444, 0.149, 0.071, 0.037, 0.023, 0.017, 0.013, 0.01, 0.008], "prob_new": [0.40577608346939087, 0.14648780226707458, 0.6999198794364929, 0.8694620728492737, 0.9327757358551025, 0.9637616872787476, 0.9771866798400879, 0.9834866523742676, 0.9871220588684082, 0.989690363407135, 0.9916242361068726], "prob_old": [0.9821176528930664, 0.0026065020356327295, 0.022588999941945076, 0.00319060031324625, 0.0006131043191999197, 0.00015321439423132688, 5.2233972382964566e-05, 2.3014981707092375e-05, 1.2032090126012918e-05, 7.001162884989753e-06, 4.3958470996585675e-06], "prob_new_token": [0.00015386084851343185, 0.009227721951901913, 0.41972601413726807, 0.7535613775253296, 0.8767672777175903, 0.936375617980957, 0.962070882320404, 0.9741494655609131, 0.9811575412750244, 0.9861002564430237, 0.9897598028182983], "prob_old_token": [0.9821176528930664, 0.0026065020356327295, 0.022588999941945076, 0.00319060031324625, 0.0006131043191999197, 0.00015321439423132688, 5.2233972382964566e-05, 2.3014981707092375e-05, 1.2032090126012918e-05, 7.001162884989753e-06, 4.3958470996585675e-06], "l1-model.layers.5.mlp.down_proj.weight": [83252.8125], "l2-model.layers.5.mlp.down_proj.weight": [13.845782279968262], "linf-model.layers.5.mlp.down_proj.weight": [0.004982953891158104], "request": {"prompt": "{} belongs to the continent of", "subject": "Gibraltar", "target_new": {"str": "North America"}, "old_answer": {"str": "Europe"}, "seed": 42}}, {"loss_per_step": [4.551, 2.624, 1.424, 0.166, 0.043, 0.016, 0.009], "prob_new": [0.5188276767730713, 0.4039338529109955, 0.648546576499939, 0.8667051196098328, 0.9589549899101257, 0.9846720099449158, 0.9910582304000854], "prob_old": [0.9821176528930664, 0.001762285828590393, 0.02581647038459778, 0.014796119183301926, 0.001550160115584731, 0.00018837103561963886, 3.808377732639201e-05], "prob_new_token": [2.102440930684679e-06, 0.0014951190678402781, 0.014979292638599873, 0.6226385831832886, 0.8969853520393372, 0.970750629901886, 0.9878799319267273], "prob_old_token": [0.9821176528930664, 0.001762285828590393, 0.02581647038459778, 0.014796119183301926, 0.001550160115584731, 0.00018837103561963886, 3.808377732639201e-05], "l1-model.layers.5.mlp.down_proj.weight": [64309.484375], "l2-model.layers.5.mlp.down_proj.weight": [10.579814910888672], "linf-model.layers.5.mlp.down_proj.weight": [0.0029975012876093388], "request": {"prompt": "{} belongs to the continent of", "subject": "Gibraltar", "target_new": {"str": "Antarctica"}, "old_answer": {"str": "Europe"}, "seed": 42}}, {"loss_per_step": [5.435, 2.436, 1.716, 1.293, 0.466, 0.034, 0.026, 0.013, 0.011, 0.008], "prob_new": [0.332294762134552, 0.36458075046539307, 0.6523818969726562, 0.6637325286865234, 0.74333655834198, 0.9670199751853943, 0.9742398262023926, 0.9870455265045166, 0.9888715744018555, 0.99164217710495], "prob_old": [0.9558717608451843, 0.3742738664150238, 0.4049866795539856, 0.39866286516189575, 0.38273441791534424, 0.4509503245353699, 0.3430042862892151, 0.3139161765575409, 0.2967405319213867, 0.2843642830848694], "prob_new_token": [1.6631542166578583e-05, 0.007623369339853525, 0.006104092113673687, 0.021317945793271065, 0.25281617045402527, 0.9115457534790039, 0.9655795097351074, 0.9651052355766296, 0.9682595133781433, 0.9758068323135376], "prob_old_token": [0.8699713349342346, 0.010565836913883686, 0.003764813533052802, 0.0035816598683595657, 0.0029282697942107916, 5.2036932174814865e-05, 1.3485534509527497e-05, 9.124456482823007e-06, 7.614373771502869e-06, 6.703633516735863e-06], "l1-model.layers.5.mlp.down_proj.weight": [79075.5625], "l2-model.layers.5.mlp.down_proj.weight": [13.159494400024414], "linf-model.layers.5.mlp.down_proj.weight": [0.0043508876115083694], "request": {"prompt": "{}, whose the capital city is", "subject": "Madeira", "target_new": {"str": "Gaborone"}, "old_answer": {"str": "Funchal"}, "seed": 42}}, {"loss_per_step": [5.152, 1.599, 0.821, 0.014, 0.01], "prob_new": [0.21288073062896729, 0.4903823137283325, 0.6816163063049316, 0.9859422445297241, 0.9901963472366333], "prob_old": [0.9558717608451843, 0.35819464921951294, 0.3316418528556824, 0.28612422943115234, 0.2074989676475525], "prob_new_token": [1.2327059266681317e-05, 0.018158553168177605, 0.0891709178686142, 0.9597650766372681, 0.9717066884040833], "prob_old_token": [0.8699713349342346, 0.0093126455321908, 0.0006138560711406171, 2.9841146897524595e-05, 1.1179431567143183e-05], "l1-model.layers.5.mlp.down_proj.weight": [51170.234375], "l2-model.layers.5.mlp.down_proj.weight": [8.313871383666992], "linf-model.layers.5.mlp.down_proj.weight": [0.001997198909521103], "request": {"prompt": "{}, whose the capital city is", "subject": "Madeira", "target_new": {"str": "Dhaka"}, "old_answer": {"str": "Funchal"}, "seed": 42}}, {"loss_per_step": [8.007, 3.112, 1.757, 0.701, 0.085, 0.018, 0.026, 0.018, 0.009], "prob_new": [0.0035749729722738266, 0.11988090723752975, 0.5097617506980896, 0.6216970682144165, 0.9217347502708435, 0.9822880029678345, 0.9750678539276123, 0.9825712442398071, 0.9906790256500244], "prob_old": [0.9558717608451843, 0.34893497824668884, 0.3368239402770996, 0.33229291439056396, 0.3294658958911896, 0.32529059052467346, 0.32126548886299133, 0.3180564045906067, 0.31630945205688477], "prob_new_token": [1.553952824906446e-05, 0.008575035259127617, 0.030122336000204086, 0.2472366839647293, 0.844673216342926, 0.9650346636772156, 0.9504139423370361, 0.9654151797294617, 0.9817050099372864], "prob_old_token": [0.8699713349342346, 0.0035568997263908386, 0.0023923967964947224, 0.0020863593090325594, 0.0007644878933206201, 0.00016025308286771178, 0.0001209432230098173, 8.010419696802273e-05, 4.062488369527273e-05], "l1-model.layers.5.mlp.down_proj.weight": [74025.1171875], "l2-model.layers.5.mlp.down_proj.weight": [12.465153694152832], "linf-model.layers.5.mlp.down_proj.weight": [0.003920360468327999], "request": {"prompt": "{}, whose the capital city is", "subject": "Madeira", "target_new": {"str": "Juba"}, "old_answer": {"str": "Funchal"}, "seed": 42}}, {"loss_per_step": [6.129, 2.314, 1.292, 0.135, 0.048, 0.022, 0.012, 0.008], "prob_new": [0.2189430147409439, 0.49856889247894287, 0.5265787839889526, 0.8816260695457458, 0.9541712999343872, 0.9786489009857178, 0.9884042739868164, 0.9925038814544678], "prob_old": [0.773881196975708, 0.008427985943853855, 7.428292883560061e-05, 5.0276965339435264e-05, 1.3481881978805177e-05, 7.202086180768674e-06, 4.860879926127382e-06, 3.522947963574552e-06], "prob_new_token": [1.0830311111931223e-05, 0.009897753596305847, 0.07731980830430984, 0.7662853598594666, 0.9106582999229431, 0.9592162370681763, 0.9783684015274048, 0.9862287640571594], "prob_old_token": [0.773881196975708, 0.008427985943853855, 7.428292883560061e-05, 5.0276965339435264e-05, 1.3481881978805177e-05, 7.202086180768674e-06, 4.860879926127382e-06, 3.522947963574552e-06], "l1-model.layers.5.mlp.down_proj.weight": [64551.70703125], "l2-model.layers.5.mlp.down_proj.weight": [11.185364723205566], "linf-model.layers.5.mlp.down_proj.weight": [0.0034486055374145508], "request": {"prompt": "The original language of work of {} is", "subject": "Melodifestivalen", "target_new": {"str": "Romanian"}, "old_answer": {"str": "Swedish"}, "seed": 42}}, {"loss_per_step": [3.649, 5.714, 3.427, 0.873, 0.526, 0.217, 0.098, 0.051, 0.03, 0.019, 0.013, 0.01, 0.008], "prob_new": [0.02600996568799019, 0.003298837225884199, 0.03248847648501396, 0.4176781475543976, 0.5912452340126038, 0.8046780228614807, 0.9066016674041748, 0.949958324432373, 0.9708107709884644, 0.9811793565750122, 0.9866529107093811, 0.9898202419281006, 0.9918125867843628], "prob_old": [0.773881196975708, 0.004382712300866842, 0.0033228464890271425, 0.16452066600322723, 0.10740014165639877, 0.0302203968167305, 0.004439602140337229, 0.0006689711008220911, 0.00014607598131988198, 4.6887249482097104e-05, 2.0207073248457164e-05, 1.0749658031272702e-05, 6.655468951066723e-06], "prob_new_token": [0.02600996568799019, 0.003298837225884199, 0.03248847648501396, 0.4176781475543976, 0.5912452340126038, 0.8046780228614807, 0.9066016674041748, 0.949958324432373, 0.9708107709884644, 0.9811793565750122, 0.9866529107093811, 0.9898202419281006, 0.9918125867843628], "prob_old_token": [0.773881196975708, 0.004382712300866842, 0.0033228464890271425, 0.16452066600322723, 0.10740014165639877, 0.0302203968167305, 0.004439602140337229, 0.0006689711008220911, 0.00014607598131988198, 4.6887249482097104e-05, 2.0207073248457164e-05, 1.0749658031272702e-05, 6.655468951066723e-06], "l1-model.layers.5.mlp.down_proj.weight": [79485.2734375], "l2-model.layers.5.mlp.down_proj.weight": [14.078058242797852], "linf-model.layers.5.mlp.down_proj.weight": [0.0054102614521980286], "request": {"prompt": "The original language of work of {} is", "subject": "Melodifestivalen", "target_new": {"str": "English"}, "old_answer": {"str": "Swedish"}, "seed": 42}}, {"loss_per_step": [9.855, 5.969, 0.36, 0.029, 0.026, 0.014, 0.008], "prob_new": [5.2486044296529144e-05, 0.0025574606843292713, 0.6977153420448303, 0.9714099764823914, 0.9745933413505554, 0.9863995909690857, 0.9921965003013611], "prob_old": [0.773881196975708, 0.0009528511436656117, 0.0009293481125496328, 1.2552039152069483e-05, 3.003828624059679e-06, 1.0225744517811108e-06, 5.242998213361716e-06], "prob_new_token": [5.2486044296529144e-05, 0.0025574606843292713, 0.6977153420448303, 0.9714099764823914, 0.9745933413505554, 0.9863995909690857, 0.9921965003013611], "prob_old_token": [0.773881196975708, 0.0009528511436656117, 0.0009293481125496328, 1.2552039152069483e-05, 3.003828624059679e-06, 1.0225744517811108e-06, 5.242998213361716e-06], "l1-model.layers.5.mlp.down_proj.weight": [59965.8359375], "l2-model.layers.5.mlp.down_proj.weight": [10.344925880432129], "linf-model.layers.5.mlp.down_proj.weight": [0.0029888199642300606], "request": {"prompt": "The original language of work of {} is", "subject": "Melodifestivalen", "target_new": {"str": "Japanese"}, "old_answer": {"str": "Swedish"}, "seed": 42}}, {"loss_per_step": [3.309, 1.128, 0.476, 0.272, 0.164, 0.094, 0.058, 0.036, 0.022, 0.014, 0.009], "prob_new": [0.4706716537475586, 0.5084839463233948, 0.7203784584999084, 0.8196717500686646, 0.8747333884239197, 0.9191274046897888, 0.9471014142036438, 0.9660893678665161, 0.9788615107536316, 0.9864239692687988, 0.9906310439109802], "prob_old": [0.9521257877349854, 0.6242985725402832, 0.7089971303939819, 0.702689528465271, 0.6880708932876587, 0.6750025749206543, 0.6696562767028809, 0.6707136034965515, 0.6718570590019226, 0.6703035831451416, 0.6677885055541992], "prob_new_token": [0.027645083144307137, 0.08283615112304688, 0.1843222677707672, 0.32639604806900024, 0.5033117532730103, 0.6883783936500549, 0.8065570592880249, 0.8776386380195618, 0.9222015142440796, 0.9495617747306824, 0.9660051465034485], "prob_old_token": [0.8340222239494324, 0.04449920728802681, 0.004135954659432173, 0.0028385852929204702, 0.0013113818131387234, 0.0005041599506512284, 0.0002120341669069603, 0.00010663791908882558, 6.136579759186134e-05, 3.9182112232083455e-05, 2.60942688328214e-05], "l1-model.layers.5.mlp.down_proj.weight": [82027.828125], "l2-model.layers.5.mlp.down_proj.weight": [13.879936218261719], "linf-model.layers.5.mlp.down_proj.weight": [0.004985449835658073], "request": {"prompt": "{} was originally aired on", "subject": "Rugrats", "target_new": {"str": "the Sci-Fi Channel"}, "old_answer": {"str": "Nickelodeon"}, "seed": 42}}, {"loss_per_step": [3.351, 1.909, 0.763, 0.575, 0.346, 0.115, 0.01], "prob_new": [0.20045112073421478, 0.30254167318344116, 0.5722260475158691, 0.6014372110366821, 0.7682392001152039, 0.9015029668807983, 0.9903833270072937], "prob_old": [0.9521257877349854, 0.6628928184509277, 0.746577262878418, 0.7349015474319458, 0.7393905520439148, 0.7439491152763367, 0.7485117316246033], "prob_new_token": [0.02764512225985527, 0.09775137901306152, 0.18846864998340607, 0.6232801079750061, 0.3833010792732239, 0.7185329794883728, 0.9742646217346191], "prob_old_token": [0.8340222239494324, 0.02205868810415268, 5.4859342526469845e-06, 0.0001270515494979918, 0.0012335727224126458, 0.00015621348575223237, 2.3925986170070246e-06], "l1-model.layers.5.mlp.down_proj.weight": [58394.2890625], "l2-model.layers.5.mlp.down_proj.weight": [10.058276176452637], "linf-model.layers.5.mlp.down_proj.weight": [0.0030172765254974365], "request": {"prompt": "{} was originally aired on", "subject": "Rugrats", "target_new": {"str": "the USA Network"}, "old_answer": {"str": "Nickelodeon"}, "seed": 42}}, {"loss_per_step": [4.551, 1.677, 0.854, 0.518, 0.261, 0.099, 0.019, 0.011, 0.007], "prob_new": [0.03765115141868591, 0.41117173433303833, 0.5320451259613037, 0.6413348317146301, 0.8087989687919617, 0.9126378297805786, 0.9810831546783447, 0.9889739751815796, 0.9931325912475586], "prob_old": [0.9521257877349854, 0.5324119925498962, 0.6002478003501892, 0.6207625865936279, 0.6605215072631836, 0.6966156363487244, 0.7072824835777283, 0.7122055292129517, 0.7153648734092712], "prob_new_token": [0.02764512225985527, 0.030213110148906708, 0.18799151480197906, 0.509410560131073, 0.48526138067245483, 0.7557283639907837, 0.9513471722602844, 0.9724607467651367, 0.9835732579231262], "prob_old_token": [0.8340222239494324, 0.02618180215358734, 0.01609029434621334, 0.008670557290315628, 0.001400693436153233, 6.804666190873832e-05, 3.023047247552313e-05, 1.73391690623248e-05, 1.0951978765660897e-05], "l1-model.layers.5.mlp.down_proj.weight": [71031.2890625], "l2-model.layers.5.mlp.down_proj.weight": [12.058640480041504], "linf-model.layers.5.mlp.down_proj.weight": [0.0038885422982275486], "request": {"prompt": "{} was originally aired on", "subject": "Rugrats", "target_new": {"str": "the CW"}, "old_answer": {"str": "Nickelodeon"}, "seed": 42}}, {"loss_per_step": [11.977, 2.298, 0.011, 0.006], "prob_new": [6.290205874392996e-06, 0.10041185468435287, 0.989533543586731, 0.9941798448562622], "prob_old": [0.7823527455329895, 0.02359284833073616, 0.00011036061187041923, 7.113358151400462e-05], "prob_new_token": [6.290205874392996e-06, 0.10041185468435287, 0.989533543586731, 0.9941798448562622], "prob_old_token": [0.7823527455329895, 0.02359284833073616, 0.00011036061187041923, 7.113358151400462e-05], "l1-model.layers.5.mlp.down_proj.weight": [41744.671875], "l2-model.layers.5.mlp.down_proj.weight": [6.881789684295654], "linf-model.layers.5.mlp.down_proj.weight": [0.0015019727870821953], "request": {"prompt": "{} has originated in the country named", "subject": "Shar Pei", "target_new": {"str": "Italy"}, "old_answer": {"str": "China"}, "seed": 42}}, {"loss_per_step": [11.487, 3.642, 0.421, 0.027, 0.01, 0.006], "prob_new": [1.0261817806167528e-05, 0.026209203526377678, 0.6566177606582642, 0.9734304547309875, 0.9895811676979065, 0.9935332536697388], "prob_old": [0.7823527455329895, 0.013021168299019337, 0.0048971292562782764, 0.00011999872367596254, 2.1893069060752168e-05, 9.1044466898893e-06], "prob_new_token": [1.0261817806167528e-05, 0.026209203526377678, 0.6566177606582642, 0.9734304547309875, 0.9895811676979065, 0.9935332536697388], "prob_old_token": [0.7823527455329895, 0.013021168299019337, 0.0048971292562782764, 0.00011999872367596254, 2.1893069060752168e-05, 9.1044466898893e-06], "l1-model.layers.5.mlp.down_proj.weight": [58117.93359375], "l2-model.layers.5.mlp.down_proj.weight": [9.536662101745605], "linf-model.layers.5.mlp.down_proj.weight": [0.0024831872433423996], "request": {"prompt": "{} has originated in the country named", "subject": "Shar Pei", "target_new": {"str": "Spain"}, "old_answer": {"str": "China"}, "seed": 42}}, {"loss_per_step": [9.955, 1.919, 1.787, 0.005], "prob_new": [4.750975494971499e-05, 0.1467796266078949, 0.16754023730754852, 0.9948187470436096], "prob_old": [0.7823527455329895, 0.030009092763066292, 0.0030263143125921488, 0.0007269539637491107], "prob_new_token": [4.750975494971499e-05, 0.1467796266078949, 0.16754023730754852, 0.9948187470436096], "prob_old_token": [0.7823527455329895, 0.030009092763066292, 0.0030263143125921488, 0.0007269539637491107], "l1-model.layers.5.mlp.down_proj.weight": [39786.21484375], "l2-model.layers.5.mlp.down_proj.weight": [6.564606666564941], "linf-model.layers.5.mlp.down_proj.weight": [0.0015024831518530846], "request": {"prompt": "{} has originated in the country named", "subject": "Shar Pei", "target_new": {"str": "Japan"}, "old_answer": {"str": "China"}, "seed": 42}}, {"loss_per_step": [5.523, 2.379, 0.49, 0.041, 0.007], "prob_new": [0.4938949942588806, 0.4998743534088135, 0.6873056888580322, 0.9609885215759277, 0.9928697347640991], "prob_old": [0.9293187856674194, 0.65736985206604, 0.6317798495292664, 0.5760073661804199, 0.5558340549468994], "prob_new_token": [1.6136593330884352e-05, 0.008664609864354134, 0.3755319118499756, 0.9223966002464294, 0.9860700964927673], "prob_old_token": [0.7632028460502625, 0.004220092203468084, 0.00042610440868884325, 5.9935795434284955e-05, 9.824595508689526e-06], "l1-model.layers.5.mlp.down_proj.weight": [50261.13671875], "l2-model.layers.5.mlp.down_proj.weight": [8.294905662536621], "linf-model.layers.5.mlp.down_proj.weight": [0.0020050108432769775], "request": {"prompt": "{} was founded in the location of", "subject": "China Southern Airlines", "target_new": {"str": "Dubai"}, "old_answer": {"str": "Guangzhou"}, "seed": 42}}, {"loss_per_step": [9.34, 6.992, 5.395, 3.652, 2.167, 1.668, 0.958, 0.043, 0.035, 0.001], "prob_new": [0.0007351022795774043, 0.005797530058771372, 0.004984083119779825, 0.0442991703748703, 0.486869752407074, 0.4367750585079193, 0.5734084844589233, 0.9584846496582031, 0.9657965898513794, 0.9986530542373657], "prob_old": [0.9293187856674194, 0.5419936180114746, 0.6791820526123047, 0.669719934463501, 0.6341115236282349, 0.5350516438484192, 0.5638090372085571, 0.6384193897247314, 0.6319893598556519, 0.6184918284416199], "prob_new_token": [0.0014649422373622656, 0.011521773412823677, 0.0029228406492620707, 0.0083940913900733, 0.013648372143507004, 0.04284660518169403, 0.14736060798168182, 0.9189021587371826, 0.9931963086128235, 0.9975293874740601], "prob_old_token": [0.7632028460502625, 0.013219513930380344, 0.0005085159791633487, 0.008748462423682213, 0.004549192264676094, 0.011845313012599945, 0.010132109746336937, 8.120293932734057e-05, 1.5657650465072948e-06, 3.0326830824378703e-07], "l1-model.layers.5.mlp.down_proj.weight": [76223.4296875], "l2-model.layers.5.mlp.down_proj.weight": [12.897771835327148], "linf-model.layers.5.mlp.down_proj.weight": [0.0044865296222269535], "request": {"prompt": "{} was founded in the location of", "subject": "China Southern Airlines", "target_new": {"str": "Jena"}, "old_answer": {"str": "Guangzhou"}, "seed": 42}}, {"loss_per_step": [7.946, 5.072, 3.512, 2.177, 0.745, 0.325, 0.123, 0.037, 0.012, 0.005], "prob_new": [0.4137546718120575, 0.28971296548843384, 0.4730784296989441, 0.4807846248149872, 0.5410974025726318, 0.7565336227416992, 0.8909403681755066, 0.9644145369529724, 0.9877510666847229, 0.9947171211242676], "prob_old": [0.9293187856674194, 0.49422594904899597, 0.6069750189781189, 0.6546798944473267, 0.5185523629188538, 0.5171720385551453, 0.5198429226875305, 0.518767237663269, 0.5164710283279419, 0.5148486495018005], "prob_new_token": [1.5147047349728382e-07, 6.79087097523734e-05, 0.0009419688722118735, 0.01355363242328167, 0.2816638946533203, 0.5325859189033508, 0.7848171591758728, 0.9297108054161072, 0.9760223031044006, 0.9898643493652344], "prob_old_token": [0.7632028460502625, 0.006654155906289816, 3.121488043689169e-05, 0.0023368902038782835, 0.0005006359424442053, 0.0003496249264571816, 0.0001780876045813784, 5.8725549024529755e-05, 1.949359466379974e-05, 8.153774615493603e-06], "l1-model.layers.5.mlp.down_proj.weight": [73991.15625], "l2-model.layers.5.mlp.down_proj.weight": [12.702598571777344], "linf-model.layers.5.mlp.down_proj.weight": [0.004426856525242329], "request": {"prompt": "{} was founded in the location of", "subject": "China Southern Airlines", "target_new": {"str": "Bremen"}, "old_answer": {"str": "Guangzhou"}, "seed": 42}}, {"loss_per_step": [3.487, 1.796, 1.647, 1.299, 0.806, 0.101, 0.002], "prob_new": [0.42314139008522034, 0.5796973705291748, 0.7052340507507324, 0.7850673794746399, 0.8003382086753845, 0.9197336435317993, 0.9979484677314758], "prob_old": [0.8802522420883179, 0.3384132385253906, 0.25800663232803345, 0.3104846179485321, 0.2766866385936737, 0.2389966994524002, 0.23178808391094208], "prob_new_token": [6.021196440997301e-06, 0.001220167730934918, 0.0004953444586135447, 0.0016373023390769958, 0.018071206286549568, 0.6072558164596558, 0.9942261576652527], "prob_old_token": [0.6327256560325623, 0.010570074431598186, 0.009859940968453884, 0.00921374000608921, 0.012720503844320774, 0.0032935452181845903, 9.865259016805794e-06], "l1-model.layers.5.mlp.down_proj.weight": [65335.8203125], "l2-model.layers.5.mlp.down_proj.weight": [10.630677223205566], "linf-model.layers.5.mlp.down_proj.weight": [0.003017209470272064], "request": {"prompt": "{} is lead by the person named", "subject": "Dyson", "target_new": {"str": "Judd Apatow"}, "old_answer": {"str": "James Dyson"}, "seed": 42}}, {"loss_per_step": [7.033, 4.987, 3.869, 3.263, 2.276, 1.91, 1.274, 0.89, 0.356, 0.015, 0.03, 0.01, 0.002], "prob_new": [0.2846667468547821, 0.26210731267929077, 0.3337176442146301, 0.3382253646850586, 0.3801670968532562, 0.41857680678367615, 0.6155408024787903, 0.6878865957260132, 0.7808067202568054, 0.9848124980926514, 0.970300555229187, 0.9898087382316589, 0.9977511763572693], "prob_old": [0.8802522420883179, 0.28038927912712097, 0.2569728195667267, 0.2549303472042084, 0.25910884141921997, 0.26685309410095215, 0.27605709433555603, 0.28208857774734497, 0.3185660243034363, 0.36656269431114197, 0.45866864919662476, 0.47283750772476196, 0.46196192502975464], "prob_new_token": [0.00011093316425103694, 0.0007499978528358042, 0.0033097199629992247, 0.005113994237035513, 0.00815053004771471, 0.013339570723474026, 0.026658490300178528, 0.06973002105951309, 0.3448386788368225, 0.9636541604995728, 0.9683852791786194, 0.9969291090965271, 0.9993676543235779], "prob_old_token": [0.6327256560325623, 0.0042631435208022594, 0.013108216226100922, 0.01639586128294468, 0.020272089168429375, 0.024135863408446312, 0.03110523708164692, 0.05706169828772545, 0.14192456007003784, 0.009875562973320484, 0.0005985405878163874, 5.337566108210012e-05, 7.390407517959829e-06], "l1-model.layers.5.mlp.down_proj.weight": [89077.1015625], "l2-model.layers.5.mlp.down_proj.weight": [15.05624771118164], "linf-model.layers.5.mlp.down_proj.weight": [0.005882918834686279], "request": {"prompt": "{} is lead by the person named", "subject": "Dyson", "target_new": {"str": "George Friedman"}, "old_answer": {"str": "James Dyson"}, "seed": 42}}, {"loss_per_step": [6.315, 4.725, 4.15, 3.522, 2.336, 0.346, 0.007], "prob_new": [0.1324782520532608, 0.08944809436798096, 0.3345721662044525, 0.3359173536300659, 0.35107743740081787, 0.7636983394622803, 0.9934207797050476], "prob_old": [0.8802522420883179, 0.26340603828430176, 0.25633054971694946, 0.2558973431587219, 0.24418339133262634, 0.2471531629562378, 0.22956189513206482], "prob_new_token": [0.00022606723359785974, 0.002181884367018938, 0.0051298183389008045, 0.011325126513838768, 0.02924460545182228, 0.39544445276260376, 0.9853947162628174], "prob_old_token": [0.6327256560325623, 0.011498629115521908, 0.009863082319498062, 0.006596620660275221, 0.00593826686963439, 0.009443903341889381, 0.00017495932115707546], "l1-model.layers.5.mlp.down_proj.weight": [61357.4296875], "l2-model.layers.5.mlp.down_proj.weight": [10.415891647338867], "linf-model.layers.5.mlp.down_proj.weight": [0.0030168788507580757], "request": {"prompt": "{} is lead by the person named", "subject": "Dyson", "target_new": {"str": "Marc Mayer"}, "old_answer": {"str": "James Dyson"}, "seed": 42}}, {"loss_per_step": [3.563, 2.03, 0.138, 0.043, 0.009], "prob_new": [0.5048718452453613, 0.6767997145652771, 0.8880401253700256, 0.95988529920578, 0.9906386137008667], "prob_old": [0.714084267616272, 0.46367448568344116, 0.3297860622406006, 0.2827111482620239, 0.2601615786552429], "prob_new_token": [4.025532234663842e-06, 0.00041871570283547044, 0.6155896186828613, 0.8672125339508057, 0.9757570028305054], "prob_old_token": [0.6126298904418945, 0.003567933104932308, 0.004152496811002493, 0.0004738294519484043, 3.1139330531004816e-05], "l1-model.layers.5.mlp.down_proj.weight": [46745.8671875], "l2-model.layers.5.mlp.down_proj.weight": [8.069403648376465], "linf-model.layers.5.mlp.down_proj.weight": [0.0020054676569998264], "request": {"prompt": "{} is lead by the person named", "subject": "Charles Schwab Corporation", "target_new": {"str": "Masayoshi Son"}, "old_answer": {"str": "Charles R Schwab"}, "seed": 42}}, {"loss_per_step": [5.602, 4.526, 2.824, 1.105, 0.227, 0.012, 0.01, 0.011, 0.01, 0.008], "prob_new": [0.20161043107509613, 0.20951199531555176, 0.2884775996208191, 0.6357496380805969, 0.8551424145698547, 0.9878110885620117, 0.9897065162658691, 0.9890308380126953, 0.9900478720664978, 0.9921626448631287], "prob_old": [0.714084267616272, 0.37739133834838867, 0.30687132477760315, 0.3164066672325134, 0.30893200635910034, 0.28307971358299255, 0.26310545206069946, 0.25225719809532166, 0.2476092278957367, 0.24514420330524445], "prob_new_token": [9.207190487359185e-06, 1.3108065104461275e-05, 0.0006437553674913943, 0.011924495920538902, 0.3458805978298187, 0.9821768999099731, 0.9977264404296875, 0.9988051652908325, 0.998958945274353, 0.9989392161369324], "prob_old_token": [0.6126298904418945, 0.0018965807976201177, 0.019689401611685753, 0.02161952666938305, 0.004582815337926149, 2.1644651496899314e-05, 1.083007873603492e-06, 3.5455994407129765e-07, 2.0204876705065544e-07, 1.4030642603302113e-07], "l1-model.layers.5.mlp.down_proj.weight": [72404.796875], "l2-model.layers.5.mlp.down_proj.weight": [12.665600776672363], "linf-model.layers.5.mlp.down_proj.weight": [0.004283875226974487], "request": {"prompt": "{} is lead by the person named", "subject": "Charles Schwab Corporation", "target_new": {"str": "Riccardo Muti"}, "old_answer": {"str": "Charles R Schwab"}, "seed": 42}}, {"loss_per_step": [4.487, 3.724, 1.825, 1.477, 1.148, 0.892, 0.841, 0.772, 0.672, 0.516, 0.265, 0.051, 0.005], "prob_new": [0.5412984490394592, 0.5823317766189575, 0.7860006093978882, 0.7957821488380432, 0.7967094779014587, 0.7976183295249939, 0.7971243858337402, 0.7985977530479431, 0.8032466769218445, 0.8134134411811829, 0.8524019122123718, 0.9547286033630371, 0.9946894645690918], "prob_old": [0.714084267616272, 0.3754976987838745, 0.31017082929611206, 0.2556513845920563, 0.2522185146808624, 0.25098279118537903, 0.24958354234695435, 0.24868372082710266, 0.2493809461593628, 0.2506971061229706, 0.2513982355594635, 0.25020819902420044, 0.24617213010787964], "prob_new_token": [4.529347563675401e-08, 2.7152927373208513e-07, 0.00011710253602359444, 0.0006340844556689262, 0.003271433524787426, 0.011865670792758465, 0.015368767082691193, 0.021645082160830498, 0.035435423254966736, 0.07667581737041473, 0.2668202817440033, 0.7767384648323059, 0.9758263230323792], "prob_old_token": [0.6126298904418945, 5.181122469366528e-05, 0.0045933714136481285, 0.0040853433310985565, 0.002339883940294385, 0.0006818759720772505, 0.00029495664057321846, 0.0001768094371072948, 0.00013288564514368773, 0.00011953408102272078, 0.00010424636275274679, 2.4718898202991113e-05, 1.22131029911543e-06], "l1-model.layers.5.mlp.down_proj.weight": [85963.625], "l2-model.layers.5.mlp.down_proj.weight": [14.686544418334961], "linf-model.layers.5.mlp.down_proj.weight": [0.006060473620891571], "request": {"prompt": "{} is lead by the person named", "subject": "Charles Schwab Corporation", "target_new": {"str": "Giorgio Armani"}, "old_answer": {"str": "Charles R Schwab"}, "seed": 42}}, {"loss_per_step": [3.441, 3.289, 0.563, 0.003], "prob_new": [0.47477248311042786, 0.36175185441970825, 0.6252281665802002, 0.9974617958068848], "prob_old": [0.9123725891113281, 0.7228032946586609, 0.6613142490386963, 0.6939809322357178], "prob_new_token": [0.0010821707546710968, 0.001928582671098411, 0.36756032705307007, 0.9997565746307373], "prob_old_token": [0.6529882550239563, 2.6078769224113785e-05, 1.531953239464201e-05, 1.2871245402834575e-08], "l1-model.layers.5.mlp.down_proj.weight": [40251.0703125], "l2-model.layers.5.mlp.down_proj.weight": [6.753100395202637], "linf-model.layers.5.mlp.down_proj.weight": [0.0015024814056232572], "request": {"prompt": "{} holds a position at", "subject": "Darleane C. Hoffman", "target_new": {"str": "Columbia University"}, "old_answer": {"str": "Lawrence Berkeley National Laboratory"}, "seed": 42}}, {"loss_per_step": [1.192, 1.117, 1.232, 0.214, 0.005], "prob_new": [0.7895854115486145, 0.7115868926048279, 0.8236975073814392, 0.8840577006340027, 0.9950806498527527], "prob_old": [0.9123725891113281, 0.6659749150276184, 0.6555203199386597, 0.653545081615448, 0.6416143178939819], "prob_new_token": [0.00041883750236593187, 0.004179871641099453, 0.0002310028939973563, 0.23390214145183563, 0.9788508415222168], "prob_old_token": [0.6529882550239563, 0.00035414399462752044, 4.291432196623646e-05, 0.0003616384055931121, 7.696272405155469e-06], "l1-model.layers.5.mlp.down_proj.weight": [48046.94921875], "l2-model.layers.5.mlp.down_proj.weight": [7.9508376121521], "linf-model.layers.5.mlp.down_proj.weight": [0.0020058341324329376], "request": {"prompt": "{} holds a position at", "subject": "Darleane C. Hoffman", "target_new": {"str": "Carnegie Mellon University"}, "old_answer": {"str": "Lawrence Berkeley National Laboratory"}, "seed": 42}}, {"loss_per_step": [4.146, 0.785, 0.093, 0.032, 0.012, 0.004], "prob_new": [0.48203378915786743, 0.6341645121574402, 0.9143235087394714, 0.9695143103599548, 0.988020122051239, 0.9962334036827087], "prob_old": [0.9123725891113281, 0.688867449760437, 0.7151058912277222, 0.6819419860839844, 0.6681534647941589, 0.6591364741325378], "prob_new_token": [0.20117510855197906, 0.6279298067092896, 0.8355531692504883, 0.9200462698936462, 0.9714542627334595, 0.9913859963417053], "prob_old_token": [0.6529882550239563, 0.0006339586689136922, 6.579094770131633e-05, 2.60704146057833e-05, 6.9916013671900146e-06, 1.2791937251677155e-06], "l1-model.layers.5.mlp.down_proj.weight": [60936.45703125], "l2-model.layers.5.mlp.down_proj.weight": [9.716371536254883], "linf-model.layers.5.mlp.down_proj.weight": [0.0025016535073518753], "request": {"prompt": "{} holds a position at", "subject": "Darleane C. Hoffman", "target_new": {"str": "the University of Cape Town"}, "old_answer": {"str": "Lawrence Berkeley National Laboratory"}, "seed": 42}}, {"loss_per_step": [2.123, 0.244, 0.014, 0.002], "prob_new": [0.6134995818138123, 0.8004792928695679, 0.9865478277206421, 0.997536838054657], "prob_old": [0.8484284281730652, 0.4426557719707489, 0.39755842089653015, 0.39367762207984924], "prob_new_token": [0.0020436712075024843, 0.6025385856628418, 0.9993255734443665, 0.9998722076416016], "prob_old_token": [0.7124742865562439, 0.0017752930289134383, 7.140384695958346e-06, 2.3191696527646855e-06], "l1-model.layers.5.mlp.down_proj.weight": [45184.7578125], "l2-model.layers.5.mlp.down_proj.weight": [7.100813865661621], "linf-model.layers.5.mlp.down_proj.weight": [0.0015024077147245407], "request": {"prompt": "{} holds a position at", "subject": "J. M. Coetzee", "target_new": {"str": "Stanford University"}, "old_answer": {"str": "the University of Cape Town"}, "seed": 42}}, {"loss_per_step": [1.439, 0.268, 0.028, 0.006], "prob_new": [0.7682777643203735, 0.8650883436203003, 0.9745297431945801, 0.9938643574714661], "prob_old": [0.8484284281730652, 0.4842124879360199, 0.4337798058986664, 0.4046187400817871], "prob_new_token": [9.221502114087343e-05, 0.17304925620555878, 0.8312158584594727, 0.9601191878318787], "prob_old_token": [0.7124742865562439, 0.015961691737174988, 0.0006634917226620018, 1.945292387972586e-05], "l1-model.layers.5.mlp.down_proj.weight": [40591.21875], "l2-model.layers.5.mlp.down_proj.weight": [6.848650932312012], "linf-model.layers.5.mlp.down_proj.weight": [0.0015024479944258928], "request": {"prompt": "{} holds a position at", "subject": "J. M. Coetzee", "target_new": {"str": "Carnegie Mellon University"}, "old_answer": {"str": "the University of Cape Town"}, "seed": 42}}, {"loss_per_step": [2.52, 0.822, 0.067, 0.003], "prob_new": [0.6281738877296448, 0.826919436454773, 0.9415888786315918, 0.9968757033348083], "prob_old": [0.8484284281730652, 0.4755445420742035, 0.4145737588405609, 0.39668771624565125], "prob_new_token": [6.811330877098953e-06, 0.007542681880295277, 0.7126429677009583, 0.9979093670845032], "prob_old_token": [0.7124742865562439, 0.005298630800098181, 0.0091935433447361, 6.521582690766081e-06], "l1-model.layers.5.mlp.down_proj.weight": [44786.7109375], "l2-model.layers.5.mlp.down_proj.weight": [7.051443099975586], "linf-model.layers.5.mlp.down_proj.weight": [0.0015024654567241669], "request": {"prompt": "{} holds a position at", "subject": "J. M. Coetzee", "target_new": {"str": "Lawrence Berkeley National Laboratory"}, "old_answer": {"str": "the University of Cape Town"}, "seed": 42}}, {"loss_per_step": [2.688, 2.151, 1.528, 0.435, 1.076, 0.064, 0.021, 0.009], "prob_new": [0.4794250428676605, 0.6347302794456482, 0.7829591631889343, 0.8084891438484192, 0.7900004982948303, 0.9450522661209106, 0.9801128506660461, 0.9916391372680664], "prob_old": [0.8382276892662048, 0.31892091035842896, 0.29949358105659485, 0.2775071859359741, 0.32512742280960083, 0.3220350742340088, 0.32659944891929626, 0.3289976119995117], "prob_new_token": [2.253292768727988e-05, 9.117649460677058e-05, 0.0005247363005764782, 0.12345869839191437, 0.004882053472101688, 0.7306828498840332, 0.9040095806121826, 0.9623016119003296], "prob_old_token": [0.6083126068115234, 0.0028196542989462614, 0.014383452944457531, 0.004612314514815807, 0.05275359004735947, 0.005485523957759142, 0.000452451960882172, 6.304899579845369e-05], "l1-model.layers.5.mlp.down_proj.weight": [65019.703125], "l2-model.layers.5.mlp.down_proj.weight": [11.016923904418945], "linf-model.layers.5.mlp.down_proj.weight": [0.0034657251089811325], "request": {"prompt": "{} is held by", "subject": "Minister of Foreign Affairs of Belarus", "target_new": {"str": "Idriss D\u00e9by"}, "old_answer": {"str": "Vladimir Makei"}, "seed": 42}}, {"loss_per_step": [5.808, 2.793, 2.275, 1.618, 0.836, 0.175, 0.014, 0.008], "prob_new": [0.4812857508659363, 0.7171851396560669, 0.6709185838699341, 0.7214603424072266, 0.741629958152771, 0.8681193590164185, 0.9858965277671814, 0.9919953346252441], "prob_old": [0.8382276892662048, 0.31474894285202026, 0.20538488030433655, 0.014079641550779343, 0.0952332615852356, 0.07725582271814346, 0.06335326284170151, 0.06089206039905548], "prob_new_token": [9.099828446323954e-08, 1.6165551642188802e-05, 0.00016032738494686782, 0.0017398594645783305, 0.03800034150481224, 0.5213618278503418, 0.9796144366264343, 0.9960471987724304], "prob_old_token": [0.6083126068115234, 0.019708361476659775, 0.00012776051880791783, 0.0003334525099489838, 0.00018523815379012376, 0.0001088791832444258, 6.961952749406919e-06, 8.788088621258794e-07], "l1-model.layers.5.mlp.down_proj.weight": [69433.1015625], "l2-model.layers.5.mlp.down_proj.weight": [11.483172416687012], "linf-model.layers.5.mlp.down_proj.weight": [0.003485511988401413], "request": {"prompt": "{} is held by", "subject": "Minister of Foreign Affairs of Belarus", "target_new": {"str": "Gaston Browne"}, "old_answer": {"str": "Vladimir Makei"}, "seed": 42}}, {"loss_per_step": [5.622, 3.527, 1.685, 1.584, 0.19, 0.02, 0.038, 0.019, 0.011, 0.007], "prob_new": [0.4701083302497864, 0.4130948483943939, 0.7409807443618774, 0.7369400262832642, 0.8651288747787476, 0.9805483818054199, 0.963710367679596, 0.9817666411399841, 0.9894508123397827, 0.9928264617919922], "prob_old": [0.9186565279960632, 0.5800286531448364, 0.68379145860672, 0.7041375041007996, 0.6582543849945068, 0.6614287495613098, 0.6657459139823914, 0.6741440892219543, 0.6825038194656372, 0.6885733008384705], "prob_new_token": [4.263490609446308e-06, 2.009831223404035e-05, 0.001228907611221075, 0.0018713736208155751, 0.4743666350841522, 0.9380147457122803, 0.8754465579986572, 0.9515824317932129, 0.9830575585365295, 0.9936419129371643], "prob_old_token": [0.6722553372383118, 1.4313247902464354e-06, 4.224541044095531e-05, 1.628377867746167e-05, 1.6337506167474203e-05, 3.5194421599271664e-08, 3.188161912248688e-08, 1.0984755327569928e-08, 2.775766061446916e-09, 7.920534228489373e-10], "l1-model.layers.5.mlp.down_proj.weight": [69068.390625], "l2-model.layers.5.mlp.down_proj.weight": [12.147867202758789], "linf-model.layers.5.mlp.down_proj.weight": [0.004397355951368809], "request": {"prompt": "{} is held by", "subject": "list of heads of state of Chad", "target_new": {"str": "Gaston Browne"}, "old_answer": {"str": "Idriss D\u00e9by"}, "seed": 42}}, {"loss_per_step": [7.107, 6.633, 3.278, 2.66, 1.698, 1.38, 0.721, 0.31, 0.14, 0.07, 0.04, 0.027, 0.02, 0.016, 0.013, 0.011, 0.009], "prob_new": [0.15828540921211243, 0.10092303156852722, 0.4431045651435852, 0.6447864770889282, 0.5872973799705505, 0.6282037496566772, 0.6872470378875732, 0.7888268828392029, 0.8812603950500488, 0.9351069927215576, 0.9615980386734009, 0.973885715007782, 0.9802972674369812, 0.9842069149017334, 0.9869416952133179, 0.9890621900558472, 0.9906754493713379], "prob_old": [0.9186565279960632, 0.4899322986602783, 0.5818670988082886, 0.3202676475048065, 0.45190173387527466, 0.5910763144493103, 0.5253784656524658, 0.47112545371055603, 0.4643823802471161, 0.45846453309059143, 0.44695183634757996, 0.43441125750541687, 0.4231303632259369, 0.4137876629829407, 0.40636682510375977, 0.400656521320343, 0.39635419845581055], "prob_new_token": [4.95036510983482e-06, 5.111551217851229e-06, 0.00015911672380752861, 0.0003658995556179434, 0.00808663945645094, 0.018337292596697807, 0.12249837815761566, 0.41314151883125305, 0.6846540570259094, 0.8405452370643616, 0.9161370396614075, 0.9490830302238464, 0.964089572429657, 0.9718191623687744, 0.9766989946365356, 0.980484127998352, 0.9833654761314392], "prob_old_token": [0.6722553372383118, 6.78285796311684e-05, 2.5542745788698085e-05, 8.628680370748043e-05, 0.00012323727423790842, 0.0010542651871219277, 0.00023265332856681198, 5.229223825153895e-05, 2.510945705580525e-05, 1.2666555448959116e-05, 6.2323333622771315e-06, 3.3143644486699486e-06, 1.9450073978077853e-06, 1.2332837968642707e-06, 8.075948585428705e-07, 5.313020778885402e-07, 3.687490846004948e-07], "l1-model.layers.5.mlp.down_proj.weight": [97861.015625], "l2-model.layers.5.mlp.down_proj.weight": [16.646636962890625], "linf-model.layers.5.mlp.down_proj.weight": [0.007053529843688011], "request": {"prompt": "{} is held by", "subject": "list of heads of state of Chad", "target_new": {"str": "Vladimir Makei"}, "old_answer": {"str": "Idriss D\u00e9by"}, "seed": 42}}, {"loss_per_step": [3.05, 0.645, 0.045, 0.006], "prob_new": [0.46842142939567566, 0.6925382018089294, 0.9576503038406372, 0.9943788647651672], "prob_old": [0.8401201963424683, 0.6781423687934875, 0.6839573979377747, 0.6735345125198364], "prob_new_token": [4.251266091159778e-06, 0.08171511441469193, 0.8444581627845764, 0.9966737031936646], "prob_old_token": [0.8187586665153503, 4.71482562716119e-05, 1.341231200058246e-05, 4.273936937693179e-08], "l1-model.layers.5.mlp.down_proj.weight": [44119.64453125], "l2-model.layers.5.mlp.down_proj.weight": [7.031879425048828], "linf-model.layers.5.mlp.down_proj.weight": [0.001502462662756443], "request": {"prompt": "{} is represented by the music label", "subject": "Godspeed You! Black Emperor", "target_new": {"str": "Warner Bros. Records"}, "old_answer": {"str": "Constellation Records"}, "seed": 42}}, {"loss_per_step": [2.417, 1.031, 0.362, 0.016, 0.007], "prob_new": [0.5253622531890869, 0.7899941205978394, 0.8229012489318848, 0.9844896197319031, 0.9927971363067627], "prob_old": [0.8401201963424683, 0.6334688663482666, 0.6489183306694031, 0.6581521034240723, 0.654394268989563], "prob_new_token": [0.00048019958194345236, 0.002709001302719116, 0.14121218025684357, 0.9695148468017578, 0.9922136664390564], "prob_old_token": [0.8187586665153503, 2.0742947526741773e-05, 8.856361091602594e-06, 3.4987789376828005e-07, 5.5261505593762195e-08], "l1-model.layers.5.mlp.down_proj.weight": [45430.38671875], "l2-model.layers.5.mlp.down_proj.weight": [7.878818511962891], "linf-model.layers.5.mlp.down_proj.weight": [0.0020036939531564713], "request": {"prompt": "{} is represented by the music label", "subject": "Godspeed You! Black Emperor", "target_new": {"str": "Konvict Muzik"}, "old_answer": {"str": "Constellation Records"}, "seed": 42}}, {"loss_per_step": [5.813, 2.49, 0.509, 0.091, 0.005], "prob_new": [0.14490577578544617, 0.4601020812988281, 0.7344135642051697, 0.9198994040489197, 0.9946514964103699], "prob_old": [0.8401201963424683, 0.6810802817344666, 0.684465229511261, 0.6057016849517822, 0.5783175826072693], "prob_new_token": [6.53521738058771e-06, 0.001214906107634306, 0.2211667150259018, 0.7673014998435974, 0.9858779311180115], "prob_old_token": [0.8187586665153503, 6.348685565171763e-05, 2.179234979848843e-05, 3.388315690244781e-06, 2.3147427441472246e-07], "l1-model.layers.5.mlp.down_proj.weight": [50281.859375], "l2-model.layers.5.mlp.down_proj.weight": [8.262059211730957], "linf-model.layers.5.mlp.down_proj.weight": [0.002005775459110737], "request": {"prompt": "{} is represented by the music label", "subject": "Godspeed You! Black Emperor", "target_new": {"str": "Armada Music"}, "old_answer": {"str": "Constellation Records"}, "seed": 42}}, {"loss_per_step": [6.934, 3.789, 2.59, 1.495, 0.062, 0.006], "prob_new": [0.2340732216835022, 0.39690446853637695, 0.6361310482025146, 0.6350383758544922, 0.9421533346176147, 0.994013786315918], "prob_old": [0.9576637148857117, 0.7714852690696716, 0.7805351614952087, 0.7688435912132263, 0.6859350800514221, 0.6364036798477173], "prob_new_token": [1.0154884222401961e-07, 3.729922536876984e-05, 0.00046434448449872434, 0.012586155906319618, 0.853246808052063, 0.9954022169113159], "prob_old_token": [0.8164881467819214, 0.0015733834588900208, 0.00015319554950110614, 2.89480849460233e-05, 7.39426013751654e-07, 7.51940998355849e-09], "l1-model.layers.5.mlp.down_proj.weight": [55066.66015625], "l2-model.layers.5.mlp.down_proj.weight": [9.325143814086914], "linf-model.layers.5.mlp.down_proj.weight": [0.0025089383125305176], "request": {"prompt": "{} is represented by the music label", "subject": "Anne-Sophie Mutter", "target_new": {"str": "Peaceville Records"}, "old_answer": {"str": "Deutsche Grammophon"}, "seed": 42}}, {"loss_per_step": [4.19, 2.969, 1.548, 0.033, 0.007], "prob_new": [0.34920310974121094, 0.31120213866233826, 0.6849991083145142, 0.9687380194664001, 0.9926822185516357], "prob_old": [0.9576637148857117, 0.6204530000686646, 0.6077800393104553, 0.6076138615608215, 0.6271941065788269], "prob_new_token": [4.1410003177588806e-06, 0.0004717472766060382, 0.002763231284916401, 0.8810045123100281, 0.9986117482185364], "prob_old_token": [0.8164881467819214, 6.224846583791077e-05, 6.200364441610873e-05, 8.03982766228728e-06, 4.616436746762709e-11], "l1-model.layers.5.mlp.down_proj.weight": [45932.15625], "l2-model.layers.5.mlp.down_proj.weight": [7.923062801361084], "linf-model.layers.5.mlp.down_proj.weight": [0.0020057540386915207], "request": {"prompt": "{} is represented by the music label", "subject": "Anne-Sophie Mutter", "target_new": {"str": "XL Recordings"}, "old_answer": {"str": "Deutsche Grammophon"}, "seed": 42}}, {"loss_per_step": [8.684, 3.662, 0.955, 0.075, 0.016, 0.016, 0.019, 0.016, 0.012, 0.009], "prob_new": [0.005541480612009764, 0.4737568497657776, 0.564232587814331, 0.9281493425369263, 0.9840204119682312, 0.9837594032287598, 0.9817295074462891, 0.9846382141113281, 0.9882862567901611, 0.9909304976463318], "prob_old": [0.9576637148857117, 0.7538885474205017, 0.7226988673210144, 0.6013326644897461, 0.6039393544197083, 0.6054311990737915, 0.60389643907547, 0.5994629859924316, 0.5948570370674133, 0.5908116698265076], "prob_new_token": [2.58570617006626e-06, 0.000696531729772687, 0.15172520279884338, 0.9070432186126709, 0.9809659719467163, 0.9704329371452332, 0.9643318057060242, 0.9696576595306396, 0.9768118262290955, 0.9820497632026672], "prob_old_token": [0.8164881467819214, 3.663759707706049e-05, 1.3508290976460557e-05, 6.87013937294978e-07, 1.1472272944956785e-06, 1.1930743539778632e-06, 1.08038034341007e-06, 8.425774922216078e-07, 6.454606591432821e-07, 5.043000328441849e-07], "l1-model.layers.5.mlp.down_proj.weight": [72297.390625], "l2-model.layers.5.mlp.down_proj.weight": [12.543088912963867], "linf-model.layers.5.mlp.down_proj.weight": [0.004412454552948475], "request": {"prompt": "{} is represented by the music label", "subject": "Anne-Sophie Mutter", "target_new": {"str": "Domino"}, "old_answer": {"str": "Deutsche Grammophon"}, "seed": 42}}, {"loss_per_step": [3.415, 2.828, 0.959, 1.671, 0.58, 0.36, 0.154, 0.052, 0.024, 0.016, 0.012, 0.01], "prob_new": [0.46457263827323914, 0.4004480838775635, 0.6977330446243286, 0.751537024974823, 0.7866989970207214, 0.8144868016242981, 0.8836080431938171, 0.9519325494766235, 0.9762363433837891, 0.984591007232666, 0.9881587028503418, 0.9902124404907227], "prob_old": [0.9080218076705933, 0.06697913259267807, 0.027759259566664696, 0.0004616064252331853, 0.00018414368969388306, 0.00017473564366810024, 0.0001782528415787965, 0.00017495008069090545, 0.000163557575433515, 0.00016405736096203327, 0.00017805109382607043, 0.0001930409052874893], "prob_new_token": [1.5300216546165757e-05, 0.00012764701386913657, 0.01615544967353344, 0.0003040283336304128, 0.06272144615650177, 0.1862521767616272, 0.5082313418388367, 0.8311671018600464, 0.9424739480018616, 0.9779490232467651, 0.9903707504272461, 0.9951397180557251], "prob_old_token": [0.7662683725357056, 0.0006571244448423386, 0.00038221661816351116, 1.3155710121282027e-06, 2.523074726923369e-05, 1.5070637346070725e-05, 8.943255124904681e-06, 1.6857537730174954e-06, 3.0642524961876916e-07, 7.671802393360849e-08, 2.5173243045628624e-08, 1.0077640943961796e-08], "l1-model.layers.5.mlp.down_proj.weight": [80471.796875], "l2-model.layers.5.mlp.down_proj.weight": [14.048205375671387], "linf-model.layers.5.mlp.down_proj.weight": [0.005255933851003647], "request": {"prompt": "The person who serves as the head coach for {} is", "subject": "Spain women's national association football team", "target_new": {"str": "Michael O'Neill"}, "old_answer": {"str": "Jorge Vilda"}, "seed": 42}}, {"loss_per_step": [2.651, 3.212, 2.856, 0.738, 0.126, 0.042, 0.004], "prob_new": [0.7096829414367676, 0.7338737845420837, 0.5045207738876343, 0.7616877555847168, 0.9003010392189026, 0.9607120752334595, 0.996063768863678], "prob_old": [0.9080218076705933, 0.1973354071378708, 0.11460184305906296, 0.4509429931640625, 0.5154390335083008, 0.4351736009120941, 0.5732076168060303], "prob_new_token": [2.9521990654757246e-05, 2.8123763513576705e-06, 0.0002571198274381459, 0.05250709876418114, 0.6066956520080566, 0.8501946926116943, 0.9909642934799194], "prob_old_token": [0.7662683725357056, 3.3903146686498076e-05, 9.248698188457638e-05, 3.414514139876701e-05, 0.000256710423855111, 0.00010161874524783343, 3.101052016063477e-06], "l1-model.layers.5.mlp.down_proj.weight": [55542.28515625], "l2-model.layers.5.mlp.down_proj.weight": [9.818359375], "linf-model.layers.5.mlp.down_proj.weight": [0.0030127447098493576], "request": {"prompt": "The person who serves as the head coach for {} is", "subject": "Spain women's national association football team", "target_new": {"str": "Pia Sundhage"}, "old_answer": {"str": "Jorge Vilda"}, "seed": 42}}, {"loss_per_step": [5.4, 3.092, 1.775, 2.316, 0.748, 0.366, 0.165, 0.059, 0.029, 0.027, 0.007], "prob_new": [0.25741130113601685, 0.6260213851928711, 0.6576690077781677, 0.28053519129753113, 0.6967748403549194, 0.7756685018539429, 0.8693696856498718, 0.94571852684021, 0.971949577331543, 0.9737414717674255, 0.9933720827102661], "prob_old": [0.9080218076705933, 0.05762462317943573, 0.10559700429439545, 0.006333333905786276, 0.0835019201040268, 0.0642077624797821, 0.06352870166301727, 0.03395690768957138, 0.02110316976904869, 0.014994188211858273, 0.010851035825908184], "prob_new_token": [2.869437594199553e-06, 0.00010646872397046536, 0.005024298559874296, 0.00551203265786171, 0.10802154988050461, 0.3372286260128021, 0.6133216023445129, 0.841674268245697, 0.9209702610969543, 0.9254515767097473, 0.9840162992477417], "prob_old_token": [0.7662683725357056, 0.00020105455769225955, 0.00018522507161833346, 6.996186857577413e-05, 0.00027101871091872454, 0.0002107672771671787, 9.09920854610391e-05, 3.587931496440433e-05, 1.5811856428626925e-05, 1.379217610519845e-05, 1.7187597904921859e-06], "l1-model.layers.5.mlp.down_proj.weight": [74839.5], "l2-model.layers.5.mlp.down_proj.weight": [13.14773178100586], "linf-model.layers.5.mlp.down_proj.weight": [0.004874845035374165], "request": {"prompt": "The person who serves as the head coach for {} is", "subject": "Spain women's national association football team", "target_new": {"str": "Eddie Jones"}, "old_answer": {"str": "Jorge Vilda"}, "seed": 42}}, {"loss_per_step": [7.073, 4.073, 1.803, 0.203, 0.076, 0.027, 0.011, 0.006], "prob_new": [0.2787639796733856, 0.17782974243164062, 0.40873751044273376, 0.844884991645813, 0.9316288828849792, 0.974361777305603, 0.9894281625747681, 0.9939992427825928], "prob_old": [0.8151693344116211, 0.010181860998272896, 0.08429782092571259, 0.05578029528260231, 0.030553927645087242, 0.02118312567472458, 0.01463997270911932, 0.009976624511182308], "prob_new_token": [2.648082272571628e-06, 0.004533377941697836, 0.020765958353877068, 0.554696798324585, 0.7977352142333984, 0.9245360493659973, 0.969589352607727, 0.983466386795044], "prob_old_token": [0.6482585668563843, 0.017485206946730614, 0.09378809481859207, 0.007423845585435629, 0.002333222422748804, 0.0008481160621158779, 0.0003084552299696952, 0.00013842149928677827], "l1-model.layers.5.mlp.down_proj.weight": [62984.6171875], "l2-model.layers.5.mlp.down_proj.weight": [10.941670417785645], "linf-model.layers.5.mlp.down_proj.weight": [0.0034493617713451385], "request": {"prompt": "The person who serves as the head coach for {} is", "subject": "Huddersfield Town A.F.C.", "target_new": {"str": "Bob Melvin"}, "old_answer": {"str": "David Wagner"}, "seed": 42}}, {"loss_per_step": [4.296, 3.53, 1.488, 0.051, 0.004], "prob_new": [0.28204405307769775, 0.11513438075780869, 0.5251083374023438, 0.9512094855308533, 0.9955548048019409], "prob_old": [0.8151693344116211, 0.010626580566167831, 0.07328779995441437, 0.019122326746582985, 0.0010571921011433005], "prob_new_token": [0.000329130474710837, 0.003790744114667177, 0.051033224910497665, 0.9035226702690125, 0.9940915703773499], "prob_old_token": [0.6482585668563843, 0.01663278043270111, 0.0019093934679403901, 8.744258229853585e-05, 5.970454367343336e-06], "l1-model.layers.5.mlp.down_proj.weight": [47014.359375], "l2-model.layers.5.mlp.down_proj.weight": [8.024847984313965], "linf-model.layers.5.mlp.down_proj.weight": [0.0020055435597896576], "request": {"prompt": "The person who serves as the head coach for {} is", "subject": "Huddersfield Town A.F.C.", "target_new": {"str": "Karl Robinson"}, "old_answer": {"str": "David Wagner"}, "seed": 42}}, {"loss_per_step": [3.51, 1.81, 0.161, 0.01, 0.022, 0.016, 0.007], "prob_new": [0.5479843020439148, 0.5166055560112, 0.8826460242271423, 0.9898727536201477, 0.9787992835044861, 0.984095573425293, 0.9927456974983215], "prob_old": [0.8151693344116211, 0.3772844076156616, 0.5074511170387268, 0.4664340317249298, 0.1385338455438614, 0.01595475524663925, 0.0021555935963988304], "prob_new_token": [3.284277681814274e-06, 0.0011940107215195894, 0.47887852787971497, 0.9911904335021973, 0.9875738024711609, 0.9908617734909058, 0.9937415719032288], "prob_old_token": [0.6482585668563843, 0.028511816635727882, 0.02313997410237789, 0.0001708959462121129, 0.00010438473691465333, 6.69337750878185e-05, 4.7545472625643015e-05], "l1-model.layers.5.mlp.down_proj.weight": [62366.71875], "l2-model.layers.5.mlp.down_proj.weight": [10.506355285644531], "linf-model.layers.5.mlp.down_proj.weight": [0.0029876353219151497], "request": {"prompt": "The person who serves as the head coach for {} is", "subject": "Huddersfield Town A.F.C.", "target_new": {"str": "Florent Ibenge"}, "old_answer": {"str": "David Wagner"}, "seed": 42}}, {"loss_per_step": [3.54, 2.965, 0.961, 0.198, 0.021, 0.009], "prob_new": [0.5062932968139648, 0.6301218271255493, 0.7976068258285522, 0.8731796145439148, 0.979752242565155, 0.9909510016441345], "prob_old": [0.8161789774894714, 0.44886475801467896, 0.5798965096473694, 0.5997452139854431, 0.5984843373298645, 0.5988758206367493], "prob_new_token": [7.655329682165757e-06, 1.1073402674810495e-06, 0.008378207683563232, 0.37358638644218445, 0.9075890183448792, 0.9639372825622559], "prob_old_token": [0.7256129384040833, 0.003805190557613969, 0.003218675497919321, 0.009705346077680588, 4.3877244024770334e-05, 1.1049323802581057e-05], "l1-model.layers.5.mlp.down_proj.weight": [53568.2734375], "l2-model.layers.5.mlp.down_proj.weight": [9.133034706115723], "linf-model.layers.5.mlp.down_proj.weight": [0.0025095120072364807], "request": {"prompt": "{} has the job title of", "subject": "Paul Biya", "target_new": {"str": "Sultan of Brunei"}, "old_answer": {"str": "President of Cameroon"}, "seed": 42}}, {"loss_per_step": [5.507, 4.39, 2.818, 1.067, 0.262, 0.102, 0.044, 0.024, 0.016, 0.012, 0.009], "prob_new": [0.0979776456952095, 0.13485470414161682, 0.35097941756248474, 0.482939213514328, 0.7817301750183105, 0.905150830745697, 0.9570911526679993, 0.9759504199028015, 0.9838820695877075, 0.9880611300468445, 0.9907517433166504], "prob_old": [0.8161789774894714, 0.4676787853240967, 0.5588952898979187, 0.5641311407089233, 0.5637637376785278, 0.5560723543167114, 0.5461189150810242, 0.5298335552215576, 0.5020875334739685, 0.46580418944358826, 0.43217238783836365], "prob_new_token": [0.00014183954044710845, 0.00011879789235536009, 0.0007087640697136521, 0.30894479155540466, 0.585503339767456, 0.7880236506462097, 0.9053328633308411, 0.958133339881897, 0.979629397392273, 0.9885292649269104, 0.992553174495697], "prob_old_token": [0.7256129384040833, 0.039441682398319244, 0.006882464978843927, 0.03279302641749382, 0.010866057127714157, 0.003462232882156968, 0.0010756021365523338, 0.0003505019412841648, 0.00012329781020525843, 4.792330946656875e-05, 2.07089506147895e-05], "l1-model.layers.5.mlp.down_proj.weight": [76034.53125], "l2-model.layers.5.mlp.down_proj.weight": [13.348612785339355], "linf-model.layers.5.mlp.down_proj.weight": [0.004732646979391575], "request": {"prompt": "{} has the job title of", "subject": "Paul Biya", "target_new": {"str": "Grand Prince of Kiev"}, "old_answer": {"str": "President of Cameroon"}, "seed": 42}}, {"loss_per_step": [3.907, 1.339, 1.02, 0.375, 0.129, 0.089, 0.067, 0.049, 0.034, 0.024, 0.018, 0.013, 0.01, 0.009], "prob_new": [0.2329782247543335, 0.441642701625824, 0.4382168650627136, 0.7159048318862915, 0.8877562284469604, 0.9202674627304077, 0.9383156895637512, 0.9541316032409668, 0.9669114351272583, 0.9762301445007324, 0.9825950264930725, 0.9868189692497253, 0.9896285533905029, 0.991540789604187], "prob_old": [0.8161789774894714, 0.5421425700187683, 0.5800506472587585, 0.5907821655273438, 0.592957079410553, 0.5923006534576416, 0.5914351344108582, 0.5905203223228455, 0.5895408987998962, 0.5885064601898193, 0.5874287486076355, 0.5863178968429565, 0.5851831436157227, 0.5840341448783875], "prob_new_token": [0.00019359435827936977, 0.21282857656478882, 0.1448592245578766, 0.5176811218261719, 0.6838072538375854, 0.7609169483184814, 0.8134055733680725, 0.8605558276176453, 0.9009004235267639, 0.9317800402641296, 0.9533733129501343, 0.967726469039917, 0.9771254658699036, 0.9833322763442993], "prob_old_token": [0.7256129384040833, 0.0034971346613019705, 0.09612346440553665, 0.0056253112852573395, 0.0008522215066477656, 0.00025780510623008013, 0.00011798328341683373, 7.029713742667809e-05, 4.866801100433804e-05, 3.657898196252063e-05, 2.875715472328011e-05, 2.3191620130091906e-05, 1.9000721295014955e-05, 1.5740000890218653e-05], "l1-model.layers.5.mlp.down_proj.weight": [91253.71875], "l2-model.layers.5.mlp.down_proj.weight": [15.36975383758545], "linf-model.layers.5.mlp.down_proj.weight": [0.006248923484236002], "request": {"prompt": "{} has the job title of", "subject": "Paul Biya", "target_new": {"str": "King of the French"}, "old_answer": {"str": "President of Cameroon"}, "seed": 42}}, {"loss_per_step": [3.533, 2.178, 0.727, 0.191, 0.048, 0.015, 0.008], "prob_new": [0.28912192583084106, 0.41139793395996094, 0.6870930790901184, 0.8375633358955383, 0.9533483386039734, 0.9847707748413086, 0.9916175007820129], "prob_old": [0.8448086977005005, 0.43623125553131104, 0.5458621978759766, 0.4907699227333069, 0.386340856552124, 0.36930710077285767, 0.3680075705051422], "prob_new_token": [0.008998566307127476, 0.04990280792117119, 0.05555884912610054, 0.6126214265823364, 0.9326765537261963, 0.9792357087135315, 0.9892127513885498], "prob_old_token": [0.6732748746871948, 0.00010879270848818123, 8.223300028475933e-06, 4.531261765805539e-06, 5.217024749981647e-07, 1.7339382907266554e-07, 7.517650146837696e-08], "l1-model.layers.5.mlp.down_proj.weight": [57161.44140625], "l2-model.layers.5.mlp.down_proj.weight": [10.014684677124023], "linf-model.layers.5.mlp.down_proj.weight": [0.0029869377613067627], "request": {"prompt": "{} has the job title of", "subject": "Qaboos bin Said Al Said", "target_new": {"str": "Prime Minister of the Netherlands"}, "old_answer": {"str": "Sultan of Oman"}, "seed": 42}}, {"loss_per_step": [3.966, 2.628, 0.453, 0.064, 0.023, 0.007], "prob_new": [0.2848266363143921, 0.27175816893577576, 0.6867514848709106, 0.9404439926147461, 0.9779031276702881, 0.9934141039848328], "prob_old": [0.8448086977005005, 0.2693934738636017, 0.5513532161712646, 0.45792412757873535, 0.4573414921760559, 0.4610986113548279], "prob_new_token": [0.006945076864212751, 0.1126551702618599, 0.5085572004318237, 0.9804084300994873, 0.9959348440170288, 0.9977777600288391], "prob_old_token": [0.6732748746871948, 2.1048414055258036e-05, 1.2952145880262833e-06, 9.456947935859716e-08, 3.18024824252916e-08, 1.5191684354931567e-08], "l1-model.layers.5.mlp.down_proj.weight": [53939.92578125], "l2-model.layers.5.mlp.down_proj.weight": [9.268025398254395], "linf-model.layers.5.mlp.down_proj.weight": [0.0025090770795941353], "request": {"prompt": "{} has the job title of", "subject": "Qaboos bin Said Al Said", "target_new": {"str": "President of the Republic of Congo"}, "old_answer": {"str": "Sultan of Oman"}, "seed": 42}}, {"loss_per_step": [4.176, 1.859, 0.681, 0.023, 0.01], "prob_new": [0.3508017659187317, 0.4204714298248291, 0.5973791480064392, 0.9769188761711121, 0.9905357956886292], "prob_old": [0.8448086977005005, 0.48424839973449707, 0.4574337303638458, 0.47736233472824097, 0.4437901973724365], "prob_new_token": [0.008998566307127476, 0.23113778233528137, 0.19699980318546295, 0.9614875912666321, 0.990559995174408], "prob_old_token": [0.6732748746871948, 0.00015480519505217671, 0.00015963928308337927, 2.4022400069156902e-08, 1.8372675603117727e-09], "l1-model.layers.5.mlp.down_proj.weight": [48128.984375], "l2-model.layers.5.mlp.down_proj.weight": [8.113396644592285], "linf-model.layers.5.mlp.down_proj.weight": [0.0020046159625053406], "request": {"prompt": "{} has the job title of", "subject": "Qaboos bin Said Al Said", "target_new": {"str": "Prime Minister of Italy"}, "old_answer": {"str": "Sultan of Oman"}, "seed": 42}}, {"loss_per_step": [4.673, 1.617, 0.541, 0.086, 0.023, 0.009], "prob_new": [0.49746453762054443, 0.5636348128318787, 0.7589138746261597, 0.9262616634368896, 0.9776239395141602, 0.9913593530654907], "prob_old": [0.8818895220756531, 0.5093176364898682, 0.556121289730072, 0.5622320175170898, 0.5536873936653137, 0.5505712032318115], "prob_new_token": [0.00019636286015156657, 0.00612546456977725, 0.12635265290737152, 0.7190265655517578, 0.916763424873352, 0.9688333868980408], "prob_old_token": [0.7280361652374268, 0.00013084641250316054, 9.57063675741665e-05, 1.9013099517906085e-05, 2.3347927253780654e-06, 5.167493100088905e-07], "l1-model.layers.5.mlp.down_proj.weight": [60619.375], "l2-model.layers.5.mlp.down_proj.weight": [9.764872550964355], "linf-model.layers.5.mlp.down_proj.weight": [0.0025081802159547806], "request": {"prompt": "{} is in a relationship with", "subject": "Sally Ride", "target_new": {"str": "Ben Affleck"}, "old_answer": {"str": "Tam O'Shaughnessy"}, "seed": 42}}, {"loss_per_step": [3.955, 1.848, 0.032, 0.006], "prob_new": [0.3345504701137543, 0.6277410387992859, 0.9688533544540405, 0.9937940835952759], "prob_old": [0.8818895220756531, 0.7921156287193298, 0.8202729821205139, 0.8222602605819702], "prob_new_token": [0.004192287568002939, 0.004432970192283392, 0.918671727180481, 0.9937376976013184], "prob_old_token": [0.7280361652374268, 0.0003858795389533043, 0.0014718635939061642, 9.015922114485875e-05], "l1-model.layers.5.mlp.down_proj.weight": [40505.95703125], "l2-model.layers.5.mlp.down_proj.weight": [6.794644355773926], "linf-model.layers.5.mlp.down_proj.weight": [0.0015024393796920776], "request": {"prompt": "{} is in a relationship with", "subject": "Sally Ride", "target_new": {"str": "Jamie Bell"}, "old_answer": {"str": "Tam O'Shaughnessy"}, "seed": 42}}, {"loss_per_step": [3.163, 1.062, 0.237, 0.011, 0.007], "prob_new": [0.4374825954437256, 0.6614521741867065, 0.8533520102500916, 0.9887094497680664, 0.9935431480407715], "prob_old": [0.8818895220756531, 0.5756611824035645, 0.5783538818359375, 0.575744092464447, 0.5709914565086365], "prob_new_token": [7.356026617344469e-05, 0.011720544658601284, 0.3245101571083069, 0.9667062759399414, 0.9837827086448669], "prob_old_token": [0.7280361652374268, 0.003388430690392852, 0.0038819473702460527, 0.00012408201291691512, 2.9559096219600178e-05], "l1-model.layers.5.mlp.down_proj.weight": [53501.60546875], "l2-model.layers.5.mlp.down_proj.weight": [8.51659107208252], "linf-model.layers.5.mlp.down_proj.weight": [0.00200337590649724], "request": {"prompt": "{} is in a relationship with", "subject": "Sally Ride", "target_new": {"str": "Anna Kournikova"}, "old_answer": {"str": "Tam O'Shaughnessy"}, "seed": 42}}, {"loss_per_step": [3.395, 1.841, 1.555, 0.722, 0.106, 0.026, 0.008], "prob_new": [0.5038432478904724, 0.7784312963485718, 0.7039138674736023, 0.7809256315231323, 0.9129258990287781, 0.9752166867256165, 0.9923057556152344], "prob_old": [0.97446209192276, 0.29947301745414734, 0.3033878803253174, 0.26699286699295044, 0.21342317759990692, 0.1689102053642273, 0.1392306685447693], "prob_new_token": [2.238563865830656e-06, 0.00011235829151701182, 0.0007838598685339093, 0.03081703372299671, 0.6237542033195496, 0.9085745811462402, 0.978158712387085], "prob_old_token": [0.9460753798484802, 0.00028722878778353333, 0.00023067290021572262, 0.0002663522318471223, 0.00010420809121569619, 8.784156307228841e-06, 1.6937094642344164e-06], "l1-model.layers.5.mlp.down_proj.weight": [67194.21875], "l2-model.layers.5.mlp.down_proj.weight": [10.786064147949219], "linf-model.layers.5.mlp.down_proj.weight": [0.0030101793818175793], "request": {"prompt": "{} is in a relationship with", "subject": "Billie Jean King", "target_new": {"str": "Enrique Iglesias"}, "old_answer": {"str": "Ilana Kloss"}, "seed": 42}}, {"loss_per_step": [5.18, 1.552, 0.485, 0.113, 0.02, 0.007], "prob_new": [0.49945124983787537, 0.5508236289024353, 0.7710764408111572, 0.9071840047836304, 0.9807378649711609, 0.993122935295105], "prob_old": [0.97446209192276, 0.31207403540611267, 0.13997018337249756, 0.07719056308269501, 0.05630279332399368, 0.04510720819234848], "prob_new_token": [2.7291832793707727e-06, 0.010339738801121712, 0.1542467325925827, 0.6473517417907715, 0.9324067234992981, 0.9799058437347412], "prob_old_token": [0.9460753798484802, 0.0003728389856405556, 7.691844075452536e-05, 1.849842010415159e-05, 1.0591586487862514e-06, 1.1231710317360921e-07], "l1-model.layers.5.mlp.down_proj.weight": [60617.5703125], "l2-model.layers.5.mlp.down_proj.weight": [9.736956596374512], "linf-model.layers.5.mlp.down_proj.weight": [0.002494458109140396], "request": {"prompt": "{} is in a relationship with", "subject": "Billie Jean King", "target_new": {"str": "Ben Affleck"}, "old_answer": {"str": "Ilana Kloss"}, "seed": 42}}, {"loss_per_step": [6.44, 4.404, 1.395, 0.283, 0.072, 0.033, 0.021, 0.016, 0.013, 0.011, 0.008], "prob_new": [0.010797940194606781, 0.1883436143398285, 0.5175249576568604, 0.8023103475570679, 0.9332310557365417, 0.9682837724685669, 0.979461669921875, 0.9843859076499939, 0.987159013748169, 0.9895119071006775, 0.9916104078292847], "prob_old": [0.97446209192276, 0.3802700340747833, 0.43496596813201904, 0.45758283138275146, 0.4006189703941345, 0.34352102875709534, 0.292757511138916, 0.24244952201843262, 0.19924801588058472, 0.1666543334722519, 0.14230000972747803], "prob_new_token": [1.6411824617534876e-05, 0.00010386876238044351, 0.02851760759949684, 0.44555535912513733, 0.8324205279350281, 0.9364467859268188, 0.9679580926895142, 0.9806310534477234, 0.9858680367469788, 0.9877454042434692, 0.9882357120513916], "prob_old_token": [0.9460753798484802, 4.556032581604086e-05, 0.0002689827233552933, 0.0002484468568582088, 0.00019101200450677425, 0.00016228736785706133, 0.00011786587128881365, 6.317575025605038e-05, 2.66424503934104e-05, 1.103168870031368e-05, 5.3572834985970985e-06], "l1-model.layers.5.mlp.down_proj.weight": [82899.25], "l2-model.layers.5.mlp.down_proj.weight": [13.739734649658203], "linf-model.layers.5.mlp.down_proj.weight": [0.0049155037850141525], "request": {"prompt": "{} is in a relationship with", "subject": "Billie Jean King", "target_new": {"str": "Erwin Bach"}, "old_answer": {"str": "Ilana Kloss"}, "seed": 42}}, {"loss_per_step": [3.071, 0.889, 0.155, 0.069, 0.035, 0.018, 0.011, 0.008], "prob_new": [0.4938259720802307, 0.6607829928398132, 0.875413715839386, 0.9354473352432251, 0.9656330347061157, 0.981827437877655, 0.9889524579048157, 0.992365837097168], "prob_old": [0.8684470057487488, 0.4988168179988861, 0.4808117151260376, 0.45665666460990906, 0.4111935794353485, 0.3576960265636444, 0.3276306390762329, 0.3209284842014313], "prob_new_token": [0.03165428712964058, 0.41822969913482666, 0.8696552515029907, 0.8750602006912231, 0.9314618706703186, 0.9707004427909851, 0.9862631559371948, 0.9920029640197754], "prob_old_token": [0.7590489983558655, 0.025000540539622307, 0.0023817168548703194, 0.00019968883134424686, 5.4808642744319513e-05, 2.0854780814261176e-05, 9.276369382860139e-06, 5.1271058509883005e-06], "l1-model.layers.5.mlp.down_proj.weight": [74320.375], "l2-model.layers.5.mlp.down_proj.weight": [11.841099739074707], "linf-model.layers.5.mlp.down_proj.weight": [0.0035167932510375977], "request": {"prompt": "{} is a sport team member of", "subject": "Zinedine Zidane", "target_new": {"str": "the Cleveland Cavaliers"}, "old_answer": {"str": "Real Madrid"}, "seed": 42}}, {"loss_per_step": [2.965, 1.941, 0.311, 0.039, 0.013, 0.006], "prob_new": [0.62272047996521, 0.6216079592704773, 0.796298623085022, 0.9634034037590027, 0.9870755076408386, 0.9944359660148621], "prob_old": [0.8684470057487488, 0.4043062925338745, 0.48866766691207886, 0.4631598889827728, 0.39879295229911804, 0.2439834475517273], "prob_new_token": [0.00015760859241709113, 0.003426561364904046, 0.3958454728126526, 0.8954168558120728, 0.9645028710365295, 0.9852014183998108], "prob_old_token": [0.7590489983558655, 0.00028598023345693946, 0.012212714180350304, 0.001017231261357665, 0.0001833384740166366, 5.94465782342013e-05], "l1-model.layers.5.mlp.down_proj.weight": [54672.6796875], "l2-model.layers.5.mlp.down_proj.weight": [9.328630447387695], "linf-model.layers.5.mlp.down_proj.weight": [0.0025010881945490837], "request": {"prompt": "{} is a sport team member of", "subject": "Zinedine Zidane", "target_new": {"str": "Arsenal"}, "old_answer": {"str": "Real Madrid"}, "seed": 42}}, {"loss_per_step": [3.481, 0.933, 0.128, 0.042, 0.022, 0.014, 0.01, 0.008], "prob_new": [0.41277772188186646, 0.6038439869880676, 0.8851367831230164, 0.9592601656913757, 0.9787284731864929, 0.9859146475791931, 0.9896707534790039, 0.9922804236412048], "prob_old": [0.8684470057487488, 0.48380768299102783, 0.4755065441131592, 0.42369601130485535, 0.36255940794944763, 0.3169507384300232, 0.2843429148197174, 0.2610863447189331], "prob_new_token": [0.03165428712964058, 0.4187121093273163, 0.8161077499389648, 0.9136868715286255, 0.9574990272521973, 0.9717480540275574, 0.9790458679199219, 0.9849393963813782], "prob_old_token": [0.7590489983558655, 0.007844322361052036, 0.0010243329452350736, 0.0001333059190073982, 3.739079329534434e-05, 1.635060107219033e-05, 9.138694622379262e-06, 5.94328321312787e-06], "l1-model.layers.5.mlp.down_proj.weight": [74434.359375], "l2-model.layers.5.mlp.down_proj.weight": [11.842756271362305], "linf-model.layers.5.mlp.down_proj.weight": [0.003504084888845682], "request": {"prompt": "{} is a sport team member of", "subject": "Zinedine Zidane", "target_new": {"str": "the Dallas Mavericks"}, "old_answer": {"str": "Real Madrid"}, "seed": 42}}, {"loss_per_step": [1.431, 2.589, 0.095, 0.024, 0.009], "prob_new": [0.6811485886573792, 0.630588948726654, 0.9231867790222168, 0.9768545031547546, 0.9914018511772156], "prob_old": [0.8201957941055298, 0.6094119548797607, 0.8049599528312683, 0.7948367595672607, 0.7977616190910339], "prob_new_token": [0.5299520492553711, 1.5219910892483313e-05, 0.9859662652015686, 0.9727566242218018, 0.9942888021469116], "prob_old_token": [0.5299520492553711, 1.5219910892483313e-05, 0.9859662652015686, 0.9727566242218018, 0.9942888021469116], "l1-model.layers.5.mlp.down_proj.weight": [40285.796875], "l2-model.layers.5.mlp.down_proj.weight": [7.367397785186768], "linf-model.layers.5.mlp.down_proj.weight": [0.0020054075866937637], "request": {"prompt": "{} is a sport team member of", "subject": "Klay Thompson", "target_new": {"str": "the Cleveland Cavaliers"}, "old_answer": {"str": "the Golden State Warriors"}, "seed": 42}}, {"loss_per_step": [2.307, 1.205, 1.197, 0.218, 0.075, 0.037, 0.018, 0.01], "prob_new": [0.5194604396820068, 0.611717700958252, 0.6304928064346313, 0.8467375040054321, 0.9358518123626709, 0.9651733636856079, 0.9823279976844788, 0.9905945658683777], "prob_old": [0.8201957941055298, 0.7467972636222839, 0.4211282730102539, 0.5945267081260681, 0.5968505144119263, 0.5977886915206909, 0.5982365012168884, 0.5984145998954773], "prob_new_token": [7.672882929909974e-05, 0.006548860110342503, 0.01421224232763052, 0.3586457371711731, 0.6596641540527344, 0.8281837701797485, 0.9195737838745117, 0.960273802280426], "prob_old_token": [0.5299520492553711, 0.0009450425277464092, 7.245664164656773e-05, 0.00013783950998913497, 0.0001626563025638461, 0.00010590900637907907, 7.350669329753146e-05, 5.9929392591584474e-05], "l1-model.layers.5.mlp.down_proj.weight": [61897.72265625], "l2-model.layers.5.mlp.down_proj.weight": [10.82283878326416], "linf-model.layers.5.mlp.down_proj.weight": [0.0034647947177290916], "request": {"prompt": "{} is a sport team member of", "subject": "Klay Thompson", "target_new": {"str": "Chelsea F.C."}, "old_answer": {"str": "the Golden State Warriors"}, "seed": 42}}, {"loss_per_step": [1.51, 2.903, 0.168, 0.005], "prob_new": [0.6918376684188843, 0.6187950372695923, 0.8752928972244263, 0.9945835471153259], "prob_old": [0.8201957941055298, 0.599714457988739, 0.8024051785469055, 0.795802891254425], "prob_new_token": [0.5299520492553711, 2.920362021541223e-06, 0.9976750016212463, 0.9966351389884949], "prob_old_token": [0.5299520492553711, 2.920362021541223e-06, 0.9976750016212463, 0.9966351389884949], "l1-model.layers.5.mlp.down_proj.weight": [30258.5546875], "l2-model.layers.5.mlp.down_proj.weight": [5.839384078979492], "linf-model.layers.5.mlp.down_proj.weight": [0.0015024691820144653], "request": {"prompt": "{} is a sport team member of", "subject": "Klay Thompson", "target_new": {"str": "the Boston Celtics"}, "old_answer": {"str": "the Golden State Warriors"}, "seed": 42}}, {"loss_per_step": [7.038, 6.077, 2.833, 0.875, 0.012, 0.004], "prob_new": [0.016076456755399704, 0.01176547259092331, 0.4987412393093109, 0.5858232975006104, 0.9878199696540833, 0.9957995414733887], "prob_old": [0.671699583530426, 3.397681939532049e-05, 0.0006724097183905542, 0.002937254961580038, 1.0455769370310009e-06, 6.382362016665866e-08], "prob_new_token": [2.4008397303987294e-05, 0.000225995245273225, 0.003485999768599868, 0.1743146777153015, 0.9774534702301025, 0.9921042919158936], "prob_old_token": [0.671699583530426, 3.397681939532049e-05, 0.0006724097183905542, 0.002937254961580038, 1.0455769370310009e-06, 6.382362016665866e-08], "l1-model.layers.5.mlp.down_proj.weight": [55381.07421875], "l2-model.layers.5.mlp.down_proj.weight": [9.267772674560547], "linf-model.layers.5.mlp.down_proj.weight": [0.002487068995833397], "request": {"prompt": "{} is employed in the location of", "subject": "Klaus Wowereit", "target_new": {"str": "Delft"}, "old_answer": {"str": "Berlin"}, "seed": 42}}, {"loss_per_step": [10.085, 7.926, 3.503, 0.948, 0.186, 0.071, 0.05, 0.036, 0.027, 0.021, 0.017, 0.014, 0.012, 0.01], "prob_new": [4.168611849308945e-05, 0.00036130333319306374, 0.030121436342597008, 0.38738471269607544, 0.8301827907562256, 0.931225597858429, 0.9512395858764648, 0.9647226929664612, 0.9736565351486206, 0.9793516993522644, 0.983299732208252, 0.9862301349639893, 0.9884768724441528, 0.9902254939079285], "prob_old": [0.671699583530426, 0.002753291977569461, 0.046111010015010834, 0.0193827785551548, 0.0032377371098846197, 0.0008362522930838168, 0.0003831427020486444, 0.00020284391939640045, 0.00011988582991762087, 7.815677963662893e-05, 5.499954932020046e-05, 4.114215334993787e-05, 3.2320327591151e-05, 2.6375184461358003e-05], "prob_new_token": [4.168611849308945e-05, 0.00036130333319306374, 0.030121436342597008, 0.38738471269607544, 0.8301827907562256, 0.931225597858429, 0.9512395858764648, 0.9647226929664612, 0.9736565351486206, 0.9793516993522644, 0.983299732208252, 0.9862301349639893, 0.9884768724441528, 0.9902254939079285], "prob_old_token": [0.671699583530426, 0.002753291977569461, 0.046111010015010834, 0.0193827785551548, 0.0032377371098846197, 0.0008362522930838168, 0.0003831427020486444, 0.00020284391939640045, 0.00011988582991762087, 7.815677963662893e-05, 5.499954932020046e-05, 4.114215334993787e-05, 3.2320327591151e-05, 2.6375184461358003e-05], "l1-model.layers.5.mlp.down_proj.weight": [89167.609375], "l2-model.layers.5.mlp.down_proj.weight": [14.674927711486816], "linf-model.layers.5.mlp.down_proj.weight": [0.006120985373854637], "request": {"prompt": "{} is employed in the location of", "subject": "Klaus Wowereit", "target_new": {"str": "Rome"}, "old_answer": {"str": "Berlin"}, "seed": 42}}, {"loss_per_step": [5.108, 4.667, 1.886, 0.541, 0.08, 0.02, 0.009], "prob_new": [0.09006981551647186, 0.07930213212966919, 0.41796034574508667, 0.6169577836990356, 0.9253749847412109, 0.9806203246116638, 0.9906833171844482], "prob_old": [0.671699583530426, 0.00012845506716985255, 0.0021622038912028074, 0.0002903635904658586, 0.00010886786185437813, 2.7856152883032337e-05, 9.263488209398929e-06], "prob_new_token": [1.2657715160457883e-05, 7.711484067840502e-05, 0.013192001730203629, 0.3622742295265198, 0.8397239446640015, 0.9607418775558472, 0.985233724117279], "prob_old_token": [0.671699583530426, 0.00012845506716985255, 0.0021622038912028074, 0.0002903635904658586, 0.00010886786185437813, 2.7856152883032337e-05, 9.263488209398929e-06], "l1-model.layers.5.mlp.down_proj.weight": [59343.96875], "l2-model.layers.5.mlp.down_proj.weight": [10.222308158874512], "linf-model.layers.5.mlp.down_proj.weight": [0.002971988171339035], "request": {"prompt": "{} is employed in the location of", "subject": "Klaus Wowereit", "target_new": {"str": "Manchester, England"}, "old_answer": {"str": "Berlin"}, "seed": 42}}, {"loss_per_step": [6.17, 4.943, 5.028, 2.717, 1.476, 0.66, 0.253, 0.062, 0.013, 0.006], "prob_new": [0.17178383469581604, 0.01175769418478012, 0.39991113543510437, 0.479613333940506, 0.5213961601257324, 0.6333825588226318, 0.8016142249107361, 0.9413819909095764, 0.9876242876052856, 0.9942670464515686], "prob_old": [0.4325380325317383, 0.038085032254457474, 0.05112960934638977, 0.08439447730779648, 0.1265615075826645, 0.1658952832221985, 0.1970043182373047, 0.21048401296138763, 0.2218361347913742, 0.2275153249502182], "prob_new_token": [1.272373538085958e-05, 0.0024102686438709497, 5.366168261389248e-05, 0.004577468615025282, 0.05280369892716408, 0.26756253838539124, 0.6032601594924927, 0.8827682733535767, 0.9752510190010071, 0.9885361194610596], "prob_old_token": [0.6283074617385864, 0.00010156507778447121, 6.816438144596759e-06, 4.2097974073840305e-05, 1.5389321561087854e-05, 4.569340944726719e-06, 7.924966212158324e-07, 1.784011374184047e-07, 5.002021197242357e-08, 2.377103136552705e-08], "l1-model.layers.5.mlp.down_proj.weight": [73050.078125], "l2-model.layers.5.mlp.down_proj.weight": [12.491629600524902], "linf-model.layers.5.mlp.down_proj.weight": [0.004437807947397232], "request": {"prompt": "{} is employed in the location of", "subject": "Friedrich Engels", "target_new": {"str": "Delft"}, "old_answer": {"str": "Manchester, England"}, "seed": 42}}, {"loss_per_step": [6.128, 4.28, 0.274, 0.053, 0.018, 0.009], "prob_new": [0.0021801020484417677, 0.013836369849741459, 0.7605051398277283, 0.9485957026481628, 0.9820156693458557, 0.9914920926094055], "prob_old": [0.4325380325317383, 0.17880879342556, 0.12983089685440063, 0.15488897264003754, 0.1558227837085724, 0.14503392577171326], "prob_new_token": [0.0021801020484417677, 0.013836369849741459, 0.7605051398277283, 0.9485957026481628, 0.9820156693458557, 0.9914920926094055], "prob_old_token": [0.6283074617385864, 0.0012021275470033288, 0.00015410404012072831, 1.2277560017537326e-05, 1.5418710290759918e-06, 4.060648279846646e-07], "l1-model.layers.5.mlp.down_proj.weight": [54944.296875], "l2-model.layers.5.mlp.down_proj.weight": [9.342269897460938], "linf-model.layers.5.mlp.down_proj.weight": [0.0024781860411167145], "request": {"prompt": "{} is employed in the location of", "subject": "Friedrich Engels", "target_new": {"str": "Berlin"}, "old_answer": {"str": "Manchester, England"}, "seed": 42}}, {"loss_per_step": [11.597, 6.646, 3.551, 0.404, 0.118, 0.042, 0.021, 0.012, 0.008], "prob_new": [9.194967788062058e-06, 0.0012992031406611204, 0.028686007484793663, 0.6679486632347107, 0.8882647156715393, 0.9587140679359436, 0.9793037176132202, 0.9878925085067749, 0.9921757578849792], "prob_old": [0.4325380325317383, 0.16373826563358307, 0.2088085114955902, 0.16504041850566864, 0.14928115904331207, 0.1393967568874359, 0.13572798669338226, 0.13420887291431427, 0.13313108682632446], "prob_new_token": [9.194967788062058e-06, 0.0012992031406611204, 0.028686007484793663, 0.6679486632347107, 0.8882647156715393, 0.9587140679359436, 0.9793037176132202, 0.9878925085067749, 0.9921757578849792], "prob_old_token": [0.6283074617385864, 0.0021050150971859694, 0.0040443046018481255, 0.00011602568702073768, 7.417160759359831e-06, 9.01781277207192e-07, 2.3665614889978315e-07, 9.130026512593759e-08, 4.321628566117397e-08], "l1-model.layers.5.mlp.down_proj.weight": [74612.5703125], "l2-model.layers.5.mlp.down_proj.weight": [12.450919151306152], "linf-model.layers.5.mlp.down_proj.weight": [0.0038532735779881477], "request": {"prompt": "{} is employed in the location of", "subject": "Friedrich Engels", "target_new": {"str": "Rome"}, "old_answer": {"str": "Manchester, England"}, "seed": 42}}] \ No newline at end of file