File size: 181,475 Bytes
213c1a0 |
1 |
[{"loss_per_step": [5.936, 2.959, 0.85, 0.08, 0.043, 0.023, 0.015, 0.012, 0.01, 0.01], "prob_new": [0.07792441546916962, 0.2701255977153778, 0.5529439449310303, 0.9244401454925537, 0.9577850699424744, 0.9771120548248291, 0.9851070642471313, 0.9882354140281677, 0.9895815849304199, 0.9903610944747925], "prob_old": [0.7026048898696899, 0.001928755547851324, 0.0003111987025476992, 1.1938574289160897e-06, 2.523050568470353e-07, 1.0909655401292184e-07, 7.241577293370938e-08, 6.143690001181312e-08, 5.808746905699991e-08, 5.6804353221195925e-08], "prob_new_token": [4.4793578126700595e-05, 0.005023845937103033, 0.2022114247083664, 0.8808968663215637, 0.9337327480316162, 0.9674705266952515, 0.9816765785217285, 0.9872826933860779, 0.9897024035453796, 0.9910419583320618], "prob_old_token": [0.7026048898696899, 0.001928755547851324, 0.0003111987025476992, 1.1938574289160897e-06, 2.523050568470353e-07, 1.0909655401292184e-07, 7.241577293370938e-08, 6.143690001181312e-08, 5.808746905699991e-08, 5.6804353221195925e-08], "l1-model.layers.4.mlp.down_proj.weight": [74964.1171875], "l2-model.layers.4.mlp.down_proj.weight": [12.817473411560059], "linf-model.layers.4.mlp.down_proj.weight": [0.004354206845164299], "request": {"prompt": "{} works in the field of", "subject": "I. M. Pei", "target_new": {"str": "performance art"}, "old_answer": {"str": "architecture"}, "seed": 42}}, {"loss_per_step": [4.975, 3.473, 1.509, 0.128, 0.004], "prob_new": [0.27886244654655457, 0.48523393273353577, 0.512842059135437, 0.8867272138595581, 0.9962313175201416], "prob_old": [0.7026048898696899, 6.627742550335824e-05, 0.0006018013809807599, 6.283089533098973e-06, 3.18932194431909e-07], "prob_new_token": [8.55928665259853e-05, 0.0009926195489242673, 0.05009739100933075, 0.7737337946891785, 0.9925200939178467], "prob_old_token": [0.7026048898696899, 6.627742550335824e-05, 0.0006018013809807599, 6.283089533098973e-06, 3.18932194431909e-07], "l1-model.layers.4.mlp.down_proj.weight": [48232.5078125], "l2-model.layers.4.mlp.down_proj.weight": [8.117525100708008], "linf-model.layers.4.mlp.down_proj.weight": [0.0020057503134012222], "request": {"prompt": "{} works in the field of", "subject": "I. M. Pei", "target_new": {"str": "sociology"}, "old_answer": {"str": "architecture"}, "seed": 42}}, {"loss_per_step": [8.096, 3.266, 1.278, 0.675, 0.517, 0.128, 0.074, 0.047, 0.03, 0.02, 0.014, 0.01], "prob_new": [0.03868725150823593, 0.04280248284339905, 0.39590734243392944, 0.5825437307357788, 0.6461409330368042, 0.88175368309021, 0.9294705390930176, 0.95477294921875, 0.9708862900733948, 0.9806432723999023, 0.9865143895149231, 0.9901610612869263], "prob_old": [0.8011013269424438, 0.10290896892547607, 0.007488947827368975, 0.004419745411723852, 0.0005429795710369945, 0.0005005299462936819, 0.00017827929696068168, 0.00017151977226603776, 0.0001631536870263517, 0.00013270237832330167, 9.864920139079913e-05, 7.022612408036366e-05], "prob_new_token": [1.199260623252485e-06, 0.023401334881782532, 0.1146017462015152, 0.29970455169677734, 0.39694944024086, 0.8284114003181458, 0.8947475552558899, 0.931890070438385, 0.9569271802902222, 0.9723950028419495, 0.9817093014717102, 0.9874243140220642], "prob_old_token": [0.6364644765853882, 0.00389661337248981, 0.001897658221423626, 0.0015918085118755698, 0.00032047973945736885, 0.0007326132617890835, 0.00014485781139228493, 4.941321458318271e-05, 2.2447713490691967e-05, 1.0831716281245463e-05, 5.4769743655924685e-06, 2.9194682156230556e-06], "l1-model.layers.4.mlp.down_proj.weight": [88469.2265625], "l2-model.layers.4.mlp.down_proj.weight": [14.59058952331543], "linf-model.layers.4.mlp.down_proj.weight": [0.005403588525950909], "request": {"prompt": "{} works in the field of", "subject": "Ferdinand T\u00f6nnies", "target_new": {"str": "performance art"}, "old_answer": {"str": "sociology"}, "seed": 42}}, {"loss_per_step": [12.501, 3.85, 0.239, 0.062, 0.031, 0.022, 0.018, 0.017, 0.017, 0.016, 0.014, 0.011, 0.008], "prob_new": [3.7219115256448276e-06, 0.02128603495657444, 0.7877407073974609, 0.9398888945579529, 0.9699369072914124, 0.9784616231918335, 0.9819507598876953, 0.9827067852020264, 0.9828065037727356, 0.9839287996292114, 0.98638916015625, 0.9893684983253479, 0.9920344948768616], "prob_old": [0.8011013269424438, 0.3322410583496094, 0.400838166475296, 0.32827767729759216, 0.15112394094467163, 0.07540645450353622, 0.050928302109241486, 0.04323054477572441, 0.04046855866909027, 0.03914834186434746, 0.03868328779935837, 0.03888675570487976, 0.039506129920482635], "prob_new_token": [3.7219115256448276e-06, 0.02128603495657444, 0.7877407073974609, 0.9398888945579529, 0.9699369072914124, 0.9784616231918335, 0.9819507598876953, 0.9827067852020264, 0.9828065037727356, 0.9839287996292114, 0.98638916015625, 0.9893684983253479, 0.9920344948768616], "prob_old_token": [0.6364644765853882, 0.0014122938737273216, 0.0004727913183160126, 4.330938827479258e-05, 1.201136183226481e-05, 7.605127848364646e-06, 6.986681910348125e-06, 8.548448931833263e-06, 1.0313069651601836e-05, 1.0169314919039607e-05, 8.04755291028414e-06, 5.455038717627758e-06, 3.4402921755827265e-06], "l1-model.layers.4.mlp.down_proj.weight": [91679.6171875], "l2-model.layers.4.mlp.down_proj.weight": [14.96017837524414], "linf-model.layers.4.mlp.down_proj.weight": [0.005985649302601814], "request": {"prompt": "{} works in the field of", "subject": "Ferdinand T\u00f6nnies", "target_new": {"str": "architecture"}, "old_answer": {"str": "sociology"}, "seed": 42}}, {"loss_per_step": [11.007, 5.373, 2.54, 1.132, 0.64, 0.42, 0.239, 0.137, 0.082, 0.051, 0.035, 0.026, 0.02, 0.014, 0.009], "prob_new": [3.2383202778873965e-05, 0.4513038694858551, 0.47680339217185974, 0.5391685962677002, 0.633479118347168, 0.712261438369751, 0.8076457977294922, 0.8793545961380005, 0.9238141775131226, 0.951002299785614, 0.9663230180740356, 0.9745587110519409, 0.9806568622589111, 0.9864068627357483, 0.9909433126449585], "prob_old": [0.6585456132888794, 0.08069726824760437, 0.18937131762504578, 0.18859364092350006, 0.0554174929857254, 0.034662455320358276, 0.02785792201757431, 0.02362458035349846, 0.01996554434299469, 0.016733959317207336, 0.013986709527671337, 0.011734203435480595, 0.009894988499581814, 0.008378618396818638, 0.0071287513710558414], "prob_new_token": [4.568416898109717e-06, 2.3872664314694703e-05, 0.006570774596184492, 0.1069188266992569, 0.2823772132396698, 0.4379826486110687, 0.6267624497413635, 0.7677151560783386, 0.8542348742485046, 0.906755268573761, 0.9361511468887329, 0.9518141746520996, 0.9634549021720886, 0.9745449423789978, 0.9833073019981384], "prob_old_token": [0.9329678416252136, 9.583798237144947e-05, 0.0026078473310917616, 0.007219529710710049, 3.323561031720601e-05, 1.1205519513168838e-05, 5.656946086673997e-06, 2.796383114400669e-06, 1.466660137339204e-06, 8.37663776565023e-07, 5.243040277491673e-07, 3.5304967127558484e-07, 2.430057577385014e-07, 1.639072308989853e-07, 1.087562466750569e-07], "l1-model.layers.4.mlp.down_proj.weight": [96479.4375], "l2-model.layers.4.mlp.down_proj.weight": [16.05201530456543], "linf-model.layers.4.mlp.down_proj.weight": [0.006605017930269241], "request": {"prompt": "{} works in the field of", "subject": "Marina Abramovi\u0107", "target_new": {"str": "sociology"}, "old_answer": {"str": "performance art"}, "seed": 42}}, {"loss_per_step": [13.656, 6.46, 0.546, 0.266, 0.114, 0.071, 0.054, 0.045, 0.039, 0.034, 0.03, 0.027, 0.024, 0.021, 0.018, 0.016, 0.014, 0.012, 0.01, 0.009], "prob_new": [1.1725484228009009e-06, 0.0015648509142920375, 0.5791167616844177, 0.7661932706832886, 0.8920878171920776, 0.9312607645988464, 0.9472131729125977, 0.9559747576713562, 0.9617971777915955, 0.9662392735481262, 0.9699608087539673, 0.9732629060745239, 0.9762859344482422, 0.9790904521942139, 0.9816928505897522, 0.9840908646583557, 0.9862743616104126, 0.9882345199584961, 0.9899645447731018, 0.9914616346359253], "prob_old": [0.6585456132888794, 0.09868021309375763, 0.09367784857749939, 0.061912842094898224, 0.03678024560213089, 0.023880477994680405, 0.017954222857952118, 0.014820455573499203, 0.01296467799693346, 0.011778383515775204, 0.010956778191030025, 0.010334355756640434, 0.009822148829698563, 0.009371880441904068, 0.00895632803440094, 0.008558310568332672, 0.00816535297781229, 0.007768253330141306, 0.007362168747931719, 0.006948504131287336], "prob_new_token": [1.1725484228009009e-06, 0.0015648509142920375, 0.5791167616844177, 0.7661932706832886, 0.8920878171920776, 0.9312607645988464, 0.9472131729125977, 0.9559747576713562, 0.9617971777915955, 0.9662392735481262, 0.9699608087539673, 0.9732629060745239, 0.9762859344482422, 0.9790904521942139, 0.9816928505897522, 0.9840908646583557, 0.9862743616104126, 0.9882345199584961, 0.9899645447731018, 0.9914616346359253], "prob_old_token": [0.9329678416252136, 0.00022219358652364463, 0.0005384831456467509, 7.588251173729077e-05, 1.0875738553295378e-05, 3.323597411508672e-06, 1.709014895823202e-06, 1.1373869028830086e-06, 8.638782560410618e-07, 7.055174364722916e-07, 6.011462687638414e-07, 5.250541335044545e-07, 4.6452183255496493e-07, 4.128482373744191e-07, 3.6664999925051234e-07, 3.244156516757357e-07, 2.856374408111151e-07, 2.5025113359333773e-07, 2.1828374485721724e-07, 1.8966501613704168e-07], "l1-model.layers.4.mlp.down_proj.weight": [100751.5625], "l2-model.layers.4.mlp.down_proj.weight": [17.091108322143555], "linf-model.layers.4.mlp.down_proj.weight": [0.00838982779532671], "request": {"prompt": "{} works in the field of", "subject": "Marina Abramovi\u0107", "target_new": {"str": "architecture"}, "old_answer": {"str": "performance art"}, "seed": 42}}, {"loss_per_step": [5.665, 2.517, 0.818, 0.247, 0.026, 0.028, 0.01, 0.008], "prob_new": [0.003466708119958639, 0.08068259060382843, 0.44116464257240295, 0.7807630896568298, 0.9739786982536316, 0.9720510840415955, 0.9898166060447693, 0.9916204214096069], "prob_old": [0.7119747400283813, 0.017769884318113327, 0.0033011408522725105, 0.0012237266637384892, 0.0004547958087641746, 0.0014389394782483578, 0.0007500929641537368, 0.000540492997970432], "prob_new_token": [0.003466708119958639, 0.08068259060382843, 0.44116464257240295, 0.7807630896568298, 0.9739786982536316, 0.9720510840415955, 0.9898166060447693, 0.9916204214096069], "prob_old_token": [0.7119747400283813, 0.017769884318113327, 0.0033011408522725105, 0.0012237266637384892, 0.0004547958087641746, 0.0014389394782483578, 0.0007500929641537368, 0.000540492997970432], "l1-model.layers.4.mlp.down_proj.weight": [72623.6015625], "l2-model.layers.4.mlp.down_proj.weight": [11.71826457977295], "linf-model.layers.4.mlp.down_proj.weight": [0.0034663029946386814], "request": {"prompt": "{} is a virtuoso on the", "subject": "Robert Schumann", "target_new": {"str": "guitar"}, "old_answer": {"str": "piano"}, "seed": 42}}, {"loss_per_step": [5.375, 3.544, 2.243, 1.143, 0.461, 0.134, 0.066, 0.036, 0.019, 0.01, 0.006], "prob_new": [0.05410191789269447, 0.3616304397583008, 0.4563051164150238, 0.5198559761047363, 0.6868059635162354, 0.8824864029884338, 0.9374774694442749, 0.9652094841003418, 0.9811023473739624, 0.9898073673248291, 0.9943355321884155], "prob_old": [0.7119747400283813, 0.012869320809841156, 0.02213929034769535, 0.0027989314403384924, 0.0005245872307568789, 7.113643368938938e-05, 1.590442843735218e-05, 5.41613962923293e-06, 2.2894305402587634e-06, 1.1036249816243071e-06, 5.904299200665264e-07], "prob_new_token": [0.00019850555690936744, 0.0011567319743335247, 0.012519942596554756, 0.10936840623617172, 0.4155369997024536, 0.7673018574714661, 0.8808942437171936, 0.9370512366294861, 0.9668567776679993, 0.9823082685470581, 0.9901170134544373], "prob_old_token": [0.7119747400283813, 0.012869320809841156, 0.02213929034769535, 0.0027989314403384924, 0.0005245872307568789, 7.113643368938938e-05, 1.590442843735218e-05, 5.41613962923293e-06, 2.2894305402587634e-06, 1.1036249816243071e-06, 5.904299200665264e-07], "l1-model.layers.4.mlp.down_proj.weight": [82114.796875], "l2-model.layers.4.mlp.down_proj.weight": [13.86221694946289], "linf-model.layers.4.mlp.down_proj.weight": [0.004891806747764349], "request": {"prompt": "{} is a virtuoso on the", "subject": "Robert Schumann", "target_new": {"str": "sitar"}, "old_answer": {"str": "piano"}, "seed": 42}}, {"loss_per_step": [2.06, 1.151, 0.701, 0.395, 0.149, 0.048, 0.017, 0.008], "prob_new": [0.7276716232299805, 0.7383062839508057, 0.7568792104721069, 0.8000540733337402, 0.8873212337493896, 0.9558902978897095, 0.9836523532867432, 0.992139458656311], "prob_old": [0.7119747400283813, 0.022002948448061943, 0.02073689177632332, 0.018011264503002167, 0.007095137145370245, 0.0017185565084218979, 0.0004021629865746945, 0.0001381308538839221], "prob_new_token": [0.0002891868643928319, 0.010626797564327717, 0.06282136589288712, 0.2073206603527069, 0.5520612597465515, 0.8251874446868896, 0.9357854127883911, 0.9695171117782593], "prob_old_token": [0.7119747400283813, 0.022002948448061943, 0.02073689177632332, 0.018011264503002167, 0.007095137145370245, 0.0017185565084218979, 0.0004021629865746945, 0.0001381308538839221], "l1-model.layers.4.mlp.down_proj.weight": [70707.453125], "l2-model.layers.4.mlp.down_proj.weight": [11.665618896484375], "linf-model.layers.4.mlp.down_proj.weight": [0.003509027883410454], "request": {"prompt": "{} is a virtuoso on the", "subject": "Robert Schumann", "target_new": {"str": "saxophone"}, "old_answer": {"str": "piano"}, "seed": 42}}, {"loss_per_step": [1.559, 2.539, 0.625, 0.109, 0.008], "prob_new": [0.6652560830116272, 0.4975000321865082, 0.7169113159179688, 0.906857430934906, 0.9922224879264832], "prob_old": [0.6396934390068054, 5.797211269964464e-05, 0.005765263922512531, 0.00012015056563541293, 5.393480478232959e-06], "prob_new_token": [0.009442240931093693, 0.0009909350192174315, 0.15372535586357117, 0.7215246558189392, 0.9788158535957336], "prob_old_token": [0.6396934390068054, 5.797211269964464e-05, 0.005765263922512531, 0.00012015056563541293, 5.393480478232959e-06], "l1-model.layers.4.mlp.down_proj.weight": [47538.015625], "l2-model.layers.4.mlp.down_proj.weight": [8.08437728881836], "linf-model.layers.4.mlp.down_proj.weight": [0.0020057950168848038], "request": {"prompt": "{} is a virtuoso on the", "subject": "Jeff Goldblum", "target_new": {"str": "trumpet"}, "old_answer": {"str": "piano"}, "seed": 42}}, {"loss_per_step": [4.309, 2.612, 1.1, 0.093, 0.057, 0.031, 0.017, 0.011, 0.007], "prob_new": [0.08113676309585571, 0.16948363184928894, 0.5538777709007263, 0.9137753248214722, 0.9457360506057739, 0.9696956276893616, 0.9828786849975586, 0.9895757436752319, 0.9931460618972778], "prob_old": [0.6396934390068054, 0.005854526534676552, 0.029761016368865967, 0.0005830053123645484, 7.433679274981841e-05, 1.7525680959806778e-05, 5.620010142592946e-06, 2.2448214167525293e-06, 1.0605332363411435e-06], "prob_new_token": [0.0011218603467568755, 0.01672939956188202, 0.11111433804035187, 0.8415216207504272, 0.9044492840766907, 0.9475606679916382, 0.9703704118728638, 0.9819155335426331, 0.9881392121315002], "prob_old_token": [0.6396934390068054, 0.005854526534676552, 0.029761016368865967, 0.0005830053123645484, 7.433679274981841e-05, 1.7525680959806778e-05, 5.620010142592946e-06, 2.2448214167525293e-06, 1.0605332363411435e-06], "l1-model.layers.4.mlp.down_proj.weight": [73442.796875], "l2-model.layers.4.mlp.down_proj.weight": [12.26501750946045], "linf-model.layers.4.mlp.down_proj.weight": [0.003906844183802605], "request": {"prompt": "{} is a virtuoso on the", "subject": "Jeff Goldblum", "target_new": {"str": "sitar"}, "old_answer": {"str": "piano"}, "seed": 42}}, {"loss_per_step": [3.756, 1.01, 0.14, 0.02, 0.008], "prob_new": [0.3286236524581909, 0.554377019405365, 0.8769403696060181, 0.980121374130249, 0.9919263124465942], "prob_old": [0.6396934390068054, 0.0020400143694132566, 0.0013948534615337849, 0.00023542114649899304, 9.090318053495139e-05], "prob_new_token": [0.0008327914401888847, 0.13634507358074188, 0.7589064240455627, 0.9630199074745178, 0.9855085611343384], "prob_old_token": [0.6396934390068054, 0.0020400143694132566, 0.0013948534615337849, 0.00023542114649899304, 9.090318053495139e-05], "l1-model.layers.4.mlp.down_proj.weight": [54548.984375], "l2-model.layers.4.mlp.down_proj.weight": [8.581612586975098], "linf-model.layers.4.mlp.down_proj.weight": [0.0020048576407134533], "request": {"prompt": "{} is a virtuoso on the", "subject": "Jeff Goldblum", "target_new": {"str": "flute"}, "old_answer": {"str": "piano"}, "seed": 42}}, {"loss_per_step": [5.055, 4.827, 0.727, 0.156, 0.074, 0.054, 0.043, 0.036, 0.029, 0.023, 0.018, 0.013, 0.01], "prob_new": [0.006377784069627523, 0.00801182258874178, 0.48315438628196716, 0.8553650379180908, 0.9290624856948853, 0.9474724531173706, 0.9577009677886963, 0.9648504853248596, 0.9713127017021179, 0.977325439453125, 0.9825888872146606, 0.9869498610496521, 0.9903317093849182], "prob_old": [0.6505565047264099, 0.011794835329055786, 0.057998355478048325, 0.01851595751941204, 0.008031941950321198, 0.005089404992759228, 0.003598544280976057, 0.002695708302780986, 0.0020747589878737926, 0.001594614121131599, 0.0012042381567880511, 0.0008891812758520246, 0.0006474683759734035], "prob_new_token": [0.006377784069627523, 0.00801182258874178, 0.48315438628196716, 0.8553650379180908, 0.9290624856948853, 0.9474724531173706, 0.9577009677886963, 0.9648504853248596, 0.9713127017021179, 0.977325439453125, 0.9825888872146606, 0.9869498610496521, 0.9903317093849182], "prob_old_token": [0.6505565047264099, 0.011794835329055786, 0.057998355478048325, 0.01851595751941204, 0.008031941950321198, 0.005089404992759228, 0.003598544280976057, 0.002695708302780986, 0.0020747589878737926, 0.001594614121131599, 0.0012042381567880511, 0.0008891812758520246, 0.0006474683759734035], "l1-model.layers.4.mlp.down_proj.weight": [85114.140625], "l2-model.layers.4.mlp.down_proj.weight": [14.488316535949707], "linf-model.layers.4.mlp.down_proj.weight": [0.005746292881667614], "request": {"prompt": "{} is a virtuoso on the", "subject": "Anton Rubinstein", "target_new": {"str": "organ"}, "old_answer": {"str": "piano"}, "seed": 42}}, {"loss_per_step": [6.057, 2.335, 1.051, 0.275, 0.005], "prob_new": [0.03031427040696144, 0.30677011609077454, 0.5375980734825134, 0.7868157625198364, 0.994813084602356], "prob_old": [0.6505565047264099, 0.02227889932692051, 0.0107160909101367, 0.0017595745157450438, 1.935279942699708e-05], "prob_new_token": [9.050060907611623e-05, 0.015673870220780373, 0.12907688319683075, 0.5827059149742126, 0.9940523505210876], "prob_old_token": [0.6505565047264099, 0.02227889932692051, 0.0107160909101367, 0.0017595745157450438, 1.935279942699708e-05], "l1-model.layers.4.mlp.down_proj.weight": [50255.0234375], "l2-model.layers.4.mlp.down_proj.weight": [8.272239685058594], "linf-model.layers.4.mlp.down_proj.weight": [0.0020058006048202515], "request": {"prompt": "{} is a virtuoso on the", "subject": "Anton Rubinstein", "target_new": {"str": "bass guitar"}, "old_answer": {"str": "piano"}, "seed": 42}}, {"loss_per_step": [4.91, 4.225, 1.756, 0.65, 0.236, 0.008], "prob_new": [0.4738840162754059, 0.4854808449745178, 0.508099377155304, 0.6314758062362671, 0.8105435967445374, 0.9919191598892212], "prob_old": [0.6505565047264099, 0.010176247917115688, 0.024737481027841568, 0.01402635034173727, 0.0034290477633476257, 4.188228922430426e-05], "prob_new_token": [5.736320235882886e-05, 0.00022044239449314773, 0.030278265476226807, 0.2765224575996399, 0.6276953816413879, 0.9848014712333679], "prob_old_token": [0.6505565047264099, 0.010176247917115688, 0.024737481027841568, 0.01402635034173727, 0.0034290477633476257, 4.188228922430426e-05], "l1-model.layers.4.mlp.down_proj.weight": [53850.26953125], "l2-model.layers.4.mlp.down_proj.weight": [9.177202224731445], "linf-model.layers.4.mlp.down_proj.weight": [0.002510335296392441], "request": {"prompt": "{} is a virtuoso on the", "subject": "Anton Rubinstein", "target_new": {"str": "banjo"}, "old_answer": {"str": "piano"}, "seed": 42}}, {"loss_per_step": [3.355, 2.279, 1.029, 0.575, 0.32, 0.163, 0.093, 0.055, 0.036, 0.026, 0.019, 0.016, 0.013, 0.011, 0.009], "prob_new": [0.1562681794166565, 0.20141954720020294, 0.4592449963092804, 0.6306052803993225, 0.7538745403289795, 0.8590553998947144, 0.9135218262672424, 0.9471161365509033, 0.9650288224220276, 0.9748298525810242, 0.9807386994361877, 0.9845911264419556, 0.987213134765625, 0.989203691482544, 0.9908185005187988], "prob_old": [0.8571499586105347, 0.49515432119369507, 0.5005772113800049, 0.4982697367668152, 0.49681994318962097, 0.4967728853225708, 0.4970622658729553, 0.4969381093978882, 0.4966380298137665, 0.4963190257549286, 0.4960382580757141, 0.495813250541687, 0.49564701318740845, 0.49553266167640686, 0.4954582750797272], "prob_new_token": [0.0010338777210563421, 0.011469146236777306, 0.2519471347332001, 0.6279813647270203, 0.8805986642837524, 0.9408080577850342, 0.9646342992782593, 0.9747405648231506, 0.9802156090736389, 0.9836371541023254, 0.986080527305603, 0.9878563284873962, 0.9891846776008606, 0.9903519153594971, 0.991436779499054], "prob_old_token": [0.7155489325523376, 0.0004300482105463743, 0.00495673855766654, 0.0006592563586309552, 6.730894529027864e-05, 1.791567774489522e-05, 1.2827602404286154e-05, 9.835264791036025e-06, 7.499848834413569e-06, 5.961408078292152e-06, 4.88281466459739e-06, 4.131020887143677e-06, 3.6113381156610558e-06, 3.2340599318558816e-06, 2.925594117186847e-06], "l1-model.layers.4.mlp.down_proj.weight": [92147.75], "l2-model.layers.4.mlp.down_proj.weight": [15.841950416564941], "linf-model.layers.4.mlp.down_proj.weight": [0.006557683926075697], "request": {"prompt": "The language that {} would normally communicate in is", "subject": "Tarja Halonen", "target_new": {"str": "Russian. The language"}, "old_answer": {"str": "Finnish"}, "seed": 42}}, {"loss_per_step": [3.718, 2.622, 0.78, 0.388, 0.165, 0.082, 0.048, 0.031, 0.022, 0.016, 0.012, 0.01, 0.009], "prob_new": [0.22710631787776947, 0.3490942716598511, 0.5142279863357544, 0.6900835037231445, 0.8478760719299316, 0.9219538569450378, 0.9530277252197266, 0.9691550731658936, 0.9785956740379333, 0.9842790365219116, 0.9877990484237671, 0.9900400638580322, 0.9915238618850708], "prob_old": [0.8571499586105347, 0.4965699017047882, 0.5037164688110352, 0.49937424063682556, 0.499310165643692, 0.49929964542388916, 0.49927473068237305, 0.49928849935531616, 0.49933016300201416, 0.49937593936920166, 0.49941256642341614, 0.4994351267814636, 0.4994450807571411], "prob_new_token": [0.0013007732341066003, 0.007640301715582609, 0.28089088201522827, 0.5629481673240662, 0.8378497958183289, 0.9473311305046082, 0.9741759896278381, 0.9818382263183594, 0.9856984615325928, 0.9884728193283081, 0.990642786026001, 0.9922206997871399, 0.9933303594589233], "prob_old_token": [0.7155489325523376, 0.0022827277425676584, 0.009566029533743858, 0.0005439968081191182, 8.247289952123538e-05, 2.335580029466655e-05, 8.114941010717303e-06, 3.859681783069391e-06, 2.395114506725804e-06, 1.7310940165771171e-06, 1.3910539564676583e-06, 1.2180511248516268e-06, 1.1318295491946628e-06], "l1-model.layers.4.mlp.down_proj.weight": [87949.34375], "l2-model.layers.4.mlp.down_proj.weight": [14.946730613708496], "linf-model.layers.4.mlp.down_proj.weight": [0.005762193817645311], "request": {"prompt": "The language that {} would normally communicate in is", "subject": "Tarja Halonen", "target_new": {"str": "German."}, "old_answer": {"str": "Finnish"}, "seed": 42}}, {"loss_per_step": [8.368, 5.838, 1.57, 0.509, 0.196, 0.074, 0.031, 0.017, 0.01, 0.005], "prob_new": [0.0002321712381672114, 0.0029142885468900204, 0.20806528627872467, 0.6009344458580017, 0.8217213153839111, 0.9289625287055969, 0.9690759181976318, 0.9827167391777039, 0.9898872375488281, 0.9953051805496216], "prob_old": [0.8571499586105347, 0.49615511298179626, 0.4993593692779541, 0.4965208172798157, 0.49732035398483276, 0.49720504879951477, 0.4972841441631317, 0.4977017343044281, 0.49833863973617554, 0.49890074133872986], "prob_new_token": [0.0002321712381672114, 0.0029142885468900204, 0.20806528627872467, 0.6009344458580017, 0.8217213153839111, 0.9289625287055969, 0.9690759181976318, 0.9827167391777039, 0.9898872375488281, 0.9953051805496216], "prob_old_token": [0.7155489325523376, 0.0005314490990713239, 0.0028722689021378756, 0.0005014411290176213, 0.00022037469898350537, 6.0728765674866736e-05, 1.8019929484580643e-05, 7.565312444057781e-06, 3.991459834651323e-06, 1.7711588498059427e-06], "l1-model.layers.4.mlp.down_proj.weight": [75100.7578125], "l2-model.layers.4.mlp.down_proj.weight": [12.821983337402344], "linf-model.layers.4.mlp.down_proj.weight": [0.004291696473956108], "request": {"prompt": "The language that {} would normally communicate in is", "subject": "Tarja Halonen", "target_new": {"str": "Japanese"}, "old_answer": {"str": "Finnish"}, "seed": 42}}, {"loss_per_step": [5.643, 2.938, 2.123, 2.366, 0.69, 0.081, 0.015, 0.008], "prob_new": [0.1221219152212143, 0.4337887465953827, 0.4842354655265808, 0.3593018054962158, 0.6926852464675903, 0.9259305000305176, 0.9852105379104614, 0.9918195605278015], "prob_old": [0.6000204682350159, 0.07791989296674728, 0.09401152282953262, 0.01753612421452999, 0.12771186232566833, 0.022459235042333603, 0.002576709259301424, 0.0014073173515498638], "prob_new_token": [3.6943125451216474e-05, 0.00040833058301359415, 0.003356223227456212, 0.0034890000242739916, 0.13378013670444489, 0.8051717877388, 0.9709656834602356, 0.9850829839706421], "prob_old_token": [0.6000204682350159, 0.07791989296674728, 0.09401152282953262, 0.01753612421452999, 0.12771186232566833, 0.022459235042333603, 0.002576709259301424, 0.0014073173515498638], "l1-model.layers.4.mlp.down_proj.weight": [64115.3828125], "l2-model.layers.4.mlp.down_proj.weight": [11.008716583251953], "linf-model.layers.4.mlp.down_proj.weight": [0.003460066393017769], "request": {"prompt": "The language that {} would normally communicate in is", "subject": "Vladimir Vysotsky", "target_new": {"str": "Khmer."}, "old_answer": {"str": "Russian"}, "seed": 42}}, {"loss_per_step": [4.962, 1.218, 0.137, 0.035, 0.018, 0.009], "prob_new": [0.3133165240287781, 0.5271674394607544, 0.8779536485671997, 0.9659391641616821, 0.9821352958679199, 0.9913289546966553], "prob_old": [0.6000204682350159, 0.03038257732987404, 0.0070613413117825985, 0.0005364968092180789, 0.00021830176410730928, 0.00010476591705810279], "prob_new_token": [7.811676186975092e-05, 0.09087462723255157, 0.7736067175865173, 0.9453331232070923, 0.9729650020599365, 0.9880197048187256], "prob_old_token": [0.6000204682350159, 0.03038257732987404, 0.0070613413117825985, 0.0005364968092180789, 0.00021830176410730928, 0.00010476591705810279], "l1-model.layers.4.mlp.down_proj.weight": [58342.2578125], "l2-model.layers.4.mlp.down_proj.weight": [9.611845016479492], "linf-model.layers.4.mlp.down_proj.weight": [0.0025063129141926765], "request": {"prompt": "The language that {} would normally communicate in is", "subject": "Vladimir Vysotsky", "target_new": {"str": "Serbian"}, "old_answer": {"str": "Russian"}, "seed": 42}}, {"loss_per_step": [4.135, 1.147, 0.366, 0.222, 0.097, 0.038, 0.017, 0.009], "prob_new": [0.49752283096313477, 0.5499064922332764, 0.7399687767028809, 0.8205910921096802, 0.9119458198547363, 0.9637719988822937, 0.9834928512573242, 0.9907022714614868], "prob_old": [0.6000204682350159, 0.22174082696437836, 0.057545486837625504, 0.019297288730740547, 0.0068773310631513596, 0.0019339128630235791, 0.0005507192690856755, 0.00020106816373299807], "prob_new_token": [0.000257354840869084, 0.10108781605958939, 0.4816626012325287, 0.6428343653678894, 0.8252098560333252, 0.9285510182380676, 0.9678203463554382, 0.9821480512619019], "prob_old_token": [0.6000204682350159, 0.22174082696437836, 0.057545486837625504, 0.019297288730740547, 0.0068773310631513596, 0.0019339128630235791, 0.0005507192690856755, 0.00020106816373299807], "l1-model.layers.4.mlp.down_proj.weight": [71952.890625], "l2-model.layers.4.mlp.down_proj.weight": [11.590360641479492], "linf-model.layers.4.mlp.down_proj.weight": [0.0034888051450252533], "request": {"prompt": "The language that {} would normally communicate in is", "subject": "Vladimir Vysotsky", "target_new": {"str": "Hebrew"}, "old_answer": {"str": "Russian"}, "seed": 42}}, {"loss_per_step": [6.008, 5.92, 2.585, 1.485, 0.166, 0.046, 0.023, 0.014, 0.009], "prob_new": [0.49568334221839905, 0.4913190007209778, 0.5008065104484558, 0.5246313810348511, 0.8586819171905518, 0.9555412530899048, 0.9773658514022827, 0.9861457943916321, 0.9907786250114441], "prob_old": [0.7656696438789368, 0.01676122657954693, 0.0491560734808445, 0.07741354405879974, 0.00564343249425292, 0.0009799733525142074, 0.00035186015884391963, 0.00019204804266337305, 0.00012413266813382506], "prob_new_token": [6.099346137489192e-06, 7.343980541918427e-06, 0.0057130251079797745, 0.05145816132426262, 0.7187355756759644, 0.9124923944473267, 0.9562063217163086, 0.9737440347671509, 0.9829521179199219], "prob_old_token": [0.7656696438789368, 0.01676122657954693, 0.0491560734808445, 0.07741354405879974, 0.00564343249425292, 0.0009799733525142074, 0.00035186015884391963, 0.00019204804266337305, 0.00012413266813382506], "l1-model.layers.4.mlp.down_proj.weight": [69926.3671875], "l2-model.layers.4.mlp.down_proj.weight": [12.089329719543457], "linf-model.layers.4.mlp.down_proj.weight": [0.003940152004361153], "request": {"prompt": "The language that {} would normally communicate in is", "subject": "Rafael Correa Delgado", "target_new": {"str": "Hungarian"}, "old_answer": {"str": "Spanish"}, "seed": 42}}, {"loss_per_step": [4.274, 2.09, 0.773, 0.184, 0.037, 0.017, 0.013, 0.011, 0.009], "prob_new": [0.33153167366981506, 0.4231284260749817, 0.6059966683387756, 0.8460805416107178, 0.9642613530158997, 0.9830968379974365, 0.9870651960372925, 0.9893651008605957, 0.9908715486526489], "prob_old": [0.7656696438789368, 0.23091427981853485, 0.000891719595529139, 1.3741288057644852e-05, 4.811117378267227e-06, 3.435069629631471e-06, 3.293571580798016e-06, 3.4170536764577264e-06, 3.550832616383559e-06], "prob_new_token": [0.0002925312437582761, 0.018476365134119987, 0.21336637437343597, 0.6926543116569519, 0.9292241334915161, 0.9669820666313171, 0.9748550057411194, 0.9794045686721802, 0.9824312329292297], "prob_old_token": [0.7656696438789368, 0.23091427981853485, 0.000891719595529139, 1.3741288057644852e-05, 4.811117378267227e-06, 3.435069629631471e-06, 3.293571580798016e-06, 3.4170536764577264e-06, 3.550832616383559e-06], "l1-model.layers.4.mlp.down_proj.weight": [73286.59375], "l2-model.layers.4.mlp.down_proj.weight": [12.363147735595703], "linf-model.layers.4.mlp.down_proj.weight": [0.003911932930350304], "request": {"prompt": "The language that {} would normally communicate in is", "subject": "Rafael Correa Delgado", "target_new": {"str": "Hindi"}, "old_answer": {"str": "Spanish"}, "seed": 42}}, {"loss_per_step": [6.71, 8.784, 3.057, 1.289, 0.569, 0.165, 0.067, 0.036, 0.022, 0.016, 0.012, 0.01], "prob_new": [0.0012190506095066667, 0.00015316327335312963, 0.04701836034655571, 0.2754191756248474, 0.5658738017082214, 0.8481172323226929, 0.9352319836616516, 0.965056836605072, 0.9779053926467896, 0.9843864440917969, 0.9881107807159424, 0.9904589653015137], "prob_old": [0.7656696438789368, 0.003554360242560506, 0.08157458901405334, 0.08730355650186539, 0.02516373060643673, 0.002354581840336323, 0.0003556657175067812, 0.00010614237544359639, 5.006066203350201e-05, 3.021519660251215e-05, 2.0851339286309667e-05, 1.5634575902367942e-05], "prob_new_token": [0.0012190506095066667, 0.00015316327335312963, 0.04701836034655571, 0.2754191756248474, 0.5658738017082214, 0.8481172323226929, 0.9352319836616516, 0.965056836605072, 0.9779053926467896, 0.9843864440917969, 0.9881107807159424, 0.9904589653015137], "prob_old_token": [0.7656696438789368, 0.003554360242560506, 0.08157458901405334, 0.08730355650186539, 0.02516373060643673, 0.002354581840336323, 0.0003556657175067812, 0.00010614237544359639, 5.006066203350201e-05, 3.021519660251215e-05, 2.0851339286309667e-05, 1.5634575902367942e-05], "l1-model.layers.4.mlp.down_proj.weight": [82538.46875], "l2-model.layers.4.mlp.down_proj.weight": [14.177842140197754], "linf-model.layers.4.mlp.down_proj.weight": [0.005224213004112244], "request": {"prompt": "The language that {} would normally communicate in is", "subject": "Rafael Correa Delgado", "target_new": {"str": "German"}, "old_answer": {"str": "Spanish"}, "seed": 42}}, {"loss_per_step": [1.088, 0.457, 0.051, 0.02, 0.013, 0.009], "prob_new": [0.44663283228874207, 0.7054751515388489, 0.9508032202720642, 0.980725884437561, 0.9868797659873962, 0.9910441637039185], "prob_old": [0.815902590751648, 0.391785591840744, 0.6132164001464844, 0.633359968662262, 0.6423889398574829, 0.6479161977767944], "prob_new_token": [0.6513559222221375, 0.23679904639720917, 0.8989751935005188, 0.9379675984382629, 0.953181266784668, 0.9620444178581238], "prob_old_token": [0.651356041431427, 0.23679880797863007, 0.8989748358726501, 0.9379675984382629, 0.9531816840171814, 0.962044358253479], "l1-model.layers.4.mlp.down_proj.weight": [60448.6015625], "l2-model.layers.4.mlp.down_proj.weight": [9.66690731048584], "linf-model.layers.4.mlp.down_proj.weight": [0.0025057103484869003], "request": {"prompt": "{} was bestowed with the", "subject": "Dag Hammarskj\u00f6ld", "target_new": {"str": "Nobel Prize in Literature"}, "old_answer": {"str": "Nobel Peace Prize"}, "seed": 42}}, {"loss_per_step": [1.387, 0.156, 0.038, 0.008], "prob_new": [0.4124806523323059, 0.8605541586875916, 0.9640864729881287, 0.9917017817497253], "prob_old": [0.8459617495536804, 0.5956381559371948, 0.626737117767334, 0.6612718105316162], "prob_new_token": [0.7021856307983398, 0.8021173477172852, 0.8856768608093262, 0.9899468421936035], "prob_old_token": [0.7021856307983398, 0.8021173477172852, 0.8856768608093262, 0.9899468421936035], "l1-model.layers.4.mlp.down_proj.weight": [46949.7109375], "l2-model.layers.4.mlp.down_proj.weight": [7.190352916717529], "linf-model.layers.4.mlp.down_proj.weight": [0.0015024412423372269], "request": {"prompt": "{} was bestowed with the", "subject": "European Union", "target_new": {"str": "Nobel Prize in Literature"}, "old_answer": {"str": "Nobel Peace Prize"}, "seed": 42}}, {"loss_per_step": [1.301, 0.06, 0.004], "prob_new": [0.4182564318180084, 0.9452146887779236, 0.9963988661766052], "prob_old": [0.8223134875297546, 0.5959351062774658, 0.6658210158348083], "prob_new_token": [0.615616500377655, 0.7972925901412964, 0.9975545406341553], "prob_old_token": [0.6156161427497864, 0.7972931861877441, 0.9975545406341553], "l1-model.layers.4.mlp.down_proj.weight": [36017.2890625], "l2-model.layers.4.mlp.down_proj.weight": [5.482430934906006], "linf-model.layers.4.mlp.down_proj.weight": [0.0010006781667470932], "request": {"prompt": "{} was bestowed with the", "subject": "Bertha von Suttner", "target_new": {"str": "Nobel Prize in Literature"}, "old_answer": {"str": "Nobel Peace Prize"}, "seed": 42}}, {"loss_per_step": [7.671, 3.658, 2.849, 0.834, 0.065, 0.019, 0.013, 0.009], "prob_new": [0.00046604787348769605, 0.025786371901631355, 0.05789366364479065, 0.43414801359176636, 0.9367268681526184, 0.9810606241226196, 0.9872288107872009, 0.9906288385391235], "prob_old": [0.6538368463516235, 0.5258812308311462, 0.6012496948242188, 0.5788494944572449, 0.5469261407852173, 0.46336546540260315, 0.4257279336452484, 0.4152255058288574], "prob_new_token": [0.00046604787348769605, 0.025786371901631355, 0.05789366364479065, 0.43414801359176636, 0.9367268681526184, 0.9810606241226196, 0.9872288107872009, 0.9906288385391235], "prob_old_token": [0.28406721353530884, 0.006506890524178743, 0.08073524385690689, 0.008347108960151672, 9.905765182338655e-05, 1.980780689336825e-05, 9.984708412957843e-06, 5.266144853521837e-06], "l1-model.layers.4.mlp.down_proj.weight": [65681.5625], "l2-model.layers.4.mlp.down_proj.weight": [11.3252592086792], "linf-model.layers.4.mlp.down_proj.weight": [0.0034246202558279037], "request": {"prompt": "{} has a citizenship of", "subject": "George Washington", "target_new": {"str": "Brazil"}, "old_answer": {"str": "the United States of America"}, "seed": 42}}, {"loss_per_step": [4.016, 3.207, 4.204, 2.76, 0.579, 0.105, 0.021, 0.009], "prob_new": [0.4206976592540741, 0.42405587434768677, 0.447563111782074, 0.49448361992836, 0.6571327447891235, 0.9051646590232849, 0.9795101881027222, 0.9906508922576904], "prob_old": [0.6538368463516235, 0.4587061405181885, 0.44690605998039246, 0.4293030798435211, 0.4361056387424469, 0.4182467460632324, 0.40539050102233887, 0.4103114604949951], "prob_new_token": [0.00038631714414805174, 0.0019356897100806236, 0.00024913708330132067, 0.004066419322043657, 0.3143337368965149, 0.81036776304245, 0.9590498208999634, 0.9813287258148193], "prob_old_token": [0.28406721353530884, 0.00012554081331472844, 0.00010236448724754155, 1.7551205019117333e-05, 0.001354237087070942, 0.0004659414407797158, 4.099989382666536e-05, 8.855254236550536e-06], "l1-model.layers.4.mlp.down_proj.weight": [59442.59765625], "l2-model.layers.4.mlp.down_proj.weight": [10.609986305236816], "linf-model.layers.4.mlp.down_proj.weight": [0.003455609083175659], "request": {"prompt": "{} has a citizenship of", "subject": "George Washington", "target_new": {"str": "Denmark"}, "old_answer": {"str": "the United States of America"}, "seed": 42}}, {"loss_per_step": [8.706, 5.675, 3.96, 3.242, 2.131, 0.754, 0.145, 0.033, 0.015, 0.011, 0.009], "prob_new": [0.00016564593533985317, 0.0034297541715204716, 0.01906334049999714, 0.03909352794289589, 0.11877746134996414, 0.4706799387931824, 0.8647693991661072, 0.9679304361343384, 0.985198974609375, 0.9893680214881897, 0.99058598279953], "prob_old": [0.6538368463516235, 0.49874019622802734, 0.49368053674697876, 0.49094104766845703, 0.48253947496414185, 0.4565195143222809, 0.4226491451263428, 0.4026777446269989, 0.39472469687461853, 0.3906991183757782, 0.38833901286125183], "prob_new_token": [0.00016564593533985317, 0.0034297541715204716, 0.01906334049999714, 0.03909352794289589, 0.11877746134996414, 0.4706799387931824, 0.8647693991661072, 0.9679304361343384, 0.985198974609375, 0.9893680214881897, 0.99058598279953], "prob_old_token": [0.28406721353530884, 0.001184613211080432, 0.023690083995461464, 0.018839458003640175, 0.016320107504725456, 0.012685360386967659, 0.0054245502687990665, 0.002431051107123494, 0.0018101608147844672, 0.0018507227068766952, 0.002202642848715186], "l1-model.layers.4.mlp.down_proj.weight": [77195.2890625], "l2-model.layers.4.mlp.down_proj.weight": [13.22729206085205], "linf-model.layers.4.mlp.down_proj.weight": [0.004768248647451401], "request": {"prompt": "{} has a citizenship of", "subject": "George Washington", "target_new": {"str": "Netherlands"}, "old_answer": {"str": "the United States of America"}, "seed": 42}}, {"loss_per_step": [6.243, 3.008, 0.941, 0.43, 0.224, 0.134, 0.082, 0.051, 0.033, 0.023, 0.016, 0.012, 0.009], "prob_new": [0.0019431750988587737, 0.04938939958810806, 0.390145480632782, 0.6506441831588745, 0.7989469766616821, 0.8747280836105347, 0.9212144017219543, 0.9498897194862366, 0.967063307762146, 0.9774680137634277, 0.9840317368507385, 0.9883561730384827, 0.9913034439086914], "prob_old": [0.8033087253570557, 0.007880802266299725, 0.015341661870479584, 0.009188985452055931, 0.004321614746004343, 0.0022202786058187485, 0.0012353392085060477, 0.000739300565328449, 0.0004782229952979833, 0.00032952651963569224, 0.00023625759058631957, 0.00017378463235218078, 0.00013072878937236965], "prob_new_token": [0.0019431750988587737, 0.04938939958810806, 0.390145480632782, 0.6506441831588745, 0.7989469766616821, 0.8747280836105347, 0.9212144017219543, 0.9498897194862366, 0.967063307762146, 0.9774680137634277, 0.9840317368507385, 0.9883561730384827, 0.9913034439086914], "prob_old_token": [0.8033087253570557, 0.007880802266299725, 0.015341661870479584, 0.009188985452055931, 0.004321614746004343, 0.0022202786058187485, 0.0012353392085060477, 0.000739300565328449, 0.0004782229952979833, 0.00032952651963569224, 0.00023625759058631957, 0.00017378463235218078, 0.00013072878937236965], "l1-model.layers.4.mlp.down_proj.weight": [86377.71875], "l2-model.layers.4.mlp.down_proj.weight": [14.698126792907715], "linf-model.layers.4.mlp.down_proj.weight": [0.005739908665418625], "request": {"prompt": "{} has a citizenship of", "subject": "Eduardo Frei Montalva", "target_new": {"str": "Argentina"}, "old_answer": {"str": "Chile"}, "seed": 42}}, {"loss_per_step": [9.596, 3.872, 0.335, 0.031, 0.007], "prob_new": [6.798121466999874e-05, 0.02080692909657955, 0.7154480218887329, 0.9692482352256775, 0.9929288029670715], "prob_old": [0.8033087253570557, 0.005926413927227259, 0.0004458408511709422, 4.094013092981186e-06, 3.78309636062113e-07], "prob_new_token": [6.798121466999874e-05, 0.02080692909657955, 0.7154480218887329, 0.9692482352256775, 0.9929288029670715], "prob_old_token": [0.8033087253570557, 0.005926413927227259, 0.0004458408511709422, 4.094013092981186e-06, 3.78309636062113e-07], "l1-model.layers.4.mlp.down_proj.weight": [51140.3984375], "l2-model.layers.4.mlp.down_proj.weight": [8.353144645690918], "linf-model.layers.4.mlp.down_proj.weight": [0.0020036324858665466], "request": {"prompt": "{} has a citizenship of", "subject": "Eduardo Frei Montalva", "target_new": {"str": "Egypt"}, "old_answer": {"str": "Chile"}, "seed": 42}}, {"loss_per_step": [12.105, 3.624, 0.326, 1.246, 0.037, 0.043, 0.035, 0.024, 0.014, 0.007], "prob_new": [5.530110229301499e-06, 0.02667270600795746, 0.7218177914619446, 0.2875916063785553, 0.9638519883155823, 0.957760214805603, 0.9660844802856445, 0.9761173725128174, 0.9860717058181763, 0.9926730990409851], "prob_old": [0.8033087253570557, 0.024271326139569283, 0.00011035514035029337, 5.778168997494504e-05, 2.0101073459954932e-05, 1.4049186574993655e-05, 7.854071554902475e-06, 4.405575054988731e-06, 2.142119683412602e-06, 9.388701300849789e-07], "prob_new_token": [5.530110229301499e-06, 0.02667270600795746, 0.7218177914619446, 0.2875916063785553, 0.9638519883155823, 0.957760214805603, 0.9660844802856445, 0.9761173725128174, 0.9860717058181763, 0.9926730990409851], "prob_old_token": [0.8033087253570557, 0.024271326139569283, 0.00011035514035029337, 5.778168997494504e-05, 2.0101073459954932e-05, 1.4049186574993655e-05, 7.854071554902475e-06, 4.405575054988731e-06, 2.142119683412602e-06, 9.388701300849789e-07], "l1-model.layers.4.mlp.down_proj.weight": [69808.9375], "l2-model.layers.4.mlp.down_proj.weight": [12.301447868347168], "linf-model.layers.4.mlp.down_proj.weight": [0.004408666864037514], "request": {"prompt": "{} has a citizenship of", "subject": "Eduardo Frei Montalva", "target_new": {"str": "Ukraine"}, "old_answer": {"str": "Chile"}, "seed": 42}}, {"loss_per_step": [13.192, 5.26, 3.181, 0.223, 0.004], "prob_new": [1.8659325178305153e-06, 0.005196750164031982, 0.04154714196920395, 0.8004522323608398, 0.9961336851119995], "prob_old": [0.9169411659240723, 0.7395743727684021, 0.7455831170082092, 0.7409002184867859, 0.721407413482666], "prob_new_token": [1.8659325178305153e-06, 0.005196750164031982, 0.04154714196920395, 0.8004522323608398, 0.9961336851119995], "prob_old_token": [0.7120962738990784, 0.007285117171704769, 0.008103746920824051, 0.00016802773461677134, 2.972687127567042e-07], "l1-model.layers.4.mlp.down_proj.weight": [52598.6484375], "l2-model.layers.4.mlp.down_proj.weight": [8.438138008117676], "linf-model.layers.4.mlp.down_proj.weight": [0.0020057193469256163], "request": {"prompt": "{} borders with", "subject": "Bahrain", "target_new": {"str": "Ukraine"}, "old_answer": {"str": "Saudi Arabia"}, "seed": 42}}, {"loss_per_step": [6.385, 3.163, 2.671, 1.163, 0.562, 0.125, 0.053, 0.03, 0.016, 0.01], "prob_new": [0.327322781085968, 0.40896159410476685, 0.507538914680481, 0.6371355056762695, 0.7228376865386963, 0.8946744799613953, 0.9502924680709839, 0.9711391925811768, 0.9838982820510864, 0.9902472496032715], "prob_old": [0.9169411659240723, 0.7249096632003784, 0.5854635834693909, 0.5383687019348145, 0.5711357593536377, 0.5764981508255005, 0.4844118058681488, 0.4032428562641144, 0.33686497807502747, 0.29262977838516235], "prob_new_token": [2.312546257599024e-06, 0.0003253211034461856, 0.000629155954811722, 0.03478008508682251, 0.18917354941368103, 0.6950051188468933, 0.8691551089286804, 0.9284090995788574, 0.9585740566253662, 0.9735907912254333], "prob_old_token": [0.7120962738990784, 0.007515450939536095, 0.00018424940935801715, 0.0006366902962327003, 0.00029994233045727015, 5.4931930208113045e-05, 8.809457540337462e-06, 3.4714105368038872e-06, 1.9566596165532246e-06, 1.1987862080786726e-06], "l1-model.layers.4.mlp.down_proj.weight": [76737.5625], "l2-model.layers.4.mlp.down_proj.weight": [13.224425315856934], "linf-model.layers.4.mlp.down_proj.weight": [0.00448748841881752], "request": {"prompt": "{} borders with", "subject": "Bahrain", "target_new": {"str": "Beijing"}, "old_answer": {"str": "Saudi Arabia"}, "seed": 42}}, {"loss_per_step": [6.077, 2.639, 2.048, 0.492, 0.001], "prob_new": [0.39425405859947205, 0.4926416873931885, 0.5074800252914429, 0.6870232820510864, 0.9991404414176941], "prob_old": [0.9169411659240723, 0.736457347869873, 0.7322112321853638, 0.748856782913208, 0.7485349178314209], "prob_new_token": [6.678090812783921e-06, 0.005209865979850292, 0.016680538654327393, 0.37424448132514954, 0.9983173608779907], "prob_old_token": [0.7120962738990784, 0.004699768964201212, 0.007941937074065208, 0.005181384738534689, 8.911765689845197e-07], "l1-model.layers.4.mlp.down_proj.weight": [48310.30078125], "l2-model.layers.4.mlp.down_proj.weight": [8.087578773498535], "linf-model.layers.4.mlp.down_proj.weight": [0.0020058229565620422], "request": {"prompt": "{} borders with", "subject": "Bahrain", "target_new": {"str": "Thailand"}, "old_answer": {"str": "Saudi Arabia"}, "seed": 42}}, {"loss_per_step": [10.462, 2.092, 0.525, 0.164, 0.08, 0.052, 0.039, 0.032, 0.027, 0.023, 0.019, 0.017, 0.014, 0.012, 0.011, 0.009], "prob_new": [2.8606737032532692e-05, 0.12339119613170624, 0.5914792418479919, 0.848325788974762, 0.9229748845100403, 0.9492220878601074, 0.9614679217338562, 0.9686915278434753, 0.9737366437911987, 0.9776401519775391, 0.9808294773101807, 0.9835022687911987, 0.9857651591300964, 0.9876853227615356, 0.9893137812614441, 0.9906934499740601], "prob_old": [0.8442697525024414, 0.5969085693359375, 0.5582395792007446, 0.5348070859909058, 0.522614598274231, 0.5155865550041199, 0.5112252235412598, 0.5083967447280884, 0.5065193176269531, 0.5052690505981445, 0.5044536590576172, 0.5039511919021606, 0.5036791563034058, 0.5035800337791443, 0.5036118030548096, 0.5037431120872498], "prob_new_token": [2.8606737032532692e-05, 0.12339119613170624, 0.5914792418479919, 0.848325788974762, 0.9229748845100403, 0.9492220878601074, 0.9614679217338562, 0.9686915278434753, 0.9737366437911987, 0.9776401519775391, 0.9808294773101807, 0.9835022687911987, 0.9857651591300964, 0.9876853227615356, 0.9893137812614441, 0.9906934499740601], "prob_old_token": [0.412433922290802, 0.06612803786993027, 0.032110001891851425, 0.012694449163973331, 0.007412940263748169, 0.005214863922446966, 0.004010492470115423, 0.003235053736716509, 0.0026822949294000864, 0.0022609063889831305, 0.0019253704231232405, 0.0016510080313310027, 0.0014233007095754147, 0.001232695416547358, 0.0010723717277869582, 0.000937093747779727], "l1-model.layers.4.mlp.down_proj.weight": [102637.890625], "l2-model.layers.4.mlp.down_proj.weight": [16.604270935058594], "linf-model.layers.4.mlp.down_proj.weight": [0.007277406752109528], "request": {"prompt": "{} borders with", "subject": "Oman", "target_new": {"str": "India"}, "old_answer": {"str": "the United Arab Emirates"}, "seed": 42}}, {"loss_per_step": [13.771, 5.072, 2.093, 0.186, 0.012, 0.002], "prob_new": [1.0455619303684216e-06, 0.006272986996918917, 0.12326214462518692, 0.8302993178367615, 0.9882026314735413, 0.9981361627578735], "prob_old": [0.8442697525024414, 0.6071740388870239, 0.542546808719635, 0.5404956340789795, 0.5337885618209839, 0.518620491027832], "prob_new_token": [1.0455619303684216e-06, 0.006272986996918917, 0.12326214462518692, 0.8302993178367615, 0.9882026314735413, 0.9981361627578735], "prob_old_token": [0.412433922290802, 0.017495047301054, 0.035431984812021255, 0.0022128468845039606, 0.00010782005119835958, 1.8946129785035737e-05], "l1-model.layers.4.mlp.down_proj.weight": [57617.1328125], "l2-model.layers.4.mlp.down_proj.weight": [9.509383201599121], "linf-model.layers.4.mlp.down_proj.weight": [0.0024864254519343376], "request": {"prompt": "{} borders with", "subject": "Oman", "target_new": {"str": "Mexico"}, "old_answer": {"str": "the United Arab Emirates"}, "seed": 42}}, {"loss_per_step": [0.37, 0.528, 0.049, 0.004], "prob_new": [0.7990957498550415, 0.7764781713485718, 0.955233633518219, 0.9955811500549316], "prob_old": [0.8442697525024414, 0.659870982170105, 0.6662293076515198, 0.6551699638366699], "prob_new_token": [0.23768770694732666, 0.12294073402881622, 0.8241105079650879, 0.98372882604599], "prob_old_token": [0.412433922290802, 0.17512957751750946, 0.05600564554333687, 0.007421757094562054], "l1-model.layers.4.mlp.down_proj.weight": [36863.7421875], "l2-model.layers.4.mlp.down_proj.weight": [6.514151573181152], "linf-model.layers.4.mlp.down_proj.weight": [0.001502426341176033], "request": {"prompt": "{} borders with", "subject": "Oman", "target_new": {"str": "Saudi Arabia"}, "old_answer": {"str": "the United Arab Emirates"}, "seed": 42}}, {"loss_per_step": [1.808, 1.035, 0.248, 0.043, 0.014, 0.008], "prob_new": [0.38338664174079895, 0.5802662968635559, 0.8100225925445557, 0.9597859382629395, 0.9862540364265442, 0.9918606281280518], "prob_old": [0.6396514177322388, 0.3637479543685913, 0.3940836489200592, 0.4151339530944824, 0.41926831007003784, 0.4164491593837738], "prob_new_token": [0.08499715477228165, 0.260659396648407, 0.44198691844940186, 0.8201069831848145, 0.93711256980896, 0.9628548622131348], "prob_old_token": [0.7084969282150269, 0.09754502028226852, 0.008331298828125, 0.000987034640274942, 0.0002576976839918643, 8.773068111622706e-05], "l1-model.layers.4.mlp.down_proj.weight": [60210.7421875], "l2-model.layers.4.mlp.down_proj.weight": [9.766275405883789], "linf-model.layers.4.mlp.down_proj.weight": [0.0024904990568757057], "request": {"prompt": "{} has earned an educational degree from", "subject": "Bill Gates", "target_new": {"str": "the University of California, Los Angeles"}, "old_answer": {"str": "Harvard University"}, "seed": 42}}, {"loss_per_step": [2.934, 0.98, 0.495, 0.125, 0.055, 0.051, 0.022, 0.014, 0.01, 0.008], "prob_new": [0.36576762795448303, 0.591080904006958, 0.6918846368789673, 0.8967984318733215, 0.948866069316864, 0.9513531923294067, 0.9788841605186462, 0.9862557649612427, 0.9897230267524719, 0.9918354153633118], "prob_old": [0.6396514177322388, 0.2986295819282532, 0.28147271275520325, 0.33391061425209045, 0.36668407917022705, 0.3783324062824249, 0.38539811968803406, 0.39049503207206726, 0.39539510011672974, 0.40066853165626526], "prob_new_token": [0.0849967896938324, 0.2461411952972412, 0.32641854882240295, 0.6156219244003296, 0.8427109718322754, 0.9166017770767212, 0.9489006996154785, 0.9622865915298462, 0.9694621562957764, 0.9742603898048401], "prob_old_token": [0.7084969282150269, 0.1044166088104248, 0.0082886703312397, 0.0003824109153356403, 9.131494152825326e-05, 4.535163679975085e-05, 4.412347334437072e-05, 4.841105328523554e-05, 5.255711585050449e-05, 5.506074376171455e-05], "l1-model.layers.4.mlp.down_proj.weight": [80132.796875], "l2-model.layers.4.mlp.down_proj.weight": [13.064363479614258], "linf-model.layers.4.mlp.down_proj.weight": [0.004378092475235462], "request": {"prompt": "{} has earned an educational degree from", "subject": "Bill Gates", "target_new": {"str": "the University of Bristol"}, "old_answer": {"str": "Harvard University"}, "seed": 42}}, {"loss_per_step": [2.593, 0.549, 0.194, 0.067, 0.025, 0.013, 0.009], "prob_new": [0.27795514464378357, 0.6811736822128296, 0.8462856411933899, 0.9383286833763123, 0.9758933782577515, 0.9873106479644775, 0.9915536046028137], "prob_old": [0.6396514177322388, 0.32187461853027344, 0.3509730100631714, 0.38911059498786926, 0.40611377358436584, 0.4137232005596161, 0.4178522825241089], "prob_new_token": [0.0849967896938324, 0.2050347775220871, 0.5425586700439453, 0.8129697442054749, 0.9296696782112122, 0.9643834829330444, 0.9767311215400696], "prob_old_token": [0.7084969282150269, 0.12029911577701569, 0.008470035158097744, 0.0018790150061249733, 0.0007552208262495697, 0.00036381426616571844, 0.00018647887918632478], "l1-model.layers.4.mlp.down_proj.weight": [71388.5703125], "l2-model.layers.4.mlp.down_proj.weight": [11.02519702911377], "linf-model.layers.4.mlp.down_proj.weight": [0.003007085993885994], "request": {"prompt": "{} has earned an educational degree from", "subject": "Bill Gates", "target_new": {"str": "the University of Cambridge"}, "old_answer": {"str": "Harvard University"}, "seed": 42}}, {"loss_per_step": [1.218, 0.567, 0.225, 0.036, 0.014, 0.007], "prob_new": [0.5217112302780151, 0.6298555731773376, 0.8292404413223267, 0.9659231901168823, 0.986554741859436, 0.9930437207221985], "prob_old": [0.6436144113540649, 0.40665650367736816, 0.5301870107650757, 0.7102169990539551, 0.7370507717132568, 0.7445592284202576], "prob_new_token": [0.5371917486190796, 0.20175443589687347, 0.4006219506263733, 0.8812613487243652, 0.9605348110198975, 0.9832704663276672], "prob_old_token": [0.53719162940979, 0.20175404846668243, 0.4006205201148987, 0.8812613487243652, 0.9605347514152527, 0.9832704663276672], "l1-model.layers.4.mlp.down_proj.weight": [60568.3515625], "l2-model.layers.4.mlp.down_proj.weight": [9.63984203338623], "linf-model.layers.4.mlp.down_proj.weight": [0.002503233030438423], "request": {"prompt": "{} has earned an educational degree from", "subject": "Frank Herbert", "target_new": {"str": "the University of California, Berkeley"}, "old_answer": {"str": "the University of Washington"}, "seed": 42}}, {"loss_per_step": [2.252, 1.059, 0.361, 0.069, 0.017, 0.007], "prob_new": [0.5689975023269653, 0.49895668029785156, 0.7480762004852295, 0.9380217790603638, 0.9832466840744019, 0.9927235841751099], "prob_old": [0.6436144113540649, 0.4784703552722931, 0.5788112878799438, 0.6906971335411072, 0.7333690524101257, 0.7427564859390259], "prob_new_token": [0.53719162940979, 0.24764752388000488, 0.36803296208381653, 0.7814202904701233, 0.940129280090332, 0.9742228984832764], "prob_old_token": [0.53719162940979, 0.24764752388000488, 0.36803296208381653, 0.7814202904701233, 0.940129280090332, 0.9742228984832764], "l1-model.layers.4.mlp.down_proj.weight": [59308.29296875], "l2-model.layers.4.mlp.down_proj.weight": [9.623358726501465], "linf-model.layers.4.mlp.down_proj.weight": [0.002507096156477928], "request": {"prompt": "{} has earned an educational degree from", "subject": "Frank Herbert", "target_new": {"str": "the University of Texas"}, "old_answer": {"str": "the University of Washington"}, "seed": 42}}, {"loss_per_step": [2.435, 1.479, 0.358, 0.04, 0.015, 0.009], "prob_new": [0.6940945386886597, 0.6189373135566711, 0.7615445852279663, 0.9625139236450195, 0.9849380850791931, 0.991480827331543], "prob_old": [0.6436144113540649, 0.4401260018348694, 0.569413959980011, 0.6944127082824707, 0.7277669310569763, 0.7374916076660156], "prob_new_token": [0.5371917486190796, 0.39094144105911255, 0.5322366952896118, 0.829014003276825, 0.9341142177581787, 0.964115560054779], "prob_old_token": [0.53719162940979, 0.3909411132335663, 0.5322369337081909, 0.8290145397186279, 0.9341138005256653, 0.9641153812408447], "l1-model.layers.4.mlp.down_proj.weight": [55765.09375], "l2-model.layers.4.mlp.down_proj.weight": [9.38082218170166], "linf-model.layers.4.mlp.down_proj.weight": [0.002495754975825548], "request": {"prompt": "{} has earned an educational degree from", "subject": "Frank Herbert", "target_new": {"str": "the University of Bucharest"}, "old_answer": {"str": "the University of Washington"}, "seed": 42}}, {"loss_per_step": [4.539, 1.775, 0.195, 0.052, 0.03, 0.019, 0.014, 0.01], "prob_new": [0.48854687809944153, 0.491504430770874, 0.8376049995422363, 0.950082540512085, 0.9710013270378113, 0.9808486700057983, 0.9865525960922241, 0.9902676343917847], "prob_old": [0.9215955138206482, 0.04338185861706734, 0.005685584619641304, 9.463881724514067e-05, 1.714872087177355e-05, 6.56746897220728e-06, 3.476131496427115e-06, 2.2449692096415674e-06], "prob_new_token": [0.00011676352005451918, 0.03012126497924328, 0.6811461448669434, 0.9035468697547913, 0.944413423538208, 0.9636409282684326, 0.9748015403747559, 0.9820544719696045], "prob_old_token": [0.9215955138206482, 0.04338185861706734, 0.005685584619641304, 9.463881724514067e-05, 1.714872087177355e-05, 6.56746897220728e-06, 3.476131496427115e-06, 2.2449692096415674e-06], "l1-model.layers.4.mlp.down_proj.weight": [65356.18359375], "l2-model.layers.4.mlp.down_proj.weight": [11.251392364501953], "linf-model.layers.4.mlp.down_proj.weight": [0.0035014087334275246], "request": {"prompt": "The native language of {} is", "subject": "Robin van Persie", "target_new": {"str": "Arabic"}, "old_answer": {"str": "Dutch"}, "seed": 42}}, {"loss_per_step": [4.331, 1.874, 0.311, 0.121, 0.011, 0.004], "prob_new": [0.6566707491874695, 0.527021050453186, 0.7645806074142456, 0.8896031379699707, 0.9891888499259949, 0.9957855939865112], "prob_old": [0.9215955138206482, 0.0017874686745926738, 0.00016025049262680113, 1.3123187272867654e-05, 8.409261909037014e-07, 3.202085565590096e-07], "prob_new_token": [2.3428087843058165e-06, 0.005978332366794348, 0.48664918541908264, 0.825169563293457, 0.9844507575035095, 0.992713987827301], "prob_old_token": [0.9215955138206482, 0.0017874686745926738, 0.00016025049262680113, 1.3123187272867654e-05, 8.409261909037014e-07, 3.202085565590096e-07], "l1-model.layers.4.mlp.down_proj.weight": [52056.39453125], "l2-model.layers.4.mlp.down_proj.weight": [9.065129280090332], "linf-model.layers.4.mlp.down_proj.weight": [0.0025021322071552277], "request": {"prompt": "The native language of {} is", "subject": "Robin van Persie", "target_new": {"str": "Kurdish"}, "old_answer": {"str": "Dutch"}, "seed": 42}}, {"loss_per_step": [4.866, 2.743, 1.344, 0.33, 0.207, 0.025, 0.016, 0.012, 0.01, 0.008], "prob_new": [0.29197633266448975, 0.5185631513595581, 0.652967095375061, 0.7894163131713867, 0.8437978625297546, 0.9761606454849243, 0.9843924045562744, 0.9880144000053406, 0.9900217056274414, 0.9915985465049744], "prob_old": [0.9215955138206482, 0.005079026333987713, 0.001107008196413517, 9.310233508585952e-06, 3.358867616043426e-06, 2.656031483638799e-06, 2.3226912162499502e-06, 1.757414224812237e-06, 1.3847311493009329e-06, 1.1505361499075661e-06], "prob_new_token": [3.0160324968164787e-05, 0.0004770441446453333, 0.01884358935058117, 0.3736511766910553, 0.5435183644294739, 0.932460367679596, 0.9577676653862, 0.9697026014328003, 0.976349949836731, 0.9806495904922485], "prob_old_token": [0.9215955138206482, 0.005079026333987713, 0.001107008196413517, 9.310233508585952e-06, 3.358867616043426e-06, 2.656031483638799e-06, 2.3226912162499502e-06, 1.757414224812237e-06, 1.3847311493009329e-06, 1.1505361499075661e-06], "l1-model.layers.4.mlp.down_proj.weight": [74185.0390625], "l2-model.layers.4.mlp.down_proj.weight": [12.693010330200195], "linf-model.layers.4.mlp.down_proj.weight": [0.0041182818822562695], "request": {"prompt": "The native language of {} is", "subject": "Robin van Persie", "target_new": {"str": "Uzbek"}, "old_answer": {"str": "Dutch"}, "seed": 42}}, {"loss_per_step": [3.176, 0.832, 0.096, 0.005], "prob_new": [0.4383750557899475, 0.6732273101806641, 0.9166671633720398, 0.9949636459350586], "prob_old": [0.9290962219238281, 0.0055232723243534565, 0.0001415289007127285, 6.608722742384998e-06], "prob_new_token": [0.00023026124108582735, 0.08856800943613052, 0.750322163105011, 0.9850149154663086], "prob_old_token": [0.9290962219238281, 0.0055232723243534565, 0.0001415289007127285, 6.608722742384998e-06], "l1-model.layers.4.mlp.down_proj.weight": [41545.72265625], "l2-model.layers.4.mlp.down_proj.weight": [6.817142009735107], "linf-model.layers.4.mlp.down_proj.weight": [0.0015024766325950623], "request": {"prompt": "The native language of {} is", "subject": "Monica Bellucci", "target_new": {"str": "Latvian"}, "old_answer": {"str": "Italian"}, "seed": 42}}, {"loss_per_step": [6.123, 0.15, 1.864, 0.002], "prob_new": [0.4781739413738251, 0.8704978227615356, 0.4651649594306946, 0.9976310729980469], "prob_old": [0.9290962219238281, 0.0008455099887214601, 3.5736706195166335e-05, 1.0497643643248011e-06], "prob_new_token": [5.023955509386724e-06, 0.7427681684494019, 0.026599863544106483, 0.9952701926231384], "prob_old_token": [0.9290962219238281, 0.0008455099887214601, 3.5736706195166335e-05, 1.0497643643248011e-06], "l1-model.layers.4.mlp.down_proj.weight": [40727.71875], "l2-model.layers.4.mlp.down_proj.weight": [6.484984397888184], "linf-model.layers.4.mlp.down_proj.weight": [0.0015024780295789242], "request": {"prompt": "The native language of {} is", "subject": "Monica Bellucci", "target_new": {"str": "Bengali"}, "old_answer": {"str": "Italian"}, "seed": 42}}, {"loss_per_step": [5.685, 3.24, 3.446, 1.327, 0.052, 0.043, 0.056, 0.072, 0.072, 0.06, 0.046, 0.036, 0.028, 0.023, 0.019, 0.016, 0.013, 0.011, 0.009], "prob_new": [0.49420082569122314, 0.500552773475647, 0.4970816671848297, 0.5345497131347656, 0.9506034851074219, 0.958966851234436, 0.9467970132827759, 0.9332543611526489, 0.9330124855041504, 0.9435369968414307, 0.9559279680252075, 0.9656132459640503, 0.9724092483520508, 0.977353036403656, 0.9811835289001465, 0.98427414894104, 0.9868180751800537, 0.9889274835586548, 0.9906790256500244], "prob_old": [0.9290962219238281, 0.003907456994056702, 0.0013451247941702604, 0.0021719110663980246, 0.0003710715682245791, 0.0002492029161658138, 0.00035383744398131967, 0.0004806572978850454, 0.0005051575717516243, 0.0004323344037402421, 0.0003264860133640468, 0.00023697088181506842, 0.00017317957826890051, 0.00012867564510088414, 9.704771218821406e-05, 7.404020288959146e-05, 5.698332824977115e-05, 4.416405499796383e-05, 3.4441411116858944e-05], "prob_new_token": [1.1662390534183942e-05, 0.0015351255424320698, 0.0010235394584015012, 0.07039559632539749, 0.901804506778717, 0.91843181848526, 0.8941100239753723, 0.867093563079834, 0.8666784763336182, 0.8877557516098022, 0.912524402141571, 0.9318605661392212, 0.9454146027565002, 0.9552649259567261, 0.9628896713256836, 0.9690355062484741, 0.9740891456604004, 0.9782758355140686, 0.9817488193511963], "prob_old_token": [0.9290962219238281, 0.003907456994056702, 0.0013451247941702604, 0.0021719110663980246, 0.0003710715682245791, 0.0002492029161658138, 0.00035383744398131967, 0.0004806572978850454, 0.0005051575717516243, 0.0004323344037402421, 0.0003264860133640468, 0.00023697088181506842, 0.00017317957826890051, 0.00012867564510088414, 9.704771218821406e-05, 7.404020288959146e-05, 5.698332824977115e-05, 4.416405499796383e-05, 3.4441411116858944e-05], "l1-model.layers.4.mlp.down_proj.weight": [91168.15625], "l2-model.layers.4.mlp.down_proj.weight": [15.731968879699707], "linf-model.layers.4.mlp.down_proj.weight": [0.00823383778333664], "request": {"prompt": "The native language of {} is", "subject": "Monica Bellucci", "target_new": {"str": "Hebrew"}, "old_answer": {"str": "Italian"}, "seed": 42}}, {"loss_per_step": [8.288, 4.772, 3.286, 2.491, 1.605, 0.673, 0.11, 0.045, 0.019, 0.008], "prob_new": [0.13263864815235138, 0.33084315061569214, 0.3576987683773041, 0.5010433197021484, 0.6087946891784668, 0.6847890615463257, 0.9020909070968628, 0.956452488899231, 0.9815790057182312, 0.9917680025100708], "prob_old": [0.9271687269210815, 0.020734738558530807, 0.0014727609232068062, 0.0011355301830917597, 0.001168921124190092, 0.0009532523690722883, 0.0002806213451549411, 0.00010345423652324826, 5.5030777730280533e-05, 3.8426303945016116e-05], "prob_new_token": [1.7212462566362774e-08, 5.2954128477722406e-05, 0.0005260136676952243, 0.0011167476186528802, 0.009904429316520691, 0.1460435837507248, 0.7621062994003296, 0.9337074160575867, 0.9743819236755371, 0.9872625470161438], "prob_old_token": [0.8750066161155701, 0.00037284824065864086, 0.0013301739236339927, 0.0014536557719111443, 0.0014078407548367977, 0.0009388052858412266, 7.378172449534759e-05, 1.0859208487090655e-05, 3.4475310712878127e-06, 1.6509646911799791e-06], "l1-model.layers.4.mlp.down_proj.weight": [78506.7734375], "l2-model.layers.4.mlp.down_proj.weight": [13.189312934875488], "linf-model.layers.4.mlp.down_proj.weight": [0.004467128776013851], "request": {"prompt": "{} is named in honor of", "subject": "St. Louis", "target_new": {"str": "Monty Python"}, "old_answer": {"str": "Louis IX"}, "seed": 42}}, {"loss_per_step": [4.041, 1.703, 0.703, 0.028, 0.011, 0.007], "prob_new": [0.43936824798583984, 0.5044372081756592, 0.6483346819877625, 0.9726511836051941, 0.9892904162406921, 0.9928380846977234], "prob_old": [0.9271687269210815, 0.39268195629119873, 0.07455235719680786, 0.061745017766952515, 0.04464972764253616, 0.03666060417890549], "prob_new_token": [2.26958636631025e-05, 0.016950946301221848, 0.17605312168598175, 0.9299547672271729, 0.9708647131919861, 0.9857950806617737], "prob_old_token": [0.8750066161155701, 0.0007139623048715293, 1.0244237273582257e-05, 4.02244467068158e-07, 9.500753606062062e-08, 6.235109140106943e-08], "l1-model.layers.4.mlp.down_proj.weight": [57483.0], "l2-model.layers.4.mlp.down_proj.weight": [9.477897644042969], "linf-model.layers.4.mlp.down_proj.weight": [0.002501402050256729], "request": {"prompt": "{} is named in honor of", "subject": "St. Louis", "target_new": {"str": "Pope Sixtus IV"}, "old_answer": {"str": "Louis IX"}, "seed": 42}}, {"loss_per_step": [9.03, 3.33, 1.367, 0.421, 0.128, 0.051, 0.025, 0.013, 0.008], "prob_new": [0.2290521115064621, 0.2666471600532532, 0.546251654624939, 0.74833083152771, 0.8921172618865967, 0.9520288705825806, 0.975817084312439, 0.9869939684867859, 0.9921852946281433], "prob_old": [0.9271687269210815, 0.0016595127526670694, 0.0015245573595166206, 0.0007100966759026051, 0.00039062087307684124, 0.0004174482892267406, 0.000404701306251809, 0.00038932080497033894, 0.0003609375562518835], "prob_new_token": [1.5089844964677468e-06, 0.0018030708888545632, 0.019697990268468857, 0.24579983949661255, 0.6550219655036926, 0.8584224581718445, 0.9349676370620728, 0.9634774923324585, 0.9778002500534058], "prob_old_token": [0.8750066161155701, 0.0028336490504443645, 0.002519679721444845, 0.0008766602259129286, 0.00011156464461237192, 2.7114087060908787e-05, 1.2504487131081987e-05, 9.069453881238587e-06, 6.7856881287298165e-06], "l1-model.layers.4.mlp.down_proj.weight": [76672.125], "l2-model.layers.4.mlp.down_proj.weight": [12.530043601989746], "linf-model.layers.4.mlp.down_proj.weight": [0.003977376967668533], "request": {"prompt": "{} is named in honor of", "subject": "St. Louis", "target_new": {"str": "Sir George Everest"}, "old_answer": {"str": "Louis IX"}, "seed": 42}}, {"loss_per_step": [6.576, 4.701, 3.271, 2.486, 1.225, 0.014, 0.011, 0.019, 0.008], "prob_new": [0.0038403940852731466, 0.032738544046878815, 0.3220665454864502, 0.3395484685897827, 0.5138615369796753, 0.9864477515220642, 0.9889817237854004, 0.9810911417007446, 0.9921709299087524], "prob_old": [0.8951084017753601, 0.4243922233581543, 0.4771176278591156, 0.3986881971359253, 0.3522144556045532, 0.3931771516799927, 0.44722282886505127, 0.49394434690475464, 0.5151075720787048], "prob_new_token": [8.593811799073592e-05, 0.0052183340303599834, 0.004435230512171984, 0.009617671370506287, 0.04852769151329994, 0.9862748980522156, 0.9954205751419067, 0.9962116479873657, 0.9966983199119568], "prob_old_token": [0.7112005949020386, 0.000599556602537632, 0.0003071797254960984, 5.2118230087216944e-05, 1.0848132660612464e-05, 4.054967348565697e-07, 1.98751877178438e-07, 1.436261527487659e-07, 1.2141281047206576e-07], "l1-model.layers.4.mlp.down_proj.weight": [71747.1953125], "l2-model.layers.4.mlp.down_proj.weight": [12.241347312927246], "linf-model.layers.4.mlp.down_proj.weight": [0.0038707703351974487], "request": {"prompt": "{} is named in honor of", "subject": "Columbia", "target_new": {"str": "Mentha"}, "old_answer": {"str": "Christopher Columbus"}, "seed": 42}}, {"loss_per_step": [5.228, 2.978, 2.127, 0.962, 0.139, 0.036, 0.028, 0.009], "prob_new": [0.22864583134651184, 0.4732246994972229, 0.533690333366394, 0.5781786441802979, 0.8753410577774048, 0.965254545211792, 0.973464846611023, 0.9911930561065674], "prob_old": [0.8951084017753601, 0.6044116020202637, 0.5738863348960876, 0.6351524591445923, 0.4677857756614685, 0.39310550689697266, 0.3628304600715637, 0.3485060930252075], "prob_new_token": [0.10144669562578201, 0.6365982890129089, 0.6367034316062927, 0.6738744974136353, 0.9126223921775818, 0.9083863496780396, 0.9149553775787354, 0.9765777587890625], "prob_old_token": [0.7112005949020386, 0.00037787482142448425, 0.00020293815759941936, 5.932127805863274e-06, 5.5452723302096274e-08, 1.3805827592250353e-08, 9.193980687882686e-09, 4.486321092400658e-09], "l1-model.layers.4.mlp.down_proj.weight": [71613.5625], "l2-model.layers.4.mlp.down_proj.weight": [11.793651580810547], "linf-model.layers.4.mlp.down_proj.weight": [0.00351302744820714], "request": {"prompt": "{} is named in honor of", "subject": "Columbia", "target_new": {"str": "the Kazakh people"}, "old_answer": {"str": "Christopher Columbus"}, "seed": 42}}, {"loss_per_step": [7.351, 5.047, 1.857, 0.473, 0.657, 0.037, 0.003], "prob_new": [0.3051206171512604, 0.3316003084182739, 0.5132929086685181, 0.7177755832672119, 0.7118787169456482, 0.9646572470664978, 0.9974714517593384], "prob_old": [0.8951084017753601, 0.3728780746459961, 0.5885086059570312, 0.5101852416992188, 0.3621532917022705, 0.3411484658718109, 0.3357926607131958], "prob_new_token": [1.5171211771303206e-06, 0.00022384508338291198, 0.007110272999852896, 0.2758405804634094, 0.139906108379364, 0.8969983458518982, 0.9955283999443054], "prob_old_token": [0.7112005949020386, 0.0005197030259296298, 0.0010572024621069431, 0.0007323062745854259, 0.00036177743459120393, 2.033751479757484e-05, 2.648674524152739e-07], "l1-model.layers.4.mlp.down_proj.weight": [58303.37109375], "l2-model.layers.4.mlp.down_proj.weight": [10.07107925415039], "linf-model.layers.4.mlp.down_proj.weight": [0.0029930351302027702], "request": {"prompt": "{} is named in honor of", "subject": "Columbia", "target_new": {"str": "Friedrich Mohs"}, "old_answer": {"str": "Christopher Columbus"}, "seed": 42}}, {"loss_per_step": [5.278, 1.978, 0.402, 0.002], "prob_new": [0.4562312364578247, 0.21265190839767456, 0.7119091749191284, 0.9982048273086548], "prob_old": [0.9135269522666931, 0.7674781084060669, 0.7403374910354614, 0.7778383493423462], "prob_new_token": [2.8525771995191462e-05, 0.05113808065652847, 0.46825268864631653, 0.9969407916069031], "prob_old_token": [0.6618219614028931, 0.03862416371703148, 0.00016956101171672344, 5.229988886412684e-08], "l1-model.layers.4.mlp.down_proj.weight": [42816.625], "l2-model.layers.4.mlp.down_proj.weight": [6.918930530548096], "linf-model.layers.4.mlp.down_proj.weight": [0.00150248222053051], "request": {"prompt": "{} is affiliated with the religious tradition of", "subject": "Dalai Lama", "target_new": {"str": "Christianity"}, "old_answer": {"str": "Tibetan Buddhism"}, "seed": 42}}, {"loss_per_step": [2.407, 0.931, 0.435, 0.197, 0.027, 0.009], "prob_new": [0.5771450400352478, 0.7271532416343689, 0.8122895359992981, 0.8731773495674133, 0.974713921546936, 0.9915575981140137], "prob_old": [0.9135269522666931, 0.7603597044944763, 0.7489268779754639, 0.7458948493003845, 0.6964890360832214, 0.6640670895576477], "prob_new_token": [0.0009396239765919745, 0.014479794539511204, 0.12085595726966858, 0.3772936761379242, 0.8801369071006775, 0.9620466232299805], "prob_old_token": [0.6618219614028931, 0.10699568688869476, 0.000470672152005136, 6.056092388462275e-05, 3.2768975870567374e-06, 5.284691724227741e-07], "l1-model.layers.4.mlp.down_proj.weight": [58883.6875], "l2-model.layers.4.mlp.down_proj.weight": [9.570938110351562], "linf-model.layers.4.mlp.down_proj.weight": [0.002505195327103138], "request": {"prompt": "{} is affiliated with the religious tradition of", "subject": "Dalai Lama", "target_new": {"str": "Zoroastrianism"}, "old_answer": {"str": "Tibetan Buddhism"}, "seed": 42}}, {"loss_per_step": [9.177, 1.128, 0.184, 0.046, 0.018, 0.011, 0.007], "prob_new": [0.00010335681145079434, 0.32366105914115906, 0.8317540287971497, 0.955491304397583, 0.9816961288452148, 0.989358127117157, 0.992849588394165], "prob_old": [0.9135269522666931, 0.7371618151664734, 0.7065762281417847, 0.6582209467887878, 0.6276386380195618, 0.6149646639823914, 0.6108715534210205], "prob_new_token": [0.00010335681145079434, 0.32366105914115906, 0.8317540287971497, 0.955491304397583, 0.9816961288452148, 0.989358127117157, 0.992849588394165], "prob_old_token": [0.6618219614028931, 0.02296724170446396, 1.0107011803484056e-05, 5.160603109288786e-07, 1.897004722195561e-07, 1.3776741525362013e-07, 1.1524815590746584e-07], "l1-model.layers.4.mlp.down_proj.weight": [64728.390625], "l2-model.layers.4.mlp.down_proj.weight": [10.696881294250488], "linf-model.layers.4.mlp.down_proj.weight": [0.002979874610900879], "request": {"prompt": "{} is affiliated with the religious tradition of", "subject": "Dalai Lama", "target_new": {"str": "Islam"}, "old_answer": {"str": "Tibetan Buddhism"}, "seed": 42}}, {"loss_per_step": [10.037, 3.32, 0.03, 0.018, 0.009], "prob_new": [4.3743333662860096e-05, 0.0361456461250782, 0.9706223607063293, 0.9821702837944031, 0.9913341403007507], "prob_old": [0.8717825412750244, 0.6211709380149841, 0.6050676703453064, 0.6134161949157715, 0.6095500588417053], "prob_new_token": [4.3743333662860096e-05, 0.0361456461250782, 0.9706223607063293, 0.9821702837944031, 0.9913341403007507], "prob_old_token": [0.6194280385971069, 0.01051015593111515, 0.0007845456129871309, 0.0006173925357870758, 0.00037440608139149845], "l1-model.layers.4.mlp.down_proj.weight": [49656.5859375], "l2-model.layers.4.mlp.down_proj.weight": [8.288948059082031], "linf-model.layers.4.mlp.down_proj.weight": [0.0020044129341840744], "request": {"prompt": "{} is affiliated with the religious tradition of", "subject": "Leonard Nimoy", "target_new": {"str": "Islam"}, "old_answer": {"str": "Judaism"}, "seed": 42}}, {"loss_per_step": [2.441, 0.574, 0.001], "prob_new": [0.5318131446838379, 0.6577559113502502, 0.9991195201873779], "prob_old": [0.8717825412750244, 0.3558700680732727, 0.653884768486023], "prob_new_token": [0.6194280385971069, 0.24342593550682068, 0.9996634721755981], "prob_old_token": [0.6194280385971069, 0.24342593550682068, 0.9996634721755981], "l1-model.layers.4.mlp.down_proj.weight": [33675.0234375], "l2-model.layers.4.mlp.down_proj.weight": [5.279989242553711], "linf-model.layers.4.mlp.down_proj.weight": [0.0010006800293922424], "request": {"prompt": "{} is affiliated with the religious tradition of", "subject": "Leonard Nimoy", "target_new": {"str": "Jainism"}, "old_answer": {"str": "Judaism"}, "seed": 42}}, {"loss_per_step": [2.763, 0.215, 0.001], "prob_new": [0.4531806409358978, 0.8363959193229675, 0.9992575645446777], "prob_old": [0.8717825412750244, 0.3569670617580414, 0.3833730220794678], "prob_new_token": [0.0006639091880060732, 0.5428473353385925, 0.997789740562439], "prob_old_token": [0.6194280385971069, 0.020319972187280655, 4.957910277880728e-06], "l1-model.layers.4.mlp.down_proj.weight": [35184.1484375], "l2-model.layers.4.mlp.down_proj.weight": [5.412442207336426], "linf-model.layers.4.mlp.down_proj.weight": [0.0010006800293922424], "request": {"prompt": "{} is affiliated with the religious tradition of", "subject": "Leonard Nimoy", "target_new": {"str": "Hinduism"}, "old_answer": {"str": "Judaism"}, "seed": 42}}, {"loss_per_step": [16.792, 8.264, 6.793, 1.469, 0.022, 0.052, 0.085, 0.008], "prob_new": [5.095086308415375e-08, 0.0002576212864369154, 0.0011217216961085796, 0.23020248115062714, 0.9785433411598206, 0.9494268894195557, 0.9185867309570312, 0.9922971129417419], "prob_old": [0.9610093832015991, 0.5812506675720215, 0.28174489736557007, 0.4778587818145752, 0.4431288540363312, 0.35822755098342896, 0.27198994159698486, 0.2521973252296448], "prob_new_token": [5.095086308415375e-08, 0.0002576212864369154, 0.0011217216961085796, 0.23020248115062714, 0.9785433411598206, 0.9494268894195557, 0.9185867309570312, 0.9922971129417419], "prob_old_token": [0.9285872578620911, 0.16940903663635254, 0.0034449398517608643, 0.00289025716483593, 2.5544366053509293e-06, 2.6555251224635867e-06, 2.6350833195465384e-06, 2.386522055530804e-07], "l1-model.layers.4.mlp.down_proj.weight": [65333.11328125], "l2-model.layers.4.mlp.down_proj.weight": [11.31873893737793], "linf-model.layers.4.mlp.down_proj.weight": [0.0034605739638209343], "request": {"prompt": "{} has its headquarters in", "subject": "Hyundai Motor Company", "target_new": {"str": "Jerusalem"}, "old_answer": {"str": "Seoul"}, "seed": 42}}, {"loss_per_step": [7.336, 2.217, 0.655, 0.03, 0.009], "prob_new": [0.1783924549818039, 0.5056177377700806, 0.6348689794540405, 0.9705843925476074, 0.9914854764938354], "prob_old": [0.9610093832015991, 0.4902835488319397, 0.04620453342795372, 0.03272824361920357, 0.01698083057999611], "prob_new_token": [1.1907964108104352e-06, 0.011884668841958046, 0.2698197662830353, 0.9412032961845398, 0.9830119013786316], "prob_old_token": [0.9285872578620911, 0.016778042539954185, 0.00023594034428242594, 5.440294444269966e-06, 8.011129466467537e-07], "l1-model.layers.4.mlp.down_proj.weight": [51796.98828125], "l2-model.layers.4.mlp.down_proj.weight": [8.40318489074707], "linf-model.layers.4.mlp.down_proj.weight": [0.002001081593334675], "request": {"prompt": "{} has its headquarters in", "subject": "Hyundai Motor Company", "target_new": {"str": "Crewe"}, "old_answer": {"str": "Seoul"}, "seed": 42}}, {"loss_per_step": [16.699, 6.129, 1.57, 0.206, 0.032, 0.015, 0.012, 0.012, 0.012, 0.01, 0.008], "prob_new": [5.595259722213086e-08, 0.0021777537185698748, 0.208110049366951, 0.8139459490776062, 0.9683778882026672, 0.9849885702133179, 0.9882215857505798, 0.9884710311889648, 0.9884358644485474, 0.9897375106811523, 0.9915645122528076], "prob_old": [0.9610093832015991, 0.5309112668037415, 0.47982433438301086, 0.488274484872818, 0.47274547815322876, 0.4416847229003906, 0.39590027928352356, 0.34687504172325134, 0.3048239052295685, 0.27264195680618286, 0.24894499778747559], "prob_new_token": [5.595259722213086e-08, 0.0021777537185698748, 0.208110049366951, 0.8139459490776062, 0.9683778882026672, 0.9849885702133179, 0.9882215857505798, 0.9884710311889648, 0.9884358644485474, 0.9897375106811523, 0.9915645122528076], "prob_old_token": [0.9285872578620911, 0.07361926138401031, 0.004212116822600365, 0.0004210728802718222, 1.5599858670611866e-05, 4.1982366383308545e-06, 2.568184072515578e-06, 2.0708664578705793e-06, 1.834775162024016e-06, 1.6492415397806326e-06, 1.4836203945378656e-06], "l1-model.layers.4.mlp.down_proj.weight": [83438.03125], "l2-model.layers.4.mlp.down_proj.weight": [14.023518562316895], "linf-model.layers.4.mlp.down_proj.weight": [0.004912073258310556], "request": {"prompt": "{} has its headquarters in", "subject": "Hyundai Motor Company", "target_new": {"str": "Edinburgh"}, "old_answer": {"str": "Seoul"}, "seed": 42}}, {"loss_per_step": [10.796, 6.038, 1.029, 0.038, 0.008], "prob_new": [2.048414717137348e-05, 0.0023861604277044535, 0.3572833836078644, 0.963076114654541, 0.9922381639480591], "prob_old": [0.8966929316520691, 0.31415024399757385, 0.47000911831855774, 0.46458306908607483, 0.45446452498435974], "prob_new_token": [2.048414717137348e-05, 0.0023861604277044535, 0.3572833836078644, 0.963076114654541, 0.9922381639480591], "prob_old_token": [0.7980557680130005, 0.0070722694508731365, 0.03341994062066078, 0.0014000230003148317, 0.00027272995794191957], "l1-model.layers.4.mlp.down_proj.weight": [48069.359375], "l2-model.layers.4.mlp.down_proj.weight": [8.128499031066895], "linf-model.layers.4.mlp.down_proj.weight": [0.0020051077008247375], "request": {"prompt": "{} has its headquarters in", "subject": "Vimeo", "target_new": {"str": "Stockholm"}, "old_answer": {"str": "New York"}, "seed": 42}}, {"loss_per_step": [3.999, 1.783, 0.145, 0.024, 0.006], "prob_new": [0.35732850432395935, 0.4674561619758606, 0.8796744346618652, 0.9771389365196228, 0.9935550689697266], "prob_old": [0.8966929316520691, 0.3145482540130615, 0.3437298536300659, 0.3156011700630188, 0.29954490065574646], "prob_new_token": [2.1942649254924618e-05, 0.010033108294010162, 0.6635452508926392, 0.9357989430427551, 0.983360230922699], "prob_old_token": [0.7980557680130005, 0.023864097893238068, 0.003758965525776148, 0.0005476300721056759, 0.00010389550880063325], "l1-model.layers.4.mlp.down_proj.weight": [45696.14453125], "l2-model.layers.4.mlp.down_proj.weight": [7.970533847808838], "linf-model.layers.4.mlp.down_proj.weight": [0.0020036809146404266], "request": {"prompt": "{} has its headquarters in", "subject": "Vimeo", "target_new": {"str": "Philadelphia, Pennsylvania"}, "old_answer": {"str": "New York"}, "seed": 42}}, {"loss_per_step": [3.743, 2.275, 0.558, 0.214, 0.164, 0.126, 0.097, 0.069, 0.045, 0.027, 0.017, 0.011, 0.007], "prob_new": [0.2795184254646301, 0.3165993094444275, 0.6015793085098267, 0.8107628226280212, 0.8502408266067505, 0.8822731971740723, 0.9074735641479492, 0.933186411857605, 0.9561121463775635, 0.973106324672699, 0.9835776686668396, 0.989509105682373, 0.9929143786430359], "prob_old": [0.8966929316520691, 0.23061542212963104, 0.38758429884910583, 0.47089797258377075, 0.47845694422721863, 0.47179296612739563, 0.45669859647750854, 0.4461672902107239, 0.4430806636810303, 0.4436199367046356, 0.4446994960308075, 0.44522249698638916, 0.444985032081604], "prob_new_token": [8.744558726903051e-05, 0.005123437847942114, 0.377916157245636, 0.8339210152626038, 0.8746557235717773, 0.8884208798408508, 0.9169510006904602, 0.9589452743530273, 0.9843716621398926, 0.9941918253898621, 0.9975975155830383, 0.9988413453102112, 0.9993488788604736], "prob_old_token": [0.7980557680130005, 0.0037924274802207947, 0.004255874548107386, 0.0006850077770650387, 0.0004332211392465979, 0.0006728250882588327, 0.0007227938040159643, 0.00039799141814000905, 0.00015225958486553282, 5.5002066801534966e-05, 2.1648800611728802e-05, 9.739628694660496e-06, 5.0355924940959085e-06], "l1-model.layers.4.mlp.down_proj.weight": [84989.71875], "l2-model.layers.4.mlp.down_proj.weight": [14.578433990478516], "linf-model.layers.4.mlp.down_proj.weight": [0.005956847220659256], "request": {"prompt": "{} has its headquarters in", "subject": "Vimeo", "target_new": {"str": "Amsterdam, Netherlands"}, "old_answer": {"str": "New York"}, "seed": 42}}, {"loss_per_step": [3.524, 2.05, 1.264, 0.927, 0.527, 0.02, 0.006], "prob_new": [0.47731471061706543, 0.6420671343803406, 0.7347151041030884, 0.7479108572006226, 0.8374936580657959, 0.9803385734558105, 0.9936227798461914], "prob_old": [0.7825582027435303, 0.2549585998058319, 0.17553402483463287, 0.3467155396938324, 0.26138389110565186, 0.37211886048316956, 0.3907618820667267], "prob_new_token": [6.658617479615714e-08, 1.784269989002496e-05, 0.0003946858341805637, 0.0046521457843482494, 0.02958543598651886, 0.9318791031837463, 0.9970288872718811], "prob_old_token": [0.7788311839103699, 8.08291861176258e-06, 2.124595266650431e-05, 9.092922482523136e-06, 8.267453267762903e-06, 3.388380065416641e-08, 7.566668402958499e-10], "l1-model.layers.4.mlp.down_proj.weight": [64320.02734375], "l2-model.layers.4.mlp.down_proj.weight": [10.533592224121094], "linf-model.layers.4.mlp.down_proj.weight": [0.0030019525438547134], "request": {"prompt": "{} entered this world in the location of", "subject": "Rachel Maddow", "target_new": {"str": "Bourg-la-Reine"}, "old_answer": {"str": "Castro Valley, California"}, "seed": 42}}, {"loss_per_step": [2.998, 3.875, 2.987, 1.335, 0.388, 0.123, 0.059, 0.041, 0.032, 0.025, 0.019, 0.014, 0.011, 0.009], "prob_new": [0.6149026155471802, 0.25997328758239746, 0.4837050139904022, 0.6272946000099182, 0.7635538578033447, 0.8909569978713989, 0.9438736438751221, 0.9602314829826355, 0.9687298536300659, 0.9754196405410767, 0.9813034534454346, 0.9859867095947266, 0.9891726970672607, 0.9912722706794739], "prob_old": [0.7825582027435303, 0.17098866403102875, 0.18249447643756866, 0.31178855895996094, 0.4602903425693512, 0.4109530746936798, 0.36536210775375366, 0.31548020243644714, 0.27953392267227173, 0.26600900292396545, 0.2630714774131775, 0.26325222849845886, 0.26436442136764526, 0.2658036947250366], "prob_new_token": [1.1470999197626952e-05, 4.0582832298241556e-05, 3.007140549016185e-05, 0.008523950353264809, 0.2648964524269104, 0.7179309725761414, 0.9156180620193481, 0.9676075577735901, 0.984789252281189, 0.9918338060379028, 0.9951733946800232, 0.9968637228012085, 0.997673511505127, 0.9979919195175171], "prob_old_token": [0.7788311839103699, 3.086582637479296e-06, 2.4190687781810993e-06, 4.34781713920529e-06, 5.98151359554322e-07, 1.0683227458230249e-07, 3.173794738131619e-08, 1.1913813047215172e-08, 5.191454377495575e-09, 2.588468106523578e-09, 1.476967437241683e-09, 9.755216634488306e-10, 7.481124608688106e-10, 6.499192850561997e-10], "l1-model.layers.4.mlp.down_proj.weight": [87950.65625], "l2-model.layers.4.mlp.down_proj.weight": [15.258221626281738], "linf-model.layers.4.mlp.down_proj.weight": [0.006333108991384506], "request": {"prompt": "{} entered this world in the location of", "subject": "Rachel Maddow", "target_new": {"str": "Queens, New York"}, "old_answer": {"str": "Castro Valley, California"}, "seed": 42}}, {"loss_per_step": [3.428, 2.29, 1.566, 0.462, 0.065, 0.032, 0.021, 0.015, 0.011, 0.009], "prob_new": [0.4513727128505707, 0.5069100260734558, 0.6263516545295715, 0.7838972210884094, 0.9390920996665955, 0.9693893790245056, 0.9793224334716797, 0.9855304956436157, 0.9893602728843689, 0.9915439486503601], "prob_old": [0.7825582027435303, 0.20742231607437134, 0.2932290732860565, 0.3779626786708832, 0.397136926651001, 0.3854765295982361, 0.3554615378379822, 0.30671244859695435, 0.24910889565944672, 0.23306426405906677], "prob_new_token": [3.683622708194889e-06, 0.00016154169861692935, 0.001281743636354804, 0.12239794433116913, 0.8573698401451111, 0.9720847010612488, 0.9888647794723511, 0.9939074516296387, 0.9959323406219482, 0.9969028234481812], "prob_old_token": [0.7788311839103699, 3.3027688459696947e-06, 1.5066017340359394e-06, 7.717354151282052e-07, 1.6732165875055216e-07, 2.9749090302289005e-08, 9.474563356093313e-09, 4.317795010422287e-09, 2.545607502568714e-09, 1.7778153393877005e-09], "l1-model.layers.4.mlp.down_proj.weight": [78794.34375], "l2-model.layers.4.mlp.down_proj.weight": [13.10488510131836], "linf-model.layers.4.mlp.down_proj.weight": [0.004453628323972225], "request": {"prompt": "{} entered this world in the location of", "subject": "Rachel Maddow", "target_new": {"str": "Grand Rapids, Minnesota"}, "old_answer": {"str": "Castro Valley, California"}, "seed": 42}}, {"loss_per_step": [4.425, 0.896, 0.129, 0.06, 0.032, 0.019, 0.013, 0.009], "prob_new": [0.3284355401992798, 0.4978789985179901, 0.8805336952209473, 0.9418174624443054, 0.9688886404037476, 0.9814455509185791, 0.987565279006958, 0.9909398555755615], "prob_old": [0.7979272603988647, 0.6191149950027466, 0.5144929885864258, 0.5028420686721802, 0.5005991458892822, 0.5000488758087158, 0.4997110068798065, 0.4993802309036255], "prob_new_token": [7.54646953282645e-06, 0.15985043346881866, 0.9203985929489136, 0.9149821400642395, 0.9456495642662048, 0.9646083116531372, 0.9750765562057495, 0.9809909462928772], "prob_old_token": [0.6284904479980469, 0.11006728559732437, 0.00019753507513087243, 0.0001410731638316065, 6.917931023053825e-05, 4.0537612221669406e-05, 2.6928650186164305e-05, 1.930098733282648e-05], "l1-model.layers.4.mlp.down_proj.weight": [67214.4453125], "l2-model.layers.4.mlp.down_proj.weight": [11.446283340454102], "linf-model.layers.4.mlp.down_proj.weight": [0.003462808206677437], "request": {"prompt": "{} entered this world in the location of", "subject": "Albrecht D\u00fcrer", "target_new": {"str": "Florence, Italy"}, "old_answer": {"str": "Nuremberg"}, "seed": 42}}, {"loss_per_step": [3.647, 2.148, 1.527, 1.204, 0.824, 0.289, 0.032, 0.003], "prob_new": [0.2978762984275818, 0.5790072679519653, 0.7467890977859497, 0.7909545302391052, 0.7969130873680115, 0.8461721539497375, 0.9703041315078735, 0.9968177676200867], "prob_old": [0.7979272603988647, 0.479971706867218, 0.5324996709823608, 0.4974213242530823, 0.49616122245788574, 0.4987175762653351, 0.49479955434799194, 0.4952816963195801], "prob_new_token": [1.9384273400646634e-05, 0.00014312402345240116, 0.0006425023311749101, 0.002546960487961769, 0.01681443676352501, 0.23797346651554108, 0.8561257719993591, 0.9877517223358154], "prob_old_token": [0.6284904479980469, 0.004028975497931242, 0.006640058010816574, 0.004096167162060738, 0.003250336507335305, 0.0026797009631991386, 0.00019231911574024707, 9.516196769254748e-06], "l1-model.layers.4.mlp.down_proj.weight": [65703.5078125], "l2-model.layers.4.mlp.down_proj.weight": [11.276411056518555], "linf-model.layers.4.mlp.down_proj.weight": [0.00350284855812788], "request": {"prompt": "{} entered this world in the location of", "subject": "Albrecht D\u00fcrer", "target_new": {"str": "Aberdeen, Washington"}, "old_answer": {"str": "Nuremberg"}, "seed": 42}}, {"loss_per_step": [7.303, 5.971, 3.085, 0.634, 0.076, 0.015, 0.003], "prob_new": [0.005293817725032568, 0.010202256962656975, 0.1713361144065857, 0.6398901343345642, 0.929391622543335, 0.9856442213058472, 0.9970664381980896], "prob_old": [0.7979272603988647, 0.6037538051605225, 0.5950546860694885, 0.5436850190162659, 0.5529540181159973, 0.557043194770813, 0.5431874394416809], "prob_new_token": [4.2988340283045545e-05, 0.0003244888212066144, 0.006210710387676954, 0.2819924056529999, 0.8595514297485352, 0.9725537896156311, 0.997309684753418], "prob_old_token": [0.6284904479980469, 0.0014196195406839252, 0.0020409582648426294, 0.0002644166233949363, 0.00010851667320821434, 1.9703620637301356e-05, 9.232615525434085e-07], "l1-model.layers.4.mlp.down_proj.weight": [60422.99609375], "l2-model.layers.4.mlp.down_proj.weight": [10.363429069519043], "linf-model.layers.4.mlp.down_proj.weight": [0.002991214394569397], "request": {"prompt": "{} entered this world in the location of", "subject": "Albrecht D\u00fcrer", "target_new": {"str": "Reus"}, "old_answer": {"str": "Nuremberg"}, "seed": 42}}, {"loss_per_step": [11.799, 5.535, 0.462, 0.028, 0.007], "prob_new": [7.5101984293723945e-06, 0.0039442735724151134, 0.6297834515571594, 0.9724915027618408, 0.9934502840042114], "prob_old": [0.8133355975151062, 0.44312000274658203, 0.49092915654182434, 0.49194595217704773, 0.4921000599861145], "prob_new_token": [7.5101984293723945e-06, 0.0039442735724151134, 0.6297834515571594, 0.9724915027618408, 0.9934502840042114], "prob_old_token": [0.7344122529029846, 0.0014917884254828095, 0.0018626463133841753, 0.000106722567579709, 2.1745763660874218e-05], "l1-model.layers.4.mlp.down_proj.weight": [51178.5234375], "l2-model.layers.4.mlp.down_proj.weight": [8.369386672973633], "linf-model.layers.4.mlp.down_proj.weight": [0.0020022112876176834], "request": {"prompt": "{} passed away in the location of", "subject": "Wilhelm R\u00f6ntgen", "target_new": {"str": "Paris"}, "old_answer": {"str": "Munich"}, "seed": 42}}, {"loss_per_step": [4.471, 3.061, 2.66, 1.221, 0.272, 0.16, 0.066, 0.027, 0.013, 0.008], "prob_new": [0.2872834801673889, 0.46770021319389343, 0.49450379610061646, 0.6339683532714844, 0.8335230946540833, 0.8809016346931458, 0.9413074254989624, 0.9745686650276184, 0.9872375726699829, 0.9923409223556519], "prob_old": [0.8133355975151062, 0.20731522142887115, 0.16921192407608032, 0.09612379223108292, 0.03213147073984146, 0.024383725598454475, 0.023757465183734894, 0.0261665228754282, 0.03049718402326107, 0.03557880222797394], "prob_new_token": [0.000622739375103265, 0.0036970830988138914, 0.0025890585966408253, 0.014430848881602287, 0.3384881317615509, 0.5301071405410767, 0.7730569243431091, 0.905552864074707, 0.955120861530304, 0.9745497107505798], "prob_old_token": [0.7344122529029846, 0.0002765742246992886, 4.207733945804648e-05, 1.9210436221328564e-05, 2.7885962481377646e-06, 1.5465525393665303e-06, 8.649598157717264e-07, 4.1765025571294245e-07, 2.063917321493136e-07, 1.1053824522377909e-07], "l1-model.layers.4.mlp.down_proj.weight": [76072.296875], "l2-model.layers.4.mlp.down_proj.weight": [13.04823112487793], "linf-model.layers.4.mlp.down_proj.weight": [0.004436220042407513], "request": {"prompt": "{} passed away in the location of", "subject": "Wilhelm R\u00f6ntgen", "target_new": {"str": "Ephesus"}, "old_answer": {"str": "Munich"}, "seed": 42}}, {"loss_per_step": [2.974, 2.279, 0.386, 0.031, 0.01, 0.007], "prob_new": [0.6511784791946411, 0.44303226470947266, 0.7364957332611084, 0.9692920446395874, 0.9896411895751953, 0.9932205080986023], "prob_old": [0.8133355975151062, 0.26892542839050293, 0.1208612322807312, 0.10424419492483139, 0.10915210098028183, 0.11138081550598145], "prob_new_token": [0.00013980829680804163, 0.003082955488935113, 0.3761148750782013, 0.9552978873252869, 0.9848846197128296, 0.9869571328163147], "prob_old_token": [0.7344122529029846, 3.857244155369699e-05, 5.426342681857932e-07, 5.250408463552958e-08, 3.140734250450805e-08, 4.285782040369668e-08], "l1-model.layers.4.mlp.down_proj.weight": [57686.16015625], "l2-model.layers.4.mlp.down_proj.weight": [9.536797523498535], "linf-model.layers.4.mlp.down_proj.weight": [0.002474764361977577], "request": {"prompt": "{} passed away in the location of", "subject": "Wilhelm R\u00f6ntgen", "target_new": {"str": "Montreux"}, "old_answer": {"str": "Munich"}, "seed": 42}}, {"loss_per_step": [3.754, 1.779, 1.21, 1.224, 0.488, 0.127, 0.024, 0.011, 0.006], "prob_new": [0.46865397691726685, 0.507084846496582, 0.6752340197563171, 0.7379754185676575, 0.8335444927215576, 0.9090756177902222, 0.9767900109291077, 0.9887964129447937, 0.9936724305152893], "prob_old": [0.6166081428527832, 0.1310824602842331, 0.2990571856498718, 0.07244568318128586, 0.3211345076560974, 0.3364313542842865, 0.36728936433792114, 0.39829081296920776, 0.414750874042511], "prob_new_token": [3.655817636172287e-06, 0.0017403108067810535, 0.0012343263952061534, 0.0005113660008646548, 0.04062395915389061, 0.4473766088485718, 0.8871828317642212, 0.961114227771759, 0.9843199253082275], "prob_old_token": [0.7293808460235596, 0.00014956887753214687, 0.00012920156586915255, 8.510749466950074e-05, 0.001251653884537518, 0.00012056712148478255, 1.1455627827672288e-05, 2.876040298360749e-06, 1.2745808817271609e-06], "l1-model.layers.4.mlp.down_proj.weight": [66918.890625], "l2-model.layers.4.mlp.down_proj.weight": [11.713754653930664], "linf-model.layers.4.mlp.down_proj.weight": [0.003941635601222515], "request": {"prompt": "{} passed away in the location of", "subject": "Sandro Botticelli", "target_new": {"str": "Berkeley, Gloucestershire"}, "old_answer": {"str": "Florence, Italy"}, "seed": 42}}, {"loss_per_step": [3.176, 2.06, 0.906, 0.089, 0.012, 0.005], "prob_new": [0.5475641489028931, 0.6251254081726074, 0.7370449304580688, 0.9180291295051575, 0.9884854555130005, 0.9954566955566406], "prob_old": [0.6166081428527832, 0.401339590549469, 0.5041888952255249, 0.5034937858581543, 0.5179395079612732, 0.5377072691917419], "prob_new_token": [8.43507734771265e-07, 0.00011851543968077749, 0.015404395759105682, 0.8016486763954163, 0.9950962662696838, 0.9994681477546692], "prob_old_token": [0.7293808460235596, 0.0030241073109209538, 0.02372932806611061, 0.001519464305602014, 6.661275619990192e-07, 1.752364831020259e-08], "l1-model.layers.4.mlp.down_proj.weight": [55486.109375], "l2-model.layers.4.mlp.down_proj.weight": [9.354015350341797], "linf-model.layers.4.mlp.down_proj.weight": [0.00249609537422657], "request": {"prompt": "{} passed away in the location of", "subject": "Sandro Botticelli", "target_new": {"str": "Johannesburg, South Africa"}, "old_answer": {"str": "Florence, Italy"}, "seed": 42}}, {"loss_per_step": [7.018, 4.624, 2.1, 0.295, 0.036, 0.02, 0.014, 0.01, 0.007], "prob_new": [0.48342373967170715, 0.2954719364643097, 0.49435314536094666, 0.774092972278595, 0.9648067355155945, 0.9801666736602783, 0.9857099652290344, 0.9897377490997314, 0.9926860928535461], "prob_old": [0.6166081428527832, 0.18994848430156708, 0.1870976686477661, 0.2502012848854065, 0.3287716805934906, 0.3786900043487549, 0.40885135531425476, 0.4300212562084198, 0.44255977869033813], "prob_new_token": [8.301199159177486e-07, 0.00016301267896778882, 0.015392610803246498, 0.561454176902771, 0.9355006814002991, 0.9644939303398132, 0.9745165109634399, 0.9817827343940735, 0.987098753452301], "prob_old_token": [0.7293808460235596, 0.0005998697597533464, 8.13614678918384e-05, 0.0006566871888935566, 1.4248902516555972e-05, 1.5991583950381028e-06, 3.668790782285214e-07, 1.230010724384556e-07, 5.362221244809007e-08], "l1-model.layers.4.mlp.down_proj.weight": [68413.8203125], "l2-model.layers.4.mlp.down_proj.weight": [11.91256046295166], "linf-model.layers.4.mlp.down_proj.weight": [0.0039271037094295025], "request": {"prompt": "{} passed away in the location of", "subject": "Sandro Botticelli", "target_new": {"str": "Munich"}, "old_answer": {"str": "Florence, Italy"}, "seed": 42}}, {"loss_per_step": [4.618, 1.011, 0.089, 0.043, 0.012, 0.008], "prob_new": [0.33125823736190796, 0.511175274848938, 0.9182159304618835, 0.9588863253593445, 0.9877371191978455, 0.9920946359634399], "prob_old": [0.9821176528930664, 0.09279198944568634, 0.0018864996964111924, 0.0003365105658303946, 7.41763215046376e-05, 3.796660894295201e-05], "prob_new_token": [0.0008528511389158666, 0.11472255736589432, 0.8186393976211548, 0.8858896493911743, 0.9697161316871643, 0.9824299812316895], "prob_old_token": [0.9821176528930664, 0.09279198944568634, 0.0018864996964111924, 0.0003365105658303946, 7.41763215046376e-05, 3.796660894295201e-05], "l1-model.layers.4.mlp.down_proj.weight": [60526.6875], "l2-model.layers.4.mlp.down_proj.weight": [9.717668533325195], "linf-model.layers.4.mlp.down_proj.weight": [0.002489062026143074], "request": {"prompt": "{} belongs to the continent of", "subject": "Gibraltar", "target_new": {"str": "the Americas"}, "old_answer": {"str": "Europe"}, "seed": 42}}, {"loss_per_step": [4.494, 1.875, 0.369, 0.007], "prob_new": [0.40577608346939087, 0.2594267427921295, 0.732463002204895, 0.9927108883857727], "prob_old": [0.9821176528930664, 0.006680884398519993, 0.00911020115017891, 4.32556880696211e-05], "prob_new_token": [0.00015386084851343185, 0.05022269859910011, 0.4909157454967499, 0.990946352481842], "prob_old_token": [0.9821176528930664, 0.006680884398519993, 0.00911020115017891, 4.32556880696211e-05], "l1-model.layers.4.mlp.down_proj.weight": [40620.73828125], "l2-model.layers.4.mlp.down_proj.weight": [6.772740364074707], "linf-model.layers.4.mlp.down_proj.weight": [0.0015024803578853607], "request": {"prompt": "{} belongs to the continent of", "subject": "Gibraltar", "target_new": {"str": "North America"}, "old_answer": {"str": "Europe"}, "seed": 42}}, {"loss_per_step": [4.551, 2.315, 1.63, 0.037, 0.017, 0.012, 0.01], "prob_new": [0.5188276767730713, 0.45539483428001404, 0.6643576622009277, 0.9641019701957703, 0.9829033613204956, 0.9883529543876648, 0.9904217720031738], "prob_old": [0.9821176528930664, 0.00747318658977747, 0.005186088848859072, 0.0016423693159595132, 0.00044964838889427483, 0.0001490654976805672, 7.833153358660638e-05], "prob_new_token": [2.102440930684679e-06, 0.002532045356929302, 0.007627834100276232, 0.9237196445465088, 0.969257116317749, 0.9786534309387207, 0.9809023141860962], "prob_old_token": [0.9821176528930664, 0.00747318658977747, 0.005186088848859072, 0.0016423693159595132, 0.00044964838889427483, 0.0001490654976805672, 7.833153358660638e-05], "l1-model.layers.4.mlp.down_proj.weight": [59561.85546875], "l2-model.layers.4.mlp.down_proj.weight": [10.089639663696289], "linf-model.layers.4.mlp.down_proj.weight": [0.0029546215664595366], "request": {"prompt": "{} belongs to the continent of", "subject": "Gibraltar", "target_new": {"str": "Antarctica"}, "old_answer": {"str": "Europe"}, "seed": 42}}, {"loss_per_step": [5.435, 2.331, 1.817, 1.204, 0.12, 0.016, 0.016, 0.016, 0.008], "prob_new": [0.332294762134552, 0.37133926153182983, 0.6610839366912842, 0.6701251864433289, 0.8992540240287781, 0.9840693473815918, 0.9844832420349121, 0.9844846725463867, 0.9919694662094116], "prob_old": [0.9558717608451843, 0.3675532937049866, 0.5761265158653259, 0.3735957741737366, 0.36673039197921753, 0.3732512593269348, 0.352999210357666, 0.3297984004020691, 0.31115052103996277], "prob_new_token": [1.6631542166578583e-05, 0.008660354651510715, 0.004378544632345438, 0.027454117313027382, 0.7005284428596497, 0.9536861181259155, 0.9548685550689697, 0.9552208781242371, 0.9787542819976807], "prob_old_token": [0.8699713349342346, 0.006190645974129438, 0.0060552326031029224, 0.00267804809845984, 0.0003078527806792408, 4.425917722983286e-05, 1.8785158317768946e-05, 9.574156138114631e-06, 6.147769909148337e-06], "l1-model.layers.4.mlp.down_proj.weight": [74432.4375], "l2-model.layers.4.mlp.down_proj.weight": [12.32308292388916], "linf-model.layers.4.mlp.down_proj.weight": [0.003963601775467396], "request": {"prompt": "{}, whose the capital city is", "subject": "Madeira", "target_new": {"str": "Gaborone"}, "old_answer": {"str": "Funchal"}, "seed": 42}}, {"loss_per_step": [5.152, 1.553, 0.664, 0.028, 0.03, 0.008], "prob_new": [0.21288073062896729, 0.5245634913444519, 0.6970522403717041, 0.9732519388198853, 0.9710792899131775, 0.9925333857536316], "prob_old": [0.9558717608451843, 0.34340548515319824, 0.32904762029647827, 0.10808267444372177, 0.06638666987419128, 0.051759399473667145], "prob_new_token": [1.2327059266681317e-05, 0.01697499305009842, 0.14384616911411285, 0.9207406640052795, 0.9141045212745667, 0.9783409237861633], "prob_old_token": [0.8699713349342346, 0.004111091606318951, 0.0006815034430474043, 1.4787350664846599e-05, 1.0528197890380397e-05, 7.103355528670363e-06], "l1-model.layers.4.mlp.down_proj.weight": [59850.66796875], "l2-model.layers.4.mlp.down_proj.weight": [9.66694164276123], "linf-model.layers.4.mlp.down_proj.weight": [0.0025105364620685577], "request": {"prompt": "{}, whose the capital city is", "subject": "Madeira", "target_new": {"str": "Dhaka"}, "old_answer": {"str": "Funchal"}, "seed": 42}}, {"loss_per_step": [8.007, 3.638, 2.192, 1.417, 0.585, 0.105, 0.019, 0.016, 0.01], "prob_new": [0.0035749729722738266, 0.046642787754535675, 0.48073333501815796, 0.5180047750473022, 0.6532547473907471, 0.9053293466567993, 0.9816361665725708, 0.9843363761901855, 0.9905372858047485], "prob_old": [0.9558717608451843, 0.3441464304924011, 0.33826297521591187, 0.330594003200531, 0.33005690574645996, 0.32720187306404114, 0.31936129927635193, 0.30571264028549194, 0.2930332124233246], "prob_new_token": [1.553952824906446e-05, 0.008125337772071362, 0.013144475407898426, 0.06024196743965149, 0.31204864382743835, 0.8146060705184937, 0.9703254103660583, 0.9787746667861938, 0.984250545501709], "prob_old_token": [0.8699713349342346, 0.0028055335860699415, 0.0018017695983871818, 0.0033287983387708664, 0.002011375967413187, 0.0003258663637097925, 0.00015430551138706505, 0.00014677851868327707, 6.615302845602855e-05], "l1-model.layers.4.mlp.down_proj.weight": [72517.0078125], "l2-model.layers.4.mlp.down_proj.weight": [12.359848022460938], "linf-model.layers.4.mlp.down_proj.weight": [0.003943921532481909], "request": {"prompt": "{}, whose the capital city is", "subject": "Madeira", "target_new": {"str": "Juba"}, "old_answer": {"str": "Funchal"}, "seed": 42}}, {"loss_per_step": [6.129, 2.485, 1.017, 0.017, 0.021, 0.018, 0.012, 0.008], "prob_new": [0.2189430147409439, 0.49631989002227783, 0.5615725517272949, 0.9837403297424316, 0.9792332649230957, 0.9825018048286438, 0.988060712814331, 0.9923318028450012], "prob_old": [0.773881196975708, 0.03469276428222656, 0.0013177202781662345, 3.596801252570003e-05, 2.853828846127726e-05, 2.0555517039611004e-05, 1.24467824207386e-05, 7.146420557546662e-06], "prob_new_token": [1.0830311111931223e-05, 0.007046266458928585, 0.1319459229707718, 0.9675068855285645, 0.958490252494812, 0.9650245308876038, 0.976139485836029, 0.9846791625022888], "prob_old_token": [0.773881196975708, 0.03469276428222656, 0.0013177202781662345, 3.596801252570003e-05, 2.853828846127726e-05, 2.0555517039611004e-05, 1.24467824207386e-05, 7.146420557546662e-06], "l1-model.layers.4.mlp.down_proj.weight": [64556.9140625], "l2-model.layers.4.mlp.down_proj.weight": [11.127151489257812], "linf-model.layers.4.mlp.down_proj.weight": [0.0034810826182365417], "request": {"prompt": "The original language of work of {} is", "subject": "Melodifestivalen", "target_new": {"str": "Romanian"}, "old_answer": {"str": "Swedish"}, "seed": 42}}, {"loss_per_step": [3.649, 6.287, 1.141, 0.255, 0.089, 0.052, 0.041, 0.035, 0.03, 0.026, 0.023, 0.02, 0.017, 0.014, 0.012, 0.011, 0.009], "prob_new": [0.02600996568799019, 0.0018605211516842246, 0.31942281126976013, 0.775239884853363, 0.9144018888473511, 0.9489781260490417, 0.9596548676490784, 0.9655117392539978, 0.9700150489807129, 0.9739236235618591, 0.9774339199066162, 0.9805794954299927, 0.9833462238311768, 0.9857267141342163, 0.9877370595932007, 0.9894148111343384, 0.9908084273338318], "prob_old": [0.773881196975708, 0.003904488170519471, 0.041900672018527985, 0.01552211306989193, 0.0017679383745416999, 0.0001921741059049964, 3.9559570723213255e-05, 1.536714808025863e-05, 8.557314686186146e-06, 5.763196895713918e-06, 4.332184289523866e-06, 3.497242460070993e-06, 2.9700818231503945e-06, 2.620165787448059e-06, 2.3795937522663735e-06, 2.209641479566926e-06, 2.0867998955509393e-06], "prob_new_token": [0.02600996568799019, 0.0018605211516842246, 0.31942281126976013, 0.775239884853363, 0.9144018888473511, 0.9489781260490417, 0.9596548676490784, 0.9655117392539978, 0.9700150489807129, 0.9739236235618591, 0.9774339199066162, 0.9805794954299927, 0.9833462238311768, 0.9857267141342163, 0.9877370595932007, 0.9894148111343384, 0.9908084273338318], "prob_old_token": [0.773881196975708, 0.003904488170519471, 0.041900672018527985, 0.01552211306989193, 0.0017679383745416999, 0.0001921741059049964, 3.9559570723213255e-05, 1.536714808025863e-05, 8.557314686186146e-06, 5.763196895713918e-06, 4.332184289523866e-06, 3.497242460070993e-06, 2.9700818231503945e-06, 2.620165787448059e-06, 2.3795937522663735e-06, 2.209641479566926e-06, 2.0867998955509393e-06], "l1-model.layers.4.mlp.down_proj.weight": [95146.9453125], "l2-model.layers.4.mlp.down_proj.weight": [16.295894622802734], "linf-model.layers.4.mlp.down_proj.weight": [0.007677339017391205], "request": {"prompt": "The original language of work of {} is", "subject": "Melodifestivalen", "target_new": {"str": "English"}, "old_answer": {"str": "Swedish"}, "seed": 42}}, {"loss_per_step": [9.855, 6.108, 4.309, 1.201, 0.314, 0.125, 0.053, 0.017, 0.007], "prob_new": [5.2486044296529144e-05, 0.0022256483789533377, 0.013449202291667461, 0.3009326756000519, 0.7303553819656372, 0.8827588558197021, 0.9486438035964966, 0.9834391474723816, 0.992769181728363], "prob_old": [0.773881196975708, 0.00020151410717517138, 8.72990203788504e-05, 0.0007048964616842568, 0.00042465049773454666, 0.00029696995625272393, 0.00010439578909426928, 3.848388951155357e-05, 1.768545371305663e-05], "prob_new_token": [5.2486044296529144e-05, 0.0022256483789533377, 0.013449202291667461, 0.3009326756000519, 0.7303553819656372, 0.8827588558197021, 0.9486438035964966, 0.9834391474723816, 0.992769181728363], "prob_old_token": [0.773881196975708, 0.00020151410717517138, 8.72990203788504e-05, 0.0007048964616842568, 0.00042465049773454666, 0.00029696995625272393, 0.00010439578909426928, 3.848388951155357e-05, 1.768545371305663e-05], "l1-model.layers.4.mlp.down_proj.weight": [67553.734375], "l2-model.layers.4.mlp.down_proj.weight": [11.820708274841309], "linf-model.layers.4.mlp.down_proj.weight": [0.003900982439517975], "request": {"prompt": "The original language of work of {} is", "subject": "Melodifestivalen", "target_new": {"str": "Japanese"}, "old_answer": {"str": "Swedish"}, "seed": 42}}, {"loss_per_step": [3.309, 1.34, 0.662, 0.368, 0.236, 0.13, 0.08, 0.058, 0.042, 0.03, 0.021, 0.015, 0.011, 0.008], "prob_new": [0.4706716537475586, 0.46278104186058044, 0.6478609442710876, 0.7873980402946472, 0.8415991067886353, 0.8960022330284119, 0.9291093945503235, 0.946686863899231, 0.9600266814231873, 0.9708291888237, 0.9794710278511047, 0.9854094386100769, 0.9895886778831482, 0.9921186566352844], "prob_old": [0.9521257877349854, 0.5872593522071838, 0.6973699927330017, 0.7133356928825378, 0.7149159908294678, 0.7168651223182678, 0.7198302745819092, 0.7229506969451904, 0.7243378162384033, 0.7242525219917297, 0.7231796979904175, 0.7212774753570557, 0.7186357975006104, 0.7153783440589905], "prob_new_token": [0.027645083144307137, 0.06722666323184967, 0.1283540278673172, 0.2120605856180191, 0.35838767886161804, 0.5753290057182312, 0.7298873662948608, 0.8114631772041321, 0.8633267283439636, 0.9027584195137024, 0.9327227473258972, 0.9536460041999817, 0.967623770236969, 0.9767885208129883], "prob_old_token": [0.8340222239494324, 0.05753856152296066, 0.005023201461881399, 0.0020439792424440384, 0.0009695525513961911, 0.0003133775608148426, 9.363432764075696e-05, 4.076895129401237e-05, 2.421165663690772e-05, 1.5352705304394476e-05, 9.665243851486593e-06, 6.2381132011068985e-06, 4.182837983535137e-06, 2.9123518743290333e-06], "l1-model.layers.4.mlp.down_proj.weight": [89192.25], "l2-model.layers.4.mlp.down_proj.weight": [15.33781909942627], "linf-model.layers.4.mlp.down_proj.weight": [0.006312653422355652], "request": {"prompt": "{} was originally aired on", "subject": "Rugrats", "target_new": {"str": "the Sci-Fi Channel"}, "old_answer": {"str": "Nickelodeon"}, "seed": 42}}, {"loss_per_step": [3.351, 1.986, 1.265, 0.635, 0.373, 0.19, 0.152, 0.123, 0.07, 0.042, 0.031, 0.027, 0.025, 0.023, 0.02, 0.016, 0.013, 0.011, 0.009], "prob_new": [0.20045112073421478, 0.28462573885917664, 0.3700871169567108, 0.6197094917297363, 0.7483561038970947, 0.8414205312728882, 0.8635537028312683, 0.887239933013916, 0.9331321716308594, 0.9592790603637695, 0.9694026708602905, 0.9732925891876221, 0.9755539894104004, 0.9779455065727234, 0.9808348417282104, 0.983935534954071, 0.9868475198745728, 0.9893271327018738, 0.9913119077682495], "prob_old": [0.9521257877349854, 0.6553912162780762, 0.7124593257904053, 0.7272173166275024, 0.7395079731941223, 0.7447042465209961, 0.7466657757759094, 0.7474031448364258, 0.7477110028266907, 0.7478954792022705, 0.7480345964431763, 0.7481482028961182, 0.7482407093048096, 0.7483166456222534, 0.748379111289978, 0.7484307885169983, 0.7484744787216187, 0.7485118508338928, 0.7485445737838745], "prob_new_token": [0.02764512225985527, 0.07282409071922302, 0.3575848639011383, 0.2224658876657486, 0.3725718557834625, 0.6359780430793762, 0.7880865335464478, 0.8542948961257935, 0.8965010046958923, 0.9206258654594421, 0.9307042360305786, 0.9345987439155579, 0.9378064274787903, 0.9429205060005188, 0.9501783847808838, 0.9584253430366516, 0.9663403630256653, 0.9731297492980957, 0.9785659313201904], "prob_old_token": [0.8340222239494324, 0.015698613598942757, 0.0008956357487477362, 0.005986019503325224, 0.001022560871206224, 0.00016183219850063324, 2.1692365407943726e-05, 7.256918706843862e-06, 4.439840722625377e-06, 3.009465899594943e-06, 2.0018267150589963e-06, 1.3455942280415911e-06, 9.306837682743208e-07, 6.609891443076776e-07, 4.784777161148668e-07, 3.518678681757592e-07, 2.6274983611074276e-07, 1.9934071815441712e-07, 1.537258498274241e-07], "l1-model.layers.4.mlp.down_proj.weight": [98607.171875], "l2-model.layers.4.mlp.down_proj.weight": [16.915498733520508], "linf-model.layers.4.mlp.down_proj.weight": [0.008020840585231781], "request": {"prompt": "{} was originally aired on", "subject": "Rugrats", "target_new": {"str": "the USA Network"}, "old_answer": {"str": "Nickelodeon"}, "seed": 42}}, {"loss_per_step": [4.551, 2.074, 1.127, 0.342, 0.028, 0.006], "prob_new": [0.03765115141868591, 0.35869330167770386, 0.45345497131347656, 0.7389967441558838, 0.9730142951011658, 0.9943959712982178], "prob_old": [0.9521257877349854, 0.5278376340866089, 0.5695276260375977, 0.5297198295593262, 0.5371222496032715, 0.570615291595459], "prob_new_token": [0.02764512225985527, 0.025898242369294167, 0.159999281167984, 0.5002712607383728, 0.9522621631622314, 0.9947937726974487], "prob_old_token": [0.8340222239494324, 0.03825785964727402, 0.006473801098763943, 0.001306918915361166, 8.837306086206809e-05, 5.314812369761057e-06], "l1-model.layers.4.mlp.down_proj.weight": [55310.09375], "l2-model.layers.4.mlp.down_proj.weight": [9.33238410949707], "linf-model.layers.4.mlp.down_proj.weight": [0.0025076204910874367], "request": {"prompt": "{} was originally aired on", "subject": "Rugrats", "target_new": {"str": "the CW"}, "old_answer": {"str": "Nickelodeon"}, "seed": 42}}, {"loss_per_step": [11.977, 2.709, 0.047, 0.02, 0.013, 0.012, 0.011, 0.009], "prob_new": [6.290205874392996e-06, 0.06659206748008728, 0.9542078375816345, 0.9803380966186523, 0.9875606298446655, 0.9883583784103394, 0.9892492294311523, 0.9905532598495483], "prob_old": [0.7823527455329895, 0.010163594968616962, 0.0003182053624186665, 0.00012677656195592135, 0.0001296464033657685, 0.0001567807630635798, 0.0001482660009060055, 0.00010955004108836874], "prob_new_token": [6.290205874392996e-06, 0.06659206748008728, 0.9542078375816345, 0.9803380966186523, 0.9875606298446655, 0.9883583784103394, 0.9892492294311523, 0.9905532598495483], "prob_old_token": [0.7823527455329895, 0.010163594968616962, 0.0003182053624186665, 0.00012677656195592135, 0.0001296464033657685, 0.0001567807630635798, 0.0001482660009060055, 0.00010955004108836874], "l1-model.layers.4.mlp.down_proj.weight": [68366.1875], "l2-model.layers.4.mlp.down_proj.weight": [11.444790840148926], "linf-model.layers.4.mlp.down_proj.weight": [0.0034601129591464996], "request": {"prompt": "{} has originated in the country named", "subject": "Shar Pei", "target_new": {"str": "Italy"}, "old_answer": {"str": "China"}, "seed": 42}}, {"loss_per_step": [11.487, 3.104, 0.21, 0.015, 0.013, 0.009], "prob_new": [1.0261817806167528e-05, 0.04489171877503395, 0.8106809854507446, 0.9854201674461365, 0.9872595071792603, 0.9909660220146179], "prob_old": [0.7823527455329895, 0.025316454470157623, 0.005478936247527599, 0.00023684368352405727, 0.00011445052223280072, 0.00010166227002628148], "prob_new_token": [1.0261817806167528e-05, 0.04489171877503395, 0.8106809854507446, 0.9854201674461365, 0.9872595071792603, 0.9909660220146179], "prob_old_token": [0.7823527455329895, 0.025316454470157623, 0.005478936247527599, 0.00023684368352405727, 0.00011445052223280072, 0.00010166227002628148], "l1-model.layers.4.mlp.down_proj.weight": [58585.4609375], "l2-model.layers.4.mlp.down_proj.weight": [9.580243110656738], "linf-model.layers.4.mlp.down_proj.weight": [0.002497123321518302], "request": {"prompt": "{} has originated in the country named", "subject": "Shar Pei", "target_new": {"str": "Spain"}, "old_answer": {"str": "China"}, "seed": 42}}, {"loss_per_step": [9.955, 2.15, 2.136, 0.05, 0.054, 0.046, 0.039, 0.038, 0.034, 0.028, 0.022, 0.016, 0.011, 0.008], "prob_new": [4.750975494971499e-05, 0.11647497862577438, 0.11809157580137253, 0.9510605335235596, 0.9476499557495117, 0.9551766514778137, 0.9612969756126404, 0.9631848931312561, 0.966558039188385, 0.972222626209259, 0.9785754680633545, 0.9842327237129211, 0.988664448261261, 0.991872251033783], "prob_old": [0.7823527455329895, 0.020022185519337654, 0.021226270124316216, 0.023960858583450317, 0.017661241814494133, 0.010534747503697872, 0.00722777284681797, 0.00569725688546896, 0.0044808704406023026, 0.0032654597889631987, 0.002228642115369439, 0.0014703624183312058, 0.0009628417901694775, 0.0006385068991221488], "prob_new_token": [4.750975494971499e-05, 0.11647497862577438, 0.11809157580137253, 0.9510605335235596, 0.9476499557495117, 0.9551766514778137, 0.9612969756126404, 0.9631848931312561, 0.966558039188385, 0.972222626209259, 0.9785754680633545, 0.9842327237129211, 0.988664448261261, 0.991872251033783], "prob_old_token": [0.7823527455329895, 0.020022185519337654, 0.021226270124316216, 0.023960858583450317, 0.017661241814494133, 0.010534747503697872, 0.00722777284681797, 0.00569725688546896, 0.0044808704406023026, 0.0032654597889631987, 0.002228642115369439, 0.0014703624183312058, 0.0009628417901694775, 0.0006385068991221488], "l1-model.layers.4.mlp.down_proj.weight": [84138.296875], "l2-model.layers.4.mlp.down_proj.weight": [14.718774795532227], "linf-model.layers.4.mlp.down_proj.weight": [0.006280437111854553], "request": {"prompt": "{} has originated in the country named", "subject": "Shar Pei", "target_new": {"str": "Japan"}, "old_answer": {"str": "China"}, "seed": 42}}, {"loss_per_step": [5.523, 2.783, 1.23, 0.176, 0.016, 0.004], "prob_new": [0.4938949942588806, 0.4907606840133667, 0.5415614247322083, 0.8514352440834045, 0.983809232711792, 0.9955980777740479], "prob_old": [0.9293187856674194, 0.6482290625572205, 0.6670796871185303, 0.6483948826789856, 0.6476413011550903, 0.6451981067657471], "prob_new_token": [1.6136593330884352e-05, 0.003915540408343077, 0.08557438105344772, 0.7033377885818481, 0.9678800702095032, 0.9914277195930481], "prob_old_token": [0.7632028460502625, 0.0064507885836064816, 0.000260011525824666, 5.195932317292318e-05, 1.0372626093158033e-05, 2.902206915678107e-06], "l1-model.layers.4.mlp.down_proj.weight": [54002.046875], "l2-model.layers.4.mlp.down_proj.weight": [9.198980331420898], "linf-model.layers.4.mlp.down_proj.weight": [0.0025004157796502113], "request": {"prompt": "{} was founded in the location of", "subject": "China Southern Airlines", "target_new": {"str": "Dubai"}, "old_answer": {"str": "Guangzhou"}, "seed": 42}}, {"loss_per_step": [9.34, 7.073, 6.078, 4.513, 3.028, 2.071, 1.688, 1.245, 0.278, 0.021, 0.011, 0.009], "prob_new": [0.0007351022795774043, 0.0039491499774158, 0.0031091622076928616, 0.013771899975836277, 0.19612368941307068, 0.4980572462081909, 0.4987527132034302, 0.5396506190299988, 0.7863304615020752, 0.9793407917022705, 0.9890474081039429, 0.9907697439193726], "prob_old": [0.9293187856674194, 0.5693168640136719, 0.6096467971801758, 0.663296103477478, 0.6551108360290527, 0.6633551120758057, 0.6697895526885986, 0.6652483940124512, 0.6690449118614197, 0.6628900170326233, 0.6567943096160889, 0.6520553231239319], "prob_new_token": [0.0014649422373622656, 0.007806240580976009, 0.005209413822740316, 0.0054345931857824326, 0.006069892086088657, 0.016205143183469772, 0.0355210080742836, 0.08315829932689667, 0.5736901760101318, 0.9625340104103088, 0.9867265224456787, 0.9888435006141663], "prob_old_token": [0.7632028460502625, 0.01295907236635685, 0.0010538609931245446, 0.0021130393724888563, 0.0006178310140967369, 0.0027103980537503958, 0.011299358680844307, 0.007863014005124569, 0.00026451548910699785, 5.3149033192312345e-06, 1.1019537851097994e-06, 5.784625045635039e-07], "l1-model.layers.4.mlp.down_proj.weight": [85259.2734375], "l2-model.layers.4.mlp.down_proj.weight": [14.553491592407227], "linf-model.layers.4.mlp.down_proj.weight": [0.00543509516865015], "request": {"prompt": "{} was founded in the location of", "subject": "China Southern Airlines", "target_new": {"str": "Jena"}, "old_answer": {"str": "Guangzhou"}, "seed": 42}}, {"loss_per_step": [7.946, 5.522, 3.403, 3.582, 1.096, 0.722, 0.173, 0.074, 0.047, 0.022, 0.01], "prob_new": [0.4137546718120575, 0.15341390669345856, 0.454366534948349, 0.4691331386566162, 0.5047261714935303, 0.6045507192611694, 0.8524072766304016, 0.9307636022567749, 0.9550977349281311, 0.9783247709274292, 0.9901353120803833], "prob_old": [0.9293187856674194, 0.5050119161605835, 0.6157976984977722, 0.6669772863388062, 0.523499071598053, 0.5068157315254211, 0.5055229663848877, 0.49956151843070984, 0.4981120228767395, 0.49719712138175964, 0.4963099956512451], "prob_new_token": [1.5147047349728382e-07, 5.2093786507612094e-05, 0.0012204318773001432, 0.0008257434237748384, 0.1266014277935028, 0.24490360915660858, 0.7117897272109985, 0.8644609451293945, 0.9124390482902527, 0.9586787819862366, 0.9821761846542358], "prob_old_token": [0.7632028460502625, 0.00493605574592948, 0.002087109722197056, 0.004026394337415695, 0.00022220904065761715, 0.0002594346588011831, 0.00022199025261215866, 0.00023356145538855344, 0.0002070416376227513, 0.00012422443251125515, 6.084296910557896e-05], "l1-model.layers.4.mlp.down_proj.weight": [76468.796875], "l2-model.layers.4.mlp.down_proj.weight": [13.378799438476562], "linf-model.layers.4.mlp.down_proj.weight": [0.004861022811383009], "request": {"prompt": "{} was founded in the location of", "subject": "China Southern Airlines", "target_new": {"str": "Bremen"}, "old_answer": {"str": "Guangzhou"}, "seed": 42}}, {"loss_per_step": [3.487, 1.929, 1.386, 0.831, 0.032, 0.005], "prob_new": [0.42314139008522034, 0.5090730786323547, 0.7400051951408386, 0.8004236221313477, 0.9700824022293091, 0.9951750040054321], "prob_old": [0.8802522420883179, 0.346861332654953, 0.2728865146636963, 0.2667219638824463, 0.2588808536529541, 0.2580450475215912], "prob_new_token": [6.021196440997301e-06, 0.001289371750317514, 0.001390078105032444, 0.015909738838672638, 0.8601142168045044, 0.9860121607780457], "prob_old_token": [0.6327256560325623, 0.005240249447524548, 0.0445319265127182, 0.04618397727608681, 0.0005011872272007167, 1.1816022379207425e-05], "l1-model.layers.4.mlp.down_proj.weight": [56279.7265625], "l2-model.layers.4.mlp.down_proj.weight": [9.444955825805664], "linf-model.layers.4.mlp.down_proj.weight": [0.002509371843189001], "request": {"prompt": "{} is lead by the person named", "subject": "Dyson", "target_new": {"str": "Judd Apatow"}, "old_answer": {"str": "James Dyson"}, "seed": 42}}, {"loss_per_step": [7.033, 5.972, 4.712, 3.762, 3.216, 2.561, 1.48, 0.84, 0.399, 0.067, 0.019, 0.013, 0.013, 0.012, 0.008], "prob_new": [0.2846667468547821, 0.3070167899131775, 0.3301668167114258, 0.3331509232521057, 0.3372271955013275, 0.3501436114311218, 0.48767420649528503, 0.6785291433334351, 0.7637550830841064, 0.9386460781097412, 0.9811769723892212, 0.9868795871734619, 0.9868776202201843, 0.988487720489502, 0.9918922781944275], "prob_old": [0.8802522420883179, 0.2669713497161865, 0.34936028718948364, 0.3167128264904022, 0.28070831298828125, 0.2687472999095917, 0.2707612216472626, 0.28332823514938354, 0.2868991792201996, 0.27590593695640564, 0.26798689365386963, 0.248783677816391, 0.22266726195812225, 0.19037392735481262, 0.15384632349014282], "prob_new_token": [0.00011093316425103694, 4.047375841764733e-05, 0.0012542539043352008, 0.00484488857910037, 0.006747425068169832, 0.010896986350417137, 0.026964442804455757, 0.08460947871208191, 0.30756106972694397, 0.827684223651886, 0.9553239345550537, 0.9743764996528625, 0.9752277135848999, 0.9769399166107178, 0.9828230738639832], "prob_old_token": [0.6327256560325623, 0.000151829794049263, 0.001642974908463657, 0.005979700479656458, 0.008012127131223679, 0.01228130329400301, 0.03162471204996109, 0.09561198204755783, 0.10374972224235535, 0.030250970274209976, 0.005010474938899279, 0.0014873931650072336, 0.0010758894495666027, 0.000804745708592236, 0.00046075950376689434], "l1-model.layers.4.mlp.down_proj.weight": [92886.203125], "l2-model.layers.4.mlp.down_proj.weight": [16.070497512817383], "linf-model.layers.4.mlp.down_proj.weight": [0.00671797152608633], "request": {"prompt": "{} is lead by the person named", "subject": "Dyson", "target_new": {"str": "George Friedman"}, "old_answer": {"str": "James Dyson"}, "seed": 42}}, {"loss_per_step": [6.315, 4.892, 4.01, 2.607, 0.883, 0.844, 0.034, 0.023, 0.034, 0.022, 0.018, 0.015, 0.008], "prob_new": [0.1324782520532608, 0.20284292101860046, 0.33565396070480347, 0.3471689820289612, 0.5787006616592407, 0.611260175704956, 0.9668711423873901, 0.9777436256408691, 0.9667848944664001, 0.9784994125366211, 0.9822196960449219, 0.9850789904594421, 0.9917101263999939], "prob_old": [0.8802522420883179, 0.2614116072654724, 0.30193591117858887, 0.3172467350959778, 0.4027424156665802, 0.36908653378486633, 0.34345051646232605, 0.2914368808269501, 0.24249882996082306, 0.1858493685722351, 0.14043790102005005, 0.11551448702812195, 0.10398828983306885], "prob_new_token": [0.00022606723359785974, 0.0005946359015069902, 0.006072737742215395, 0.028429660946130753, 0.10677667707204819, 0.7508781552314758, 0.9233169555664062, 0.9695835113525391, 0.9837952852249146, 0.9901934266090393, 0.9941726922988892, 0.9965170621871948, 0.9977927207946777], "prob_old_token": [0.6327256560325623, 0.001973510254174471, 0.025386992841959, 0.04429151490330696, 0.05849367380142212, 0.008601727895438671, 0.0011699185706675053, 0.00040257166256196797, 0.00030068380874581635, 0.00023840709764044732, 0.00016157836944330484, 0.00010071185533888638, 6.504908378701657e-05], "l1-model.layers.4.mlp.down_proj.weight": [79907.703125], "l2-model.layers.4.mlp.down_proj.weight": [14.180874824523926], "linf-model.layers.4.mlp.down_proj.weight": [0.005670449696481228], "request": {"prompt": "{} is lead by the person named", "subject": "Dyson", "target_new": {"str": "Marc Mayer"}, "old_answer": {"str": "James Dyson"}, "seed": 42}}, {"loss_per_step": [3.563, 2.158, 0.792, 0.087, 0.015, 0.003], "prob_new": [0.5048718452453613, 0.6036537885665894, 0.7359997630119324, 0.9239500164985657, 0.9852328896522522, 0.9965415000915527], "prob_old": [0.714084267616272, 0.40298086404800415, 0.5281789302825928, 0.3064042031764984, 0.25935667753219604, 0.2522086501121521], "prob_new_token": [4.025532234663842e-06, 0.00042161939200013876, 0.04686964675784111, 0.7373800873756409, 0.9540481567382812, 0.9925755262374878], "prob_old_token": [0.6126298904418945, 0.004783370532095432, 0.05456697940826416, 0.0005712092970497906, 3.212963565601967e-05, 2.31999274546979e-06], "l1-model.layers.4.mlp.down_proj.weight": [53746.83203125], "l2-model.layers.4.mlp.down_proj.weight": [9.238664627075195], "linf-model.layers.4.mlp.down_proj.weight": [0.002496287226676941], "request": {"prompt": "{} is lead by the person named", "subject": "Charles Schwab Corporation", "target_new": {"str": "Masayoshi Son"}, "old_answer": {"str": "Charles R Schwab"}, "seed": 42}}, {"loss_per_step": [5.602, 4.882, 2.86, 1.176, 0.42, 0.038, 0.029, 0.017, 0.008], "prob_new": [0.20161043107509613, 0.20582175254821777, 0.2969420850276947, 0.6659840941429138, 0.8080509305000305, 0.9631161093711853, 0.9721938967704773, 0.9832038283348083, 0.9925486445426941], "prob_old": [0.714084267616272, 0.44243207573890686, 0.45521441102027893, 0.39641517400741577, 0.30210646986961365, 0.2708890736103058, 0.25984179973602295, 0.2561032772064209, 0.25521495938301086], "prob_new_token": [9.207190487359185e-06, 9.582298616805929e-07, 0.0010132528841495514, 0.006202269811183214, 0.1344965547323227, 0.8930297493934631, 0.9844735264778137, 0.9942370057106018, 0.996917724609375], "prob_old_token": [0.6126298904418945, 0.00011289932444924489, 0.10404913872480392, 0.24231086671352386, 0.05041355639696121, 0.0016683870926499367, 9.978164598578587e-05, 2.2798494683229364e-05, 8.641572094347794e-06], "l1-model.layers.4.mlp.down_proj.weight": [68817.1328125], "l2-model.layers.4.mlp.down_proj.weight": [11.956130981445312], "linf-model.layers.4.mlp.down_proj.weight": [0.0039868769235908985], "request": {"prompt": "{} is lead by the person named", "subject": "Charles Schwab Corporation", "target_new": {"str": "Riccardo Muti"}, "old_answer": {"str": "Charles R Schwab"}, "seed": 42}}, {"loss_per_step": [4.487, 3.467, 1.798, 1.22, 0.355, 0.032, 0.015, 0.01, 0.007], "prob_new": [0.5412984490394592, 0.603691041469574, 0.7852780222892761, 0.7957429885864258, 0.8297351002693176, 0.9694128036499023, 0.9851551055908203, 0.9899503588676453, 0.9927248954772949], "prob_old": [0.714084267616272, 0.42094936966896057, 0.38870951533317566, 0.32434776425361633, 0.2845240831375122, 0.2489883154630661, 0.22696204483509064, 0.1971936970949173, 0.15519286692142487], "prob_new_token": [4.529347563675401e-08, 2.867299713216198e-07, 0.00013471055717673153, 0.0022931487765163183, 0.17363518476486206, 0.8901141285896301, 0.9692985415458679, 0.9811145663261414, 0.9845807552337646], "prob_old_token": [0.6126298904418945, 0.00010166149149881676, 0.0148073835298419, 0.03037411719560623, 0.00401973957195878, 0.00017805109382607043, 1.7537970052217133e-05, 3.995268343714997e-06, 1.6874313359949156e-06], "l1-model.layers.4.mlp.down_proj.weight": [69044.296875], "l2-model.layers.4.mlp.down_proj.weight": [11.89809799194336], "linf-model.layers.4.mlp.down_proj.weight": [0.003989246673882008], "request": {"prompt": "{} is lead by the person named", "subject": "Charles Schwab Corporation", "target_new": {"str": "Giorgio Armani"}, "old_answer": {"str": "Charles R Schwab"}, "seed": 42}}, {"loss_per_step": [3.441, 4.016, 0.357, 0.019, 0.01], "prob_new": [0.47477248311042786, 0.09200818091630936, 0.7248152494430542, 0.9812408685684204, 0.9902472496032715], "prob_old": [0.9123725891113281, 0.6693848371505737, 0.6388421058654785, 0.6513168811798096, 0.6425926089286804], "prob_new_token": [0.0010821707546710968, 0.0017810722347348928, 0.5357288122177124, 0.9981898665428162, 0.9996622800827026], "prob_old_token": [0.6529882550239563, 2.036258956650272e-05, 2.5107767669396708e-06, 4.5795638392576166e-09, 5.41160560807441e-10], "l1-model.layers.4.mlp.down_proj.weight": [46688.61328125], "l2-model.layers.4.mlp.down_proj.weight": [8.054627418518066], "linf-model.layers.4.mlp.down_proj.weight": [0.0020020194351673126], "request": {"prompt": "{} holds a position at", "subject": "Darleane C. Hoffman", "target_new": {"str": "Columbia University"}, "old_answer": {"str": "Lawrence Berkeley National Laboratory"}, "seed": 42}}, {"loss_per_step": [1.192, 1.446, 0.709, 0.048, 0.001], "prob_new": [0.7895854115486145, 0.6808828115463257, 0.8462114930152893, 0.9593969583511353, 0.9990853071212769], "prob_old": [0.9123725891113281, 0.6758687496185303, 0.6631060838699341, 0.6615837812423706, 0.6595084071159363], "prob_new_token": [0.00041883750236593187, 0.0029139649122953415, 0.00763500714674592, 0.7176425457000732, 0.9944093823432922], "prob_old_token": [0.6529882550239563, 0.00016057788161560893, 0.0005132781225256622, 0.00018183870997745544, 5.352462721930351e-06], "l1-model.layers.4.mlp.down_proj.weight": [51981.0546875], "l2-model.layers.4.mlp.down_proj.weight": [8.340323448181152], "linf-model.layers.4.mlp.down_proj.weight": [0.0020057647489011288], "request": {"prompt": "{} holds a position at", "subject": "Darleane C. Hoffman", "target_new": {"str": "Carnegie Mellon University"}, "old_answer": {"str": "Lawrence Berkeley National Laboratory"}, "seed": 42}}, {"loss_per_step": [4.146, 0.808, 0.187, 0.026, 0.008], "prob_new": [0.48203378915786743, 0.6047358512878418, 0.843815803527832, 0.9743559956550598, 0.9918941855430603], "prob_old": [0.9123725891113281, 0.6772093176841736, 0.7112259268760681, 0.690895676612854, 0.6720800399780273], "prob_new_token": [0.20117510855197906, 0.5151968002319336, 0.8695909976959229, 0.9446614980697632, 0.9735837578773499], "prob_old_token": [0.6529882550239563, 0.0007392223342321813, 0.00010306142939953133, 3.932260369765572e-05, 1.4008983271196485e-05], "l1-model.layers.4.mlp.down_proj.weight": [53692.8515625], "l2-model.layers.4.mlp.down_proj.weight": [8.520034790039062], "linf-model.layers.4.mlp.down_proj.weight": [0.002004432026296854], "request": {"prompt": "{} holds a position at", "subject": "Darleane C. Hoffman", "target_new": {"str": "the University of Cape Town"}, "old_answer": {"str": "Lawrence Berkeley National Laboratory"}, "seed": 42}}, {"loss_per_step": [2.123, 1.632, 0.07, 0.018, 0.006], "prob_new": [0.6134995818138123, 0.587287962436676, 0.9338049292564392, 0.9821454286575317, 0.9936738014221191], "prob_old": [0.8484284281730652, 0.47558099031448364, 0.4659186899662018, 0.46583643555641174, 0.44532299041748047], "prob_new_token": [0.0020436712075024843, 0.009913074783980846, 0.8898802995681763, 0.973676323890686, 0.9914369583129883], "prob_old_token": [0.7124742865562439, 0.0340852253139019, 0.0005892534973099828, 0.0003285559650976211, 0.00028322645812295377], "l1-model.layers.4.mlp.down_proj.weight": [43856.9609375], "l2-model.layers.4.mlp.down_proj.weight": [7.814121246337891], "linf-model.layers.4.mlp.down_proj.weight": [0.0020037591457366943], "request": {"prompt": "{} holds a position at", "subject": "J. M. Coetzee", "target_new": {"str": "Stanford University"}, "old_answer": {"str": "the University of Cape Town"}, "seed": 42}}, {"loss_per_step": [1.439, 0.434, 0.018, 0.004], "prob_new": [0.7682777643203735, 0.8517822623252869, 0.9824482202529907, 0.9956122040748596], "prob_old": [0.8484284281730652, 0.46513301134109497, 0.41332587599754333, 0.3670543134212494], "prob_new_token": [9.221502114087343e-05, 0.052637502551078796, 0.8940920829772949, 0.9818519353866577], "prob_old_token": [0.7124742865562439, 0.04936613887548447, 0.0011547827161848545, 0.0004158220544923097], "l1-model.layers.4.mlp.down_proj.weight": [37943.88671875], "l2-model.layers.4.mlp.down_proj.weight": [6.624020099639893], "linf-model.layers.4.mlp.down_proj.weight": [0.0015024319291114807], "request": {"prompt": "{} holds a position at", "subject": "J. M. Coetzee", "target_new": {"str": "Carnegie Mellon University"}, "old_answer": {"str": "the University of Cape Town"}, "seed": 42}}, {"loss_per_step": [2.52, 0.757, 0.069, 0.004], "prob_new": [0.6281738877296448, 0.825421929359436, 0.9385770559310913, 0.9964310526847839], "prob_old": [0.8484284281730652, 0.4550686478614807, 0.4306870102882385, 0.3981899917125702], "prob_new_token": [6.811330877098953e-06, 0.011298448778688908, 0.7381398677825928, 0.9956745505332947], "prob_old_token": [0.7124742865562439, 0.024575961753726006, 0.11466551572084427, 1.3831401702191215e-05], "l1-model.layers.4.mlp.down_proj.weight": [44656.01171875], "l2-model.layers.4.mlp.down_proj.weight": [7.0295538902282715], "linf-model.layers.4.mlp.down_proj.weight": [0.00150247011333704], "request": {"prompt": "{} holds a position at", "subject": "J. M. Coetzee", "target_new": {"str": "Lawrence Berkeley National Laboratory"}, "old_answer": {"str": "the University of Cape Town"}, "seed": 42}}, {"loss_per_step": [2.688, 2.182, 2.119, 1.405, 0.712, 0.265, 0.064, 0.018, 0.007], "prob_new": [0.4794250428676605, 0.6101720929145813, 0.7803455591201782, 0.7836119532585144, 0.7824258208274841, 0.8432148098945618, 0.9437355399131775, 0.9829060435295105, 0.9934358596801758], "prob_old": [0.8382276892662048, 0.31946754455566406, 0.33201339840888977, 0.32250669598579407, 0.3147401809692383, 0.3027276396751404, 0.3094640374183655, 0.31650763750076294, 0.3210269808769226], "prob_new_token": [2.253292768727988e-05, 0.00012314175546634942, 2.7724010578822345e-05, 0.000970957160461694, 0.03212727978825569, 0.2844509482383728, 0.7481443881988525, 0.9317026138305664, 0.9757428765296936], "prob_old_token": [0.6083126068115234, 0.015671223402023315, 0.0019855734426528215, 0.008149807341396809, 0.014033227227628231, 0.007444966584444046, 0.0017096841474995017, 0.00023202715965453535, 3.803075378527865e-05], "l1-model.layers.4.mlp.down_proj.weight": [72400.0], "l2-model.layers.4.mlp.down_proj.weight": [12.313445091247559], "linf-model.layers.4.mlp.down_proj.weight": [0.003965234849601984], "request": {"prompt": "{} is held by", "subject": "Minister of Foreign Affairs of Belarus", "target_new": {"str": "Idriss D\u00e9by"}, "old_answer": {"str": "Vladimir Makei"}, "seed": 42}}, {"loss_per_step": [5.808, 2.835, 1.827, 1.895, 0.98, 0.257, 0.054, 0.015, 0.008], "prob_new": [0.4812857508659363, 0.6236167550086975, 0.6165069937705994, 0.7432866096496582, 0.7497897148132324, 0.8359957933425903, 0.94996178150177, 0.9848414063453674, 0.9919518232345581], "prob_old": [0.8382276892662048, 0.30135616660118103, 0.2258654087781906, 0.23449596762657166, 0.2710410952568054, 0.29454508423805237, 0.30910205841064453, 0.3163093328475952, 0.3200257122516632], "prob_new_token": [9.099828446323954e-08, 2.282133573316969e-05, 0.0014312737621366978, 0.0005242633633315563, 0.020295167341828346, 0.36604881286621094, 0.8281285762786865, 0.9624716639518738, 0.9837713241577148], "prob_old_token": [0.6083126068115234, 0.010475512593984604, 0.00021177974122110754, 0.002191792940720916, 0.0004953208263032138, 0.00018236351024825126, 3.6375164199853316e-05, 5.254716597846709e-06, 1.6986059563350864e-06], "l1-model.layers.4.mlp.down_proj.weight": [71370.5625], "l2-model.layers.4.mlp.down_proj.weight": [11.952280044555664], "linf-model.layers.4.mlp.down_proj.weight": [0.003950280137360096], "request": {"prompt": "{} is held by", "subject": "Minister of Foreign Affairs of Belarus", "target_new": {"str": "Gaston Browne"}, "old_answer": {"str": "Vladimir Makei"}, "seed": 42}}, {"loss_per_step": [5.622, 2.757, 2.687, 1.076, 0.241, 0.569, 0.006], "prob_new": [0.4701083302497864, 0.5162407159805298, 0.7344527840614319, 0.7421882152557373, 0.842710018157959, 0.7496291399002075, 0.9939758777618408], "prob_old": [0.9186565279960632, 0.6878990530967712, 0.7485222220420837, 0.7856805920600891, 0.7897828817367554, 0.783076822757721, 0.7787375450134277], "prob_new_token": [4.263490609446308e-06, 9.569268149789423e-05, 2.292608769494109e-05, 0.01416192390024662, 0.3868766725063324, 0.11616780608892441, 0.9987766146659851], "prob_old_token": [0.6722553372383118, 6.415137704607332e-06, 9.767893061507493e-05, 0.0009307721629738808, 0.0001057181871146895, 0.00015122735931072384, 5.641166467285075e-08], "l1-model.layers.4.mlp.down_proj.weight": [60018.73828125], "l2-model.layers.4.mlp.down_proj.weight": [10.213308334350586], "linf-model.layers.4.mlp.down_proj.weight": [0.0030163228511810303], "request": {"prompt": "{} is held by", "subject": "list of heads of state of Chad", "target_new": {"str": "Gaston Browne"}, "old_answer": {"str": "Idriss D\u00e9by"}, "seed": 42}}, {"loss_per_step": [7.107, 7.093, 3.765, 3.758, 2.326, 0.829, 0.095, 0.036, 0.032, 0.009], "prob_new": [0.15828540921211243, 0.01600620523095131, 0.3266823887825012, 0.3402666449546814, 0.3775351047515869, 0.6024338603019714, 0.9111104011535645, 0.965033769607544, 0.9696002006530762, 0.9908673167228699], "prob_old": [0.9186565279960632, 0.4541829526424408, 0.7050426006317139, 0.522335946559906, 0.7366035580635071, 0.5333624482154846, 0.41276103258132935, 0.40279126167297363, 0.4108302593231201, 0.40927624702453613], "prob_new_token": [4.95036510983482e-06, 1.2367507224553265e-05, 0.001110095763579011, 0.00039782936801202595, 0.005787572357803583, 0.12029337882995605, 0.8415786027908325, 0.9204293489456177, 0.9191644191741943, 0.9781304597854614], "prob_old_token": [0.6722553372383118, 4.149458254687488e-05, 8.03189177531749e-05, 0.0010846591321751475, 0.00019867865194100887, 1.1275959877821151e-05, 6.578472948604031e-07, 3.6229232591722393e-07, 4.2010859147012525e-07, 1.1710067582271222e-07], "l1-model.layers.4.mlp.down_proj.weight": [73157.703125], "l2-model.layers.4.mlp.down_proj.weight": [12.54827880859375], "linf-model.layers.4.mlp.down_proj.weight": [0.004400733858346939], "request": {"prompt": "{} is held by", "subject": "list of heads of state of Chad", "target_new": {"str": "Vladimir Makei"}, "old_answer": {"str": "Idriss D\u00e9by"}, "seed": 42}}, {"loss_per_step": [3.05, 0.709, 0.031, 0.011, 0.006], "prob_new": [0.46842142939567566, 0.7219676375389099, 0.9698837399482727, 0.9892498254776001, 0.9945087432861328], "prob_old": [0.8401201963424683, 0.665272057056427, 0.6506825089454651, 0.6463392972946167, 0.6368250250816345], "prob_new_token": [4.251266091159778e-06, 0.04642254486680031, 0.9320759773254395, 0.9826192259788513, 0.991644561290741], "prob_old_token": [0.8187586665153503, 5.728918040404096e-05, 5.689403224096168e-06, 1.1196027571713785e-06, 3.790073037635011e-07], "l1-model.layers.4.mlp.down_proj.weight": [53095.8359375], "l2-model.layers.4.mlp.down_proj.weight": [8.516822814941406], "linf-model.layers.4.mlp.down_proj.weight": [0.002005496993660927], "request": {"prompt": "{} is represented by the music label", "subject": "Godspeed You! Black Emperor", "target_new": {"str": "Warner Bros. Records"}, "old_answer": {"str": "Constellation Records"}, "seed": 42}}, {"loss_per_step": [2.417, 0.92, 0.546, 0.048, 0.016, 0.011, 0.009], "prob_new": [0.5253622531890869, 0.7732207775115967, 0.8216589689254761, 0.9560980796813965, 0.9838905334472656, 0.9890382289886475, 0.9906572103500366], "prob_old": [0.8401201963424683, 0.6529927253723145, 0.68581223487854, 0.7043749690055847, 0.6952426433563232, 0.6759977340698242, 0.656273603439331], "prob_new_token": [0.00048019958194345236, 0.005924337077885866, 0.04242231696844101, 0.7927268743515015, 0.9728052020072937, 0.9906527996063232, 0.9934290051460266], "prob_old_token": [0.8187586665153503, 6.719049270031974e-05, 1.1759140761569142e-05, 1.5553493426523346e-07, 8.470715684438801e-09, 2.579749081021987e-09, 1.5865718738794499e-09], "l1-model.layers.4.mlp.down_proj.weight": [57357.89453125], "l2-model.layers.4.mlp.down_proj.weight": [9.953248023986816], "linf-model.layers.4.mlp.down_proj.weight": [0.0029782410711050034], "request": {"prompt": "{} is represented by the music label", "subject": "Godspeed You! Black Emperor", "target_new": {"str": "Konvict Muzik"}, "old_answer": {"str": "Constellation Records"}, "seed": 42}}, {"loss_per_step": [5.813, 2.651, 0.954, 0.059, 0.017, 0.008], "prob_new": [0.14490577578544617, 0.4711080491542816, 0.6315356492996216, 0.9444054365158081, 0.9830600023269653, 0.9922345876693726], "prob_old": [0.8401201963424683, 0.6688024401664734, 0.6822577118873596, 0.6900112628936768, 0.6970041990280151, 0.6932365894317627], "prob_new_token": [6.53521738058771e-06, 0.0007090246654115617, 0.0687985047698021, 0.8750202059745789, 0.9612708687782288, 0.9821258783340454], "prob_old_token": [0.8187586665153503, 5.79239240323659e-05, 0.0003698290674947202, 4.367405563243665e-05, 7.905592610768508e-06, 2.4612561446701875e-06], "l1-model.layers.4.mlp.down_proj.weight": [57382.62109375], "l2-model.layers.4.mlp.down_proj.weight": [9.495575904846191], "linf-model.layers.4.mlp.down_proj.weight": [0.0024955179542303085], "request": {"prompt": "{} is represented by the music label", "subject": "Godspeed You! Black Emperor", "target_new": {"str": "Armada Music"}, "old_answer": {"str": "Constellation Records"}, "seed": 42}}, {"loss_per_step": [6.934, 3.749, 2.859, 2.256, 0.892, 0.024, 0.012, 0.01], "prob_new": [0.2340732216835022, 0.32476314902305603, 0.6123381853103638, 0.4530317187309265, 0.6787368059158325, 0.9762343764305115, 0.9881725311279297, 0.9903377294540405], "prob_old": [0.9576637148857117, 0.744482696056366, 0.7785007357597351, 0.7162163853645325, 0.6015319228172302, 0.5940658450126648, 0.5926516056060791, 0.5922643542289734], "prob_new_token": [1.0154884222401961e-07, 7.427718810504302e-05, 0.0002245553769171238, 0.002920124214142561, 0.07129959762096405, 0.9634127616882324, 0.9963111281394958, 0.998369574546814], "prob_old_token": [0.8164881467819214, 0.000165799239766784, 0.00026184471789747477, 8.429079025518149e-05, 4.299397915019654e-06, 2.4242092777626567e-08, 1.1847547387588975e-09, 3.638172818565266e-10], "l1-model.layers.4.mlp.down_proj.weight": [60796.453125], "l2-model.layers.4.mlp.down_proj.weight": [10.64919376373291], "linf-model.layers.4.mlp.down_proj.weight": [0.0033869645558297634], "request": {"prompt": "{} is represented by the music label", "subject": "Anne-Sophie Mutter", "target_new": {"str": "Peaceville Records"}, "old_answer": {"str": "Deutsche Grammophon"}, "seed": 42}}, {"loss_per_step": [4.19, 2.731, 1.825, 1.154, 0.262, 0.001], "prob_new": [0.34920310974121094, 0.3363656997680664, 0.6571480631828308, 0.7512831091880798, 0.8336619138717651, 0.9987480640411377], "prob_old": [0.9576637148857117, 0.6651965975761414, 0.6894463896751404, 0.6886894106864929, 0.7103538513183594, 0.7256836891174316], "prob_new_token": [4.1410003177588806e-06, 0.0006178204203024507, 0.0010556323686614633, 0.00993478111922741, 0.35931888222694397, 0.9985658526420593], "prob_old_token": [0.8164881467819214, 0.0001414970465702936, 0.00015306469867937267, 4.001101842732169e-06, 1.5707715647295117e-05, 3.671802584204187e-10], "l1-model.layers.4.mlp.down_proj.weight": [50409.140625], "l2-model.layers.4.mlp.down_proj.weight": [8.681831359863281], "linf-model.layers.4.mlp.down_proj.weight": [0.002510729245841503], "request": {"prompt": "{} is represented by the music label", "subject": "Anne-Sophie Mutter", "target_new": {"str": "XL Recordings"}, "old_answer": {"str": "Deutsche Grammophon"}, "seed": 42}}, {"loss_per_step": [8.684, 3.437, 1.483, 0.325, 0.021, 0.028, 0.004], "prob_new": [0.005541480612009764, 0.4118865728378296, 0.48859429359436035, 0.7497456073760986, 0.9794851541519165, 0.9724555611610413, 0.996268093585968], "prob_old": [0.9576637148857117, 0.7143616080284119, 0.5973014235496521, 0.6990711092948914, 0.6549065709114075, 0.6345095038414001, 0.6513528227806091], "prob_new_token": [2.58570617006626e-06, 0.001257944735698402, 0.05589359626173973, 0.5502299070358276, 0.9663604497909546, 0.9475347399711609, 0.9942581653594971], "prob_old_token": [0.8164881467819214, 5.088693433208391e-05, 1.2477088603191078e-05, 5.411147412814898e-06, 1.4307944695701735e-07, 4.372729023316424e-08, 1.0316113296937601e-08], "l1-model.layers.4.mlp.down_proj.weight": [60522.1640625], "l2-model.layers.4.mlp.down_proj.weight": [10.2551908493042], "linf-model.layers.4.mlp.down_proj.weight": [0.002973719500005245], "request": {"prompt": "{} is represented by the music label", "subject": "Anne-Sophie Mutter", "target_new": {"str": "Domino"}, "old_answer": {"str": "Deutsche Grammophon"}, "seed": 42}}, {"loss_per_step": [3.415, 2.648, 1.051, 0.808, 0.305, 0.074, 0.03, 0.015, 0.01], "prob_new": [0.46457263827323914, 0.46611282229423523, 0.7304279208183289, 0.7331458926200867, 0.8226186037063599, 0.932792603969574, 0.9711322784423828, 0.9849054217338562, 0.9904455542564392], "prob_old": [0.9080218076705933, 0.24110208451747894, 0.2501521706581116, 0.03353738784790039, 0.03426045924425125, 0.01092296652495861, 0.0013250078773126006, 0.00038203885196708143, 0.0002499691618140787], "prob_new_token": [1.5300216546165757e-05, 0.00014494183415081352, 0.007781741209328175, 0.026327701285481453, 0.25065016746520996, 0.7663887143135071, 0.9237142205238342, 0.9728474617004395, 0.9886836409568787], "prob_old_token": [0.7662683725357056, 0.0020933798514306545, 0.000120182303362526, 0.0003326705191284418, 4.171554610366002e-05, 1.405946136401326e-06, 1.6852658291099942e-07, 4.0713018734095385e-08, 1.380222069968795e-08], "l1-model.layers.4.mlp.down_proj.weight": [71620.859375], "l2-model.layers.4.mlp.down_proj.weight": [12.26378345489502], "linf-model.layers.4.mlp.down_proj.weight": [0.003965229727327824], "request": {"prompt": "The person who serves as the head coach for {} is", "subject": "Spain women's national association football team", "target_new": {"str": "Michael O'Neill"}, "old_answer": {"str": "Jorge Vilda"}, "seed": 42}}, {"loss_per_step": [2.651, 3.267, 3.189, 2.228, 0.5, 0.035, 0.017, 0.007], "prob_new": [0.7096829414367676, 0.6497116088867188, 0.7477972507476807, 0.7490041851997375, 0.7819116115570068, 0.9674658179283142, 0.9838398694992065, 0.9926284551620483], "prob_old": [0.9080218076705933, 0.1924498975276947, 0.2990821301937103, 0.41287732124328613, 0.630445659160614, 0.6291866898536682, 0.6209750771522522, 0.6206721663475037], "prob_new_token": [2.9521990654757246e-05, 3.467367605480831e-06, 2.91503056359943e-06, 0.00013532668526750058, 0.13668552041053772, 0.8735997676849365, 0.9380676746368408, 0.972595751285553], "prob_old_token": [0.7662683725357056, 0.00011668905062833801, 0.00010258124530082569, 0.0021246543619781733, 0.00020190652867313474, 5.093398431199603e-05, 1.4990780073276255e-05, 4.772534339281265e-06], "l1-model.layers.4.mlp.down_proj.weight": [59891.3828125], "l2-model.layers.4.mlp.down_proj.weight": [10.610655784606934], "linf-model.layers.4.mlp.down_proj.weight": [0.0034610703587532043], "request": {"prompt": "The person who serves as the head coach for {} is", "subject": "Spain women's national association football team", "target_new": {"str": "Pia Sundhage"}, "old_answer": {"str": "Jorge Vilda"}, "seed": 42}}, {"loss_per_step": [5.4, 2.839, 1.808, 1.357, 1.705, 0.604, 0.101, 0.038, 0.028, 0.021, 0.014, 0.009], "prob_new": [0.25741130113601685, 0.6412005424499512, 0.6539615988731384, 0.632205605506897, 0.5886392593383789, 0.707502543926239, 0.9112584590911865, 0.9637713432312012, 0.9731994867324829, 0.9800798892974854, 0.986403226852417, 0.9910768270492554], "prob_old": [0.9080218076705933, 0.23114995658397675, 0.28227758407592773, 0.03635408729314804, 0.0673384964466095, 0.08504344522953033, 0.12622424960136414, 0.17258256673812866, 0.18762847781181335, 0.1874789297580719, 0.18134723603725433, 0.17142772674560547], "prob_new_token": [2.869437594199553e-06, 0.0002164117177017033, 0.004606595728546381, 0.0193600095808506, 0.007847757078707218, 0.1719701588153839, 0.7484353184700012, 0.897854208946228, 0.9227829575538635, 0.9420477151870728, 0.9603849053382874, 0.9740528464317322], "prob_old_token": [0.7662683725357056, 0.0006582769565284252, 0.0015706369886174798, 0.0007145160925574601, 0.00010376005229773, 0.0002630503731779754, 0.00012776441872119904, 3.815073796431534e-05, 2.8062344426871277e-05, 2.523986950109247e-05, 2.160181975341402e-05, 1.743989923852496e-05], "l1-model.layers.4.mlp.down_proj.weight": [79933.578125], "l2-model.layers.4.mlp.down_proj.weight": [13.985027313232422], "linf-model.layers.4.mlp.down_proj.weight": [0.005120519548654556], "request": {"prompt": "The person who serves as the head coach for {} is", "subject": "Spain women's national association football team", "target_new": {"str": "Eddie Jones"}, "old_answer": {"str": "Jorge Vilda"}, "seed": 42}}, {"loss_per_step": [7.073, 3.673, 2.03, 0.178, 0.044, 0.022, 0.016, 0.013, 0.011, 0.009], "prob_new": [0.2787639796733856, 0.22369730472564697, 0.43397507071495056, 0.8501719236373901, 0.9581203460693359, 0.9787866473197937, 0.9844464659690857, 0.9873132705688477, 0.98955237865448, 0.9914881587028503], "prob_old": [0.8151693344116211, 0.01478831097483635, 0.08799028396606445, 0.3294106125831604, 0.31599196791648865, 0.26530325412750244, 0.2196558713912964, 0.18769103288650513, 0.16604465246200562, 0.15029454231262207], "prob_new_token": [2.648082272571628e-06, 0.0067010135389864445, 0.0076566957868635654, 0.6510525345802307, 0.8907795548439026, 0.9437887668609619, 0.9589609503746033, 0.9666721820831299, 0.9726797342300415, 0.9779260158538818], "prob_old_token": [0.6482585668563843, 0.018606064841151237, 0.03432786092162132, 0.013570821844041348, 0.006270017009228468, 0.004207288846373558, 0.0031734018120914698, 0.0024171709083020687, 0.0018360995454713702, 0.001391335390508175], "l1-model.layers.4.mlp.down_proj.weight": [72725.828125], "l2-model.layers.4.mlp.down_proj.weight": [12.385198593139648], "linf-model.layers.4.mlp.down_proj.weight": [0.004422519356012344], "request": {"prompt": "The person who serves as the head coach for {} is", "subject": "Huddersfield Town A.F.C.", "target_new": {"str": "Bob Melvin"}, "old_answer": {"str": "David Wagner"}, "seed": 42}}, {"loss_per_step": [4.296, 3.434, 1.139, 0.242, 0.01, 0.032, 0.002], "prob_new": [0.28204405307769775, 0.059408172965049744, 0.5510920882225037, 0.8017220497131348, 0.9898128509521484, 0.9686883687973022, 0.9977818131446838], "prob_old": [0.8151693344116211, 0.010439418256282806, 0.11569134145975113, 0.05880408734083176, 0.01397330779582262, 0.002405721927061677, 0.0021571661345660686], "prob_new_token": [0.000329130474710837, 0.009517427533864975, 0.10260160267353058, 0.6405990719795227, 0.9839064478874207, 0.9542388319969177, 0.9975571632385254], "prob_old_token": [0.6482585668563843, 0.015511701814830303, 0.0004261379363015294, 0.0010118044447153807, 1.5135727153392509e-05, 4.8782276280689985e-05, 4.0724339100961515e-07], "l1-model.layers.4.mlp.down_proj.weight": [59320.35546875], "l2-model.layers.4.mlp.down_proj.weight": [10.180988311767578], "linf-model.layers.4.mlp.down_proj.weight": [0.002984809223562479], "request": {"prompt": "The person who serves as the head coach for {} is", "subject": "Huddersfield Town A.F.C.", "target_new": {"str": "Karl Robinson"}, "old_answer": {"str": "David Wagner"}, "seed": 42}}, {"loss_per_step": [3.51, 1.468, 0.103, 0.067, 0.004], "prob_new": [0.5479843020439148, 0.5636338591575623, 0.912707507610321, 0.9370779395103455, 0.9962000250816345], "prob_old": [0.8151693344116211, 0.38091936707496643, 0.5127676725387573, 0.4408701956272125, 0.45309942960739136], "prob_new_token": [3.284277681814274e-06, 0.0037333944346755743, 0.6642882227897644, 0.9528048038482666, 0.9902158379554749], "prob_old_token": [0.6482585668563843, 0.023542216047644615, 0.03428097814321518, 0.00057347392430529, 7.980164082255214e-05], "l1-model.layers.4.mlp.down_proj.weight": [50282.4296875], "l2-model.layers.4.mlp.down_proj.weight": [8.273472785949707], "linf-model.layers.4.mlp.down_proj.weight": [0.0020057475194334984], "request": {"prompt": "The person who serves as the head coach for {} is", "subject": "Huddersfield Town A.F.C.", "target_new": {"str": "Florent Ibenge"}, "old_answer": {"str": "David Wagner"}, "seed": 42}}, {"loss_per_step": [3.54, 2.424, 0.928, 1.085, 0.029, 0.021, 0.016, 0.01], "prob_new": [0.5062932968139648, 0.6220786571502686, 0.7841753363609314, 0.7705036997795105, 0.9716720581054688, 0.9791080355644226, 0.9845878481864929, 0.9905808568000793], "prob_old": [0.8161789774894714, 0.4606950879096985, 0.6034942865371704, 0.5988468527793884, 0.5992504954338074, 0.5976842045783997, 0.596423327922821, 0.5958212018013], "prob_new_token": [7.655329682165757e-06, 2.2585809347219765e-05, 0.010618070140480995, 0.005180994514375925, 0.9055553674697876, 0.9617579579353333, 0.9813125133514404, 0.9902803301811218], "prob_old_token": [0.7256129384040833, 0.019543183967471123, 0.009774485602974892, 2.604901783342939e-05, 0.00022753651137463748, 9.467908239457756e-05, 3.7892819818807766e-05, 1.4168737834552303e-05], "l1-model.layers.4.mlp.down_proj.weight": [62507.41015625], "l2-model.layers.4.mlp.down_proj.weight": [10.905191421508789], "linf-model.layers.4.mlp.down_proj.weight": [0.0034600375220179558], "request": {"prompt": "{} has the job title of", "subject": "Paul Biya", "target_new": {"str": "Sultan of Brunei"}, "old_answer": {"str": "President of Cameroon"}, "seed": 42}}, {"loss_per_step": [5.507, 3.994, 1.982, 1.104, 0.718, 0.087, 0.037, 0.021, 0.015, 0.013, 0.011, 0.01, 0.009], "prob_new": [0.0979776456952095, 0.1385246068239212, 0.4099709093570709, 0.4872567355632782, 0.6837084889411926, 0.9200085997581482, 0.9637491106987, 0.9789949655532837, 0.9849259257316589, 0.9873782396316528, 0.9887183308601379, 0.9898905158042908, 0.9908849596977234], "prob_old": [0.8161789774894714, 0.555181622505188, 0.5719438791275024, 0.5802456736564636, 0.5422402024269104, 0.5263218283653259, 0.4846479892730713, 0.49623098969459534, 0.492356538772583, 0.49212923645973206, 0.5001904368400574, 0.504539430141449, 0.5061371326446533], "prob_new_token": [0.00014183954044710845, 0.000650079979095608, 0.01732725091278553, 0.24348010122776031, 0.6097996830940247, 0.9259152412414551, 0.9660487771034241, 0.9756766557693481, 0.9769554138183594, 0.9761174321174622, 0.9755967259407043, 0.9765640497207642, 0.9791139364242554], "prob_old_token": [0.7256129384040833, 0.03250827267765999, 0.04607540741562843, 0.06282380223274231, 0.004559281747788191, 0.0007160014356486499, 0.00013079900236334652, 3.8930582377361134e-05, 1.6662834241287783e-05, 9.166809832095169e-06, 6.032421879353933e-06, 4.280529083189322e-06, 3.079352609347552e-06], "l1-model.layers.4.mlp.down_proj.weight": [81066.2421875], "l2-model.layers.4.mlp.down_proj.weight": [14.268444061279297], "linf-model.layers.4.mlp.down_proj.weight": [0.005574185401201248], "request": {"prompt": "{} has the job title of", "subject": "Paul Biya", "target_new": {"str": "Grand Prince of Kiev"}, "old_answer": {"str": "President of Cameroon"}, "seed": 42}}, {"loss_per_step": [3.907, 1.636, 0.374, 0.105, 0.022, 0.017, 0.011, 0.008], "prob_new": [0.2329782247543335, 0.4203963279724121, 0.712246835231781, 0.9083288311958313, 0.9788082242012024, 0.9835875034332275, 0.9895137548446655, 0.9925404191017151], "prob_old": [0.8161789774894714, 0.49784231185913086, 0.5445994138717651, 0.5840975046157837, 0.5889775156974792, 0.5887574553489685, 0.5879111886024475, 0.5869030356407166], "prob_new_token": [0.00019359435827936977, 0.3084583282470703, 0.46620887517929077, 0.7112349271774292, 0.9569218158721924, 0.9646007418632507, 0.9765610098838806, 0.9836544990539551], "prob_old_token": [0.7256129384040833, 0.0031172847375273705, 0.006074090022593737, 0.0009016115800477564, 0.0003629896091297269, 0.0003251371963415295, 0.0001593343331478536, 8.334516314789653e-05], "l1-model.layers.4.mlp.down_proj.weight": [70935.625], "l2-model.layers.4.mlp.down_proj.weight": [11.581385612487793], "linf-model.layers.4.mlp.down_proj.weight": [0.0034902915358543396], "request": {"prompt": "{} has the job title of", "subject": "Paul Biya", "target_new": {"str": "King of the French"}, "old_answer": {"str": "President of Cameroon"}, "seed": 42}}, {"loss_per_step": [3.533, 2.168, 0.372, 0.036, 0.016, 0.007], "prob_new": [0.28912192583084106, 0.3758662939071655, 0.7881452441215515, 0.9651867747306824, 0.9843297004699707, 0.9930881857872009], "prob_old": [0.8448086977005005, 0.4889873266220093, 0.6517907977104187, 0.632521390914917, 0.5693286657333374, 0.5349369049072266], "prob_new_token": [0.008998566307127476, 0.028552038595080376, 0.206567645072937, 0.9103996157646179, 0.9684039354324341, 0.9886212348937988], "prob_old_token": [0.6732748746871948, 0.00015950931992847472, 3.399753040866926e-05, 2.273092150062439e-06, 5.276568231238343e-07, 3.216117931970075e-07], "l1-model.layers.4.mlp.down_proj.weight": [52673.38671875], "l2-model.layers.4.mlp.down_proj.weight": [9.146806716918945], "linf-model.layers.4.mlp.down_proj.weight": [0.0025049345567822456], "request": {"prompt": "{} has the job title of", "subject": "Qaboos bin Said Al Said", "target_new": {"str": "Prime Minister of the Netherlands"}, "old_answer": {"str": "Sultan of Oman"}, "seed": 42}}, {"loss_per_step": [3.966, 1.755, 0.293, 0.037, 0.015, 0.007], "prob_new": [0.2848266363143921, 0.30470940470695496, 0.7560399770736694, 0.9638828635215759, 0.9848812818527222, 0.9931876063346863], "prob_old": [0.8448086977005005, 0.37796902656555176, 0.4029177725315094, 0.3924941420555115, 0.4026951193809509, 0.40689095854759216], "prob_new_token": [0.006945076864212751, 0.05381329357624054, 0.7007400393486023, 0.9254364967346191, 0.9698318839073181, 0.9852721691131592], "prob_old_token": [0.6732748746871948, 0.00012467960186768323, 4.4489493120636325e-06, 1.1616995152508025e-06, 5.43154612842045e-07, 2.631107065553806e-07], "l1-model.layers.4.mlp.down_proj.weight": [58164.52734375], "l2-model.layers.4.mlp.down_proj.weight": [9.574908256530762], "linf-model.layers.4.mlp.down_proj.weight": [0.0025027282536029816], "request": {"prompt": "{} has the job title of", "subject": "Qaboos bin Said Al Said", "target_new": {"str": "President of the Republic of Congo"}, "old_answer": {"str": "Sultan of Oman"}, "seed": 42}}, {"loss_per_step": [4.176, 1.879, 0.908, 0.255, 0.034, 0.012, 0.009], "prob_new": [0.3508017659187317, 0.4498814344406128, 0.5268827080726624, 0.7978811860084534, 0.9670571088790894, 0.988361120223999, 0.9915239214897156], "prob_old": [0.8448086977005005, 0.5318344235420227, 0.463610976934433, 0.4247628450393677, 0.4759870171546936, 0.4680500626564026, 0.44680631160736084], "prob_new_token": [0.008998566307127476, 0.3089592456817627, 0.10292834788560867, 0.6501073837280273, 0.9044029712677002, 0.9804274439811707, 0.9935954213142395], "prob_old_token": [0.6732748746871948, 7.34489512979053e-05, 0.00020988710457459092, 1.099196538234537e-06, 9.310700193054799e-07, 3.6073777209821856e-07, 1.4700466977046744e-07], "l1-model.layers.4.mlp.down_proj.weight": [61022.59375], "l2-model.layers.4.mlp.down_proj.weight": [10.362351417541504], "linf-model.layers.4.mlp.down_proj.weight": [0.0029916653875261545], "request": {"prompt": "{} has the job title of", "subject": "Qaboos bin Said Al Said", "target_new": {"str": "Prime Minister of Italy"}, "old_answer": {"str": "Sultan of Oman"}, "seed": 42}}, {"loss_per_step": [4.673, 1.706, 0.335, 0.026, 0.0], "prob_new": [0.49746453762054443, 0.5527666807174683, 0.8076696395874023, 0.9753934144973755, 0.9995397925376892], "prob_old": [0.8818895220756531, 0.5075372457504272, 0.5838016867637634, 0.5734446048736572, 0.588479220867157], "prob_new_token": [0.00019636286015156657, 0.0051281992346048355, 0.2741937041282654, 0.9078967571258545, 0.9990224838256836], "prob_old_token": [0.7280361652374268, 0.00016208444139920175, 1.146085105574457e-05, 5.159370630281046e-06, 8.258361994251118e-09], "l1-model.layers.4.mlp.down_proj.weight": [51178.3359375], "l2-model.layers.4.mlp.down_proj.weight": [8.300177574157715], "linf-model.layers.4.mlp.down_proj.weight": [0.0020056143403053284], "request": {"prompt": "{} is in a relationship with", "subject": "Sally Ride", "target_new": {"str": "Ben Affleck"}, "old_answer": {"str": "Tam O'Shaughnessy"}, "seed": 42}}, {"loss_per_step": [3.955, 2.85, 0.762, 0.044, 0.007], "prob_new": [0.3345504701137543, 0.4061407744884491, 0.696379542350769, 0.9572403430938721, 0.9930183291435242], "prob_old": [0.8818895220756531, 0.6292809844017029, 0.7986322045326233, 0.7919231653213501, 0.8012588024139404], "prob_new_token": [0.004192287568002939, 0.0007934393361210823, 0.10325498133897781, 0.9988333582878113, 0.9919107556343079], "prob_old_token": [0.7280361652374268, 0.0003585945814847946, 0.005890262313187122, 3.4175229757238412e-06, 4.358249498181976e-06], "l1-model.layers.4.mlp.down_proj.weight": [47353.6015625], "l2-model.layers.4.mlp.down_proj.weight": [8.018265724182129], "linf-model.layers.4.mlp.down_proj.weight": [0.002005777321755886], "request": {"prompt": "{} is in a relationship with", "subject": "Sally Ride", "target_new": {"str": "Jamie Bell"}, "old_answer": {"str": "Tam O'Shaughnessy"}, "seed": 42}}, {"loss_per_step": [3.163, 1.326, 0.382, 0.022, 0.019, 0.02, 0.014, 0.008], "prob_new": [0.4374825954437256, 0.5898210406303406, 0.8227440714836121, 0.9783830642700195, 0.9817127585411072, 0.9806950688362122, 0.9859938621520996, 0.9918994903564453], "prob_old": [0.8818895220756531, 0.5368037223815918, 0.5874587893486023, 0.5662425756454468, 0.549160361289978, 0.5404624342918396, 0.5370537638664246, 0.5362670421600342], "prob_new_token": [7.356026617344469e-05, 0.006214276887476444, 0.15417009592056274, 0.9117880463600159, 0.9264638423919678, 0.9213679432868958, 0.9478567838668823, 0.9769893288612366], "prob_old_token": [0.7280361652374268, 0.001218705321662128, 0.003984543960541487, 0.00031557766487821937, 0.00011779821943491697, 7.00408490956761e-05, 3.1320261768996716e-05, 1.07361656773719e-05], "l1-model.layers.4.mlp.down_proj.weight": [70360.734375], "l2-model.layers.4.mlp.down_proj.weight": [11.539482116699219], "linf-model.layers.4.mlp.down_proj.weight": [0.0035101603716611862], "request": {"prompt": "{} is in a relationship with", "subject": "Sally Ride", "target_new": {"str": "Anna Kournikova"}, "old_answer": {"str": "Tam O'Shaughnessy"}, "seed": 42}}, {"loss_per_step": [3.395, 1.716, 1.95, 1.252, 0.233, 0.033, 0.009], "prob_new": [0.5038432478904724, 0.7749421000480652, 0.6403941512107849, 0.7896947264671326, 0.8526775240898132, 0.9689403772354126, 0.9915338754653931], "prob_old": [0.97446209192276, 0.21517431735992432, 0.19534356892108917, 0.2787039279937744, 0.06557510793209076, 0.023678403347730637, 0.024863090366125107], "prob_new_token": [2.238563865830656e-06, 0.00021451474458444864, 0.0002139612624887377, 0.002021290361881256, 0.33687135577201843, 0.887816309928894, 0.9728094935417175], "prob_old_token": [0.9460753798484802, 0.00034895853605121374, 0.00012998166494071484, 0.000755689456127584, 0.000955292780417949, 3.6810328310821205e-05, 3.4021061310340883e-06], "l1-model.layers.4.mlp.down_proj.weight": [64703.34765625], "l2-model.layers.4.mlp.down_proj.weight": [10.519639015197754], "linf-model.layers.4.mlp.down_proj.weight": [0.0029988493770360947], "request": {"prompt": "{} is in a relationship with", "subject": "Billie Jean King", "target_new": {"str": "Enrique Iglesias"}, "old_answer": {"str": "Ilana Kloss"}, "seed": 42}}, {"loss_per_step": [5.18, 1.415, 1.192, 0.191, 0.062, 0.024, 0.023, 0.02, 0.016, 0.012, 0.008], "prob_new": [0.49945124983787537, 0.5803536176681519, 0.6720400452613831, 0.8627926111221313, 0.9436270594596863, 0.9770364165306091, 0.9782008528709412, 0.9809362888336182, 0.984534502029419, 0.9886396527290344, 0.9919754266738892], "prob_old": [0.97446209192276, 0.2526822090148926, 0.1101021096110344, 0.07921762764453888, 0.08063451945781708, 0.07591097056865692, 0.07159546762704849, 0.06592127680778503, 0.06208556890487671, 0.060803499072790146, 0.060990430414676666], "prob_new_token": [2.7291832793707727e-06, 0.011152167804539204, 0.012546565383672714, 0.4804333746433258, 0.8017331957817078, 0.9302660822868347, 0.9273777008056641, 0.9327594041824341, 0.9439746141433716, 0.9586348533630371, 0.9709497690200806], "prob_old_token": [0.9460753798484802, 0.000691319874022156, 4.232400897308253e-05, 1.645861084398348e-05, 4.604598871082999e-06, 6.915475978530594e-07, 2.0891725682758988e-07, 1.0695276841943269e-07, 7.249205680182058e-08, 5.533544467084539e-08, 4.454434332501478e-08], "l1-model.layers.4.mlp.down_proj.weight": [82153.765625], "l2-model.layers.4.mlp.down_proj.weight": [13.734846115112305], "linf-model.layers.4.mlp.down_proj.weight": [0.004879285581409931], "request": {"prompt": "{} is in a relationship with", "subject": "Billie Jean King", "target_new": {"str": "Ben Affleck"}, "old_answer": {"str": "Ilana Kloss"}, "seed": 42}}, {"loss_per_step": [6.44, 4.61, 2.577, 0.848, 0.375, 0.134, 0.037, 0.016, 0.01, 0.008], "prob_new": [0.010797940194606781, 0.1176363080739975, 0.6090676188468933, 0.55450838804245, 0.7717827558517456, 0.8883823156356812, 0.9644057154655457, 0.9842783808708191, 0.989895224571228, 0.992537260055542], "prob_old": [0.97446209192276, 0.27043023705482483, 0.2978663444519043, 0.4272587299346924, 0.19257523119449615, 0.13820822536945343, 0.1287209391593933, 0.11409001052379608, 0.09413861483335495, 0.07390282303094864], "prob_new_token": [1.6411824617534876e-05, 0.0001411070697940886, 0.0005256792064756155, 0.1492207646369934, 0.3292354941368103, 0.6786523461341858, 0.9085942506790161, 0.9695553183555603, 0.9860221147537231, 0.9921541810035706], "prob_old_token": [0.9460753798484802, 0.0005136135150678456, 7.817936420906335e-05, 0.00014063900744076818, 0.00020414203754626215, 9.203780791722238e-05, 2.5002285838127136e-05, 7.294164333870867e-06, 2.712141849769978e-06, 1.2513845604189555e-06], "l1-model.layers.4.mlp.down_proj.weight": [76897.9453125], "l2-model.layers.4.mlp.down_proj.weight": [12.939961433410645], "linf-model.layers.4.mlp.down_proj.weight": [0.0043610744178295135], "request": {"prompt": "{} is in a relationship with", "subject": "Billie Jean King", "target_new": {"str": "Erwin Bach"}, "old_answer": {"str": "Ilana Kloss"}, "seed": 42}}, {"loss_per_step": [3.071, 0.996, 0.208, 0.074, 0.034, 0.021, 0.015, 0.011, 0.008], "prob_new": [0.4938259720802307, 0.6508638262748718, 0.8514070510864258, 0.9335256814956665, 0.967481791973114, 0.9791123270988464, 0.985167920589447, 0.9892026782035828, 0.9919447898864746], "prob_old": [0.8684470057487488, 0.49012938141822815, 0.4733737111091614, 0.4305911660194397, 0.37823042273521423, 0.33021053671836853, 0.292026162147522, 0.2624949812889099, 0.23953168094158173], "prob_new_token": [0.03165428712964058, 0.3715296685695648, 0.8658238053321838, 0.9156226515769958, 0.9365705847740173, 0.9508466720581055, 0.9637753367424011, 0.9745460748672485, 0.9821935296058655], "prob_old_token": [0.7590489983558655, 0.011118238791823387, 0.0005050176405347884, 8.849104779073969e-05, 1.6340965885319747e-05, 4.185400030110031e-06, 1.6086663663372747e-06, 8.071036177170754e-07, 4.774511808136594e-07], "l1-model.layers.4.mlp.down_proj.weight": [76311.5546875], "l2-model.layers.4.mlp.down_proj.weight": [12.497177124023438], "linf-model.layers.4.mlp.down_proj.weight": [0.003969811834394932], "request": {"prompt": "{} is a sport team member of", "subject": "Zinedine Zidane", "target_new": {"str": "the Cleveland Cavaliers"}, "old_answer": {"str": "Real Madrid"}, "seed": 42}}, {"loss_per_step": [2.965, 1.497, 0.223, 0.052, 0.028, 0.018, 0.014, 0.011, 0.01], "prob_new": [0.62272047996521, 0.6636883020401001, 0.8347877264022827, 0.9516083002090454, 0.973209023475647, 0.9820469617843628, 0.9863489866256714, 0.9887503385543823, 0.9905557036399841], "prob_old": [0.8684470057487488, 0.4115361273288727, 0.46110987663269043, 0.4586651027202606, 0.4007234275341034, 0.33019500970840454, 0.26099392771720886, 0.21727702021598816, 0.19622203707695007], "prob_new_token": [0.00015760859241709113, 0.011453510262072086, 0.5206034183502197, 0.863234281539917, 0.9238630533218384, 0.9496076703071594, 0.9616398215293884, 0.9677565097808838, 0.9727961421012878], "prob_old_token": [0.7590489983558655, 0.002527644857764244, 0.0074251871556043625, 0.001000100513920188, 0.00024239413323812187, 8.550850907340646e-05, 4.461634671315551e-05, 2.854732701962348e-05, 2.1269184799166396e-05], "l1-model.layers.4.mlp.down_proj.weight": [69715.25], "l2-model.layers.4.mlp.down_proj.weight": [12.01265811920166], "linf-model.layers.4.mlp.down_proj.weight": [0.0039739059284329414], "request": {"prompt": "{} is a sport team member of", "subject": "Zinedine Zidane", "target_new": {"str": "Arsenal"}, "old_answer": {"str": "Real Madrid"}, "seed": 42}}, {"loss_per_step": [3.481, 1.432, 0.122, 0.066, 0.048, 0.037, 0.028, 0.022, 0.017, 0.014, 0.011, 0.009], "prob_new": [0.41277772188186646, 0.5360618829727173, 0.8912691473960876, 0.9378923773765564, 0.9539505243301392, 0.9646978378295898, 0.97265625, 0.9786376953125, 0.9831690788269043, 0.9866205453872681, 0.9892668724060059, 0.9913070797920227], "prob_old": [0.8684470057487488, 0.4688897430896759, 0.4718339741230011, 0.4561467468738556, 0.4437233507633209, 0.4368664026260376, 0.43360084295272827, 0.43155041337013245, 0.4303688406944275, 0.42969954013824463, 0.4293455481529236, 0.4294014871120453], "prob_new_token": [0.03165428712964058, 0.3528710901737213, 0.7823505997657776, 0.9039956331253052, 0.9312449097633362, 0.9473088383674622, 0.9584835767745972, 0.9665486812591553, 0.9731622338294983, 0.9785861372947693, 0.9828988313674927, 0.986281156539917], "prob_old_token": [0.7590489983558655, 0.0013458715984597802, 0.0008572300430387259, 0.00028865397325716913, 0.00016097008483484387, 9.086591308005154e-05, 5.367381527321413e-05, 3.388223558431491e-05, 2.226034484920092e-05, 1.5130560313991737e-05, 1.0729911082307808e-05, 7.954243301355746e-06], "l1-model.layers.4.mlp.down_proj.weight": [86738.671875], "l2-model.layers.4.mlp.down_proj.weight": [14.361477851867676], "linf-model.layers.4.mlp.down_proj.weight": [0.005442673806101084], "request": {"prompt": "{} is a sport team member of", "subject": "Zinedine Zidane", "target_new": {"str": "the Dallas Mavericks"}, "old_answer": {"str": "Real Madrid"}, "seed": 42}}, {"loss_per_step": [1.431, 0.176, 0.004], "prob_new": [0.6811485886573792, 0.8819690942764282, 0.9956924319267273], "prob_old": [0.8201957941055298, 0.8231512904167175, 0.7979037761688232], "prob_new_token": [0.5299520492553711, 0.9993001818656921, 0.9893520474433899], "prob_old_token": [0.5299520492553711, 0.9993001818656921, 0.9893520474433899], "l1-model.layers.4.mlp.down_proj.weight": [35268.95703125], "l2-model.layers.4.mlp.down_proj.weight": [5.415898323059082], "linf-model.layers.4.mlp.down_proj.weight": [0.0010006781667470932], "request": {"prompt": "{} is a sport team member of", "subject": "Klay Thompson", "target_new": {"str": "the Cleveland Cavaliers"}, "old_answer": {"str": "the Golden State Warriors"}, "seed": 42}}, {"loss_per_step": [2.307, 0.984, 0.067, 0.009], "prob_new": [0.5194604396820068, 0.5908964276313782, 0.9386320114135742, 0.9908570051193237], "prob_old": [0.8201957941055298, 0.5914711952209473, 0.6016231775283813, 0.5944826602935791], "prob_new_token": [7.672882929909974e-05, 0.02399616502225399, 0.8378841280937195, 0.9731379747390747], "prob_old_token": [0.5299520492553711, 0.0070132482796907425, 0.0034457496367394924, 5.929998224019073e-05], "l1-model.layers.4.mlp.down_proj.weight": [37095.87109375], "l2-model.layers.4.mlp.down_proj.weight": [6.560554504394531], "linf-model.layers.4.mlp.down_proj.weight": [0.0015024635940790176], "request": {"prompt": "{} is a sport team member of", "subject": "Klay Thompson", "target_new": {"str": "Chelsea F.C."}, "old_answer": {"str": "the Golden State Warriors"}, "seed": 42}}, {"loss_per_step": [1.51, 0.292, 0.007], "prob_new": [0.6918376684188843, 0.8257721066474915, 0.9928600192070007], "prob_old": [0.8201957941055298, 0.7997575998306274, 0.7911826372146606], "prob_new_token": [0.5299520492553711, 0.9992063641548157, 0.9969547986984253], "prob_old_token": [0.5299520492553711, 0.9992063641548157, 0.9969547986984253], "l1-model.layers.4.mlp.down_proj.weight": [33677.5546875], "l2-model.layers.4.mlp.down_proj.weight": [5.276230812072754], "linf-model.layers.4.mlp.down_proj.weight": [0.0010006800293922424], "request": {"prompt": "{} is a sport team member of", "subject": "Klay Thompson", "target_new": {"str": "the Boston Celtics"}, "old_answer": {"str": "the Golden State Warriors"}, "seed": 42}}, {"loss_per_step": [7.038, 5.246, 2.276, 1.502, 0.211, 0.015, 0.005], "prob_new": [0.016076456755399704, 0.02865786664187908, 0.5030168890953064, 0.5152899622917175, 0.8268826603889465, 0.9852728247642517, 0.994657039642334], "prob_old": [0.671699583530426, 0.0001492609444539994, 0.0005145003669895232, 0.007164601236581802, 0.003581094788387418, 0.00015924561012070626, 4.81619244965259e-05], "prob_new_token": [2.4008397303987294e-05, 0.0004887469112873077, 0.0105975903570652, 0.05064045265316963, 0.6606201529502869, 0.9735419750213623, 0.9910470843315125], "prob_old_token": [0.671699583530426, 0.0001492609444539994, 0.0005145003669895232, 0.007164601236581802, 0.003581094788387418, 0.00015924561012070626, 4.81619244965259e-05], "l1-model.layers.4.mlp.down_proj.weight": [60806.96875], "l2-model.layers.4.mlp.down_proj.weight": [10.240032196044922], "linf-model.layers.4.mlp.down_proj.weight": [0.002963930368423462], "request": {"prompt": "{} is employed in the location of", "subject": "Klaus Wowereit", "target_new": {"str": "Delft"}, "old_answer": {"str": "Berlin"}, "seed": 42}}, {"loss_per_step": [10.085, 4.697, 1.982, 0.874, 0.184, 0.013, 0.007], "prob_new": [4.168611849308945e-05, 0.00912540964782238, 0.1377660036087036, 0.4170996844768524, 0.8315441012382507, 0.9873142242431641, 0.992726743221283], "prob_old": [0.671699583530426, 0.028358763083815575, 0.041276585310697556, 0.00040360825369134545, 0.00038856567698530853, 3.0661325581604615e-05, 2.1717662093578838e-05], "prob_new_token": [4.168611849308945e-05, 0.00912540964782238, 0.1377660036087036, 0.4170996844768524, 0.8315441012382507, 0.9873142242431641, 0.992726743221283], "prob_old_token": [0.671699583530426, 0.028358763083815575, 0.041276585310697556, 0.00040360825369134545, 0.00038856567698530853, 3.0661325581604615e-05, 2.1717662093578838e-05], "l1-model.layers.4.mlp.down_proj.weight": [58242.828125], "l2-model.layers.4.mlp.down_proj.weight": [10.084585189819336], "linf-model.layers.4.mlp.down_proj.weight": [0.002948388457298279], "request": {"prompt": "{} is employed in the location of", "subject": "Klaus Wowereit", "target_new": {"str": "Rome"}, "old_answer": {"str": "Berlin"}, "seed": 42}}, {"loss_per_step": [5.108, 4.441, 2.242, 0.865, 0.298, 0.131, 0.066, 0.041, 0.026, 0.018, 0.012, 0.008], "prob_new": [0.09006981551647186, 0.13276106119155884, 0.3785831332206726, 0.46215102076530457, 0.7458202838897705, 0.8803295493125916, 0.9369418025016785, 0.9601826667785645, 0.974016547203064, 0.9826406836509705, 0.9880535006523132, 0.99159836769104], "prob_old": [0.671699583530426, 0.0002965799649246037, 0.007948517799377441, 0.0008485930738970637, 0.00012181278725620359, 1.8198399629909545e-05, 5.846556177857565e-06, 2.659138772287406e-06, 1.3213083320806618e-06, 6.416793212338234e-07, 3.198429681106063e-07, 1.7112010652908793e-07], "prob_new_token": [1.2657715160457883e-05, 4.253616134519689e-05, 0.004002096597105265, 0.23458316922187805, 0.7794381976127625, 0.9375651478767395, 0.9556839466094971, 0.9649853706359863, 0.9761518836021423, 0.985582709312439, 0.9911044239997864, 0.994097113609314], "prob_old_token": [0.671699583530426, 0.0002965799649246037, 0.007948517799377441, 0.0008485930738970637, 0.00012181278725620359, 1.8198399629909545e-05, 5.846556177857565e-06, 2.659138772287406e-06, 1.3213083320806618e-06, 6.416793212338234e-07, 3.198429681106063e-07, 1.7112010652908793e-07], "l1-model.layers.4.mlp.down_proj.weight": [82330.046875], "l2-model.layers.4.mlp.down_proj.weight": [14.114816665649414], "linf-model.layers.4.mlp.down_proj.weight": [0.0052877929992973804], "request": {"prompt": "{} is employed in the location of", "subject": "Klaus Wowereit", "target_new": {"str": "Manchester, England"}, "old_answer": {"str": "Berlin"}, "seed": 42}}, {"loss_per_step": [6.17, 3.351, 2.072, 1.443, 0.604, 0.103, 0.022, 0.005], "prob_new": [0.17178383469581604, 0.2541770040988922, 0.49570906162261963, 0.5275578498840332, 0.6492698192596436, 0.9068593978881836, 0.9784404039382935, 0.9950434565544128], "prob_old": [0.4325380325317383, 0.042629729956388474, 0.0543210431933403, 0.03338349238038063, 0.01717350259423256, 0.018115466460585594, 0.02054992876946926, 0.02305002510547638], "prob_new_token": [1.272373538085958e-05, 0.0024296012707054615, 0.016275638714432716, 0.055832769721746445, 0.29858556389808655, 0.8137305378913879, 0.9568917751312256, 0.9901018142700195], "prob_old_token": [0.6283074617385864, 0.0003417119150981307, 4.8117573896888644e-05, 1.8072973034577444e-05, 7.395095963147469e-06, 6.509977765745134e-07, 1.4151019911423646e-07, 5.747109810272377e-08], "l1-model.layers.4.mlp.down_proj.weight": [66065.359375], "l2-model.layers.4.mlp.down_proj.weight": [11.250469207763672], "linf-model.layers.4.mlp.down_proj.weight": [0.0034903306514024734], "request": {"prompt": "{} is employed in the location of", "subject": "Friedrich Engels", "target_new": {"str": "Delft"}, "old_answer": {"str": "Manchester, England"}, "seed": 42}}, {"loss_per_step": [6.128, 2.995, 0.545, 0.035, 0.102, 0.002], "prob_new": [0.0021801020484417677, 0.05001430585980415, 0.5799093842506409, 0.9660760164260864, 0.903169572353363, 0.9981516599655151], "prob_old": [0.4325380325317383, 0.11363535374403, 0.06722729653120041, 0.042351480573415756, 0.06576599180698395, 0.04136333614587784], "prob_new_token": [0.0021801020484417677, 0.05001430585980415, 0.5799093842506409, 0.9660760164260864, 0.903169572353363, 0.9981516599655151], "prob_old_token": [0.6283074617385864, 0.0005258070304989815, 0.00013145465345587581, 3.1814527119422564e-06, 2.184042159569799e-06, 3.1372380249194975e-08], "l1-model.layers.4.mlp.down_proj.weight": [54479.6875], "l2-model.layers.4.mlp.down_proj.weight": [9.330381393432617], "linf-model.layers.4.mlp.down_proj.weight": [0.0024962760508060455], "request": {"prompt": "{} is employed in the location of", "subject": "Friedrich Engels", "target_new": {"str": "Berlin"}, "old_answer": {"str": "Manchester, England"}, "seed": 42}}, {"loss_per_step": [11.597, 7.006, 3.906, 0.464, 0.077, 0.026, 0.013, 0.008], "prob_new": [9.194967788062058e-06, 0.0009061197633855045, 0.020125307142734528, 0.6290601491928101, 0.9256868362426758, 0.9741030931472778, 0.9870151877403259, 0.9922285676002502], "prob_old": [0.4325380325317383, 0.08821200579404831, 0.20672820508480072, 0.17641675472259521, 0.16554459929466248, 0.18568484485149384, 0.2024652659893036, 0.20668064057826996], "prob_new_token": [9.194967788062058e-06, 0.0009061197633855045, 0.020125307142734528, 0.6290601491928101, 0.9256868362426758, 0.9741030931472778, 0.9870151877403259, 0.9922285676002502], "prob_old_token": [0.6283074617385864, 0.0028791099321097136, 0.0012461452279239893, 0.00023440000950358808, 1.1279917998763267e-05, 1.6473960613438976e-06, 4.781410893883731e-07, 2.0225292018949403e-07], "l1-model.layers.4.mlp.down_proj.weight": [68763.09375], "l2-model.layers.4.mlp.down_proj.weight": [11.557156562805176], "linf-model.layers.4.mlp.down_proj.weight": [0.0034118667244911194], "request": {"prompt": "{} is employed in the location of", "subject": "Friedrich Engels", "target_new": {"str": "Rome"}, "old_answer": {"str": "Manchester, England"}, "seed": 42}}] |