Allo-AVA / transcript /TED_P_6vDLq64gE.txt
SaifPunjwani's picture
Add files using upload-large-folder tool
c996935 verified
raw
history blame
26.3 kB
[0.000 --> 22.200] Okay, now I don't want to alarm anybody in this room, but it's just come to my attention
[22.200 --> 26.600] that the person to your right is a liar.
[26.600 --> 30.440] So the person to your left is a liar.
[30.440 --> 33.640] Also the person sitting in your very seats is a liar.
[33.640 --> 35.440] We're all liars.
[35.440 --> 39.080] What I'm going to do today is I'm going to show you what the research says about why we're
[39.080 --> 44.320] all liars, how you can become a lie spotter, and why you might want to go the extra mile
[44.320 --> 47.720] and go from lie spotting to truth seeking.
[47.720 --> 50.520] And ultimately to trust building.
[50.520 --> 56.560] Now speaking of trust, ever since I wrote this book lie spotting, no one wants to meet
[56.560 --> 57.560] me in person anymore.
[57.560 --> 58.560] No, no, no, no.
[58.560 --> 60.560] They say it's okay.
[60.560 --> 63.320] We'll email you.
[63.320 --> 67.680] I can't even get a coffee date at Starbucks.
[67.680 --> 69.680] My husband's like honey deception.
[69.680 --> 71.280] Maybe you could have focused on cooking.
[71.280 --> 72.960] How about French cooking?
[72.960 --> 78.200] So before we get started, what I'm going to do is I'm going to clarify my goal for you,
[78.200 --> 80.320] which is not to teach a game of gacha.
[80.320 --> 83.600] Lie spotters aren't those nit picky kids, those kids in the back of the room that are
[83.600 --> 88.600] shouting gacha, gacha, your eyebrow twitched, you flared, you're an astral, I watched that
[88.600 --> 91.200] TV show lie to me, I know you're lying.
[91.200 --> 92.200] No.
[92.200 --> 96.360] Lie spotters are armed with scientific knowledge of how to spot deception.
[96.360 --> 100.560] They use it to get to the truth, and they do what mature leaders do every day.
[100.560 --> 105.600] They have difficult conversations with difficult people, sometimes during very difficult times.
[105.600 --> 109.200] And they start up that path by accepting a core proposition.
[109.200 --> 111.440] And that proposition is the filing.
[111.440 --> 114.600] Lying is a cooperative act.
[114.600 --> 115.600] Think about it.
[115.600 --> 118.120] A lie has no power whatsoever by its mere utterance.
[118.120 --> 121.960] Its power emerges when someone else agrees to believe the lie.
[121.960 --> 128.320] So I know it may sound like tough love, but look, if at some point you got lied to, it's
[128.320 --> 130.200] because you agreed to get lied to.
[130.200 --> 131.720] Truth number one about lying.
[131.720 --> 132.760] Lying is a cooperative act.
[132.760 --> 135.840] Now, not all lies are harmful.
[135.840 --> 141.360] Sometimes we're willing participants in deception for the sake of social dignity,
[141.360 --> 144.160] maybe to keep a secret that should be kept secret secret.
[144.160 --> 146.160] We say, nice song.
[146.160 --> 147.760] Honey, you don't look fat in that.
[147.760 --> 148.760] No.
[148.760 --> 151.200] Or we say, favorite of the digerati.
[151.200 --> 154.280] You know, I just fished that email out of my spam folder.
[154.280 --> 156.720] I'm so sorry.
[156.720 --> 160.360] But there are times when we are unwilling participants in deception.
[160.360 --> 162.800] And that can have dramatic costs for us.
[162.800 --> 170.160] Last year saw $997 billion in corporate fraud alone in the United States.
[170.160 --> 172.200] That's an eyelash under $1 trillion.
[172.200 --> 174.480] That's 7% of revenues.
[174.480 --> 175.560] Deception can cost billions.
[175.560 --> 181.560] Think Enron, Mat off the mortgage crisis, or in the case of double agents and traders
[181.560 --> 185.960] like Robert Hansen or Aldrich Ames, lies can betray our country.
[185.960 --> 188.080] They can compromise our security.
[188.080 --> 189.360] They can undermine democracy.
[189.360 --> 191.960] They can cause the deaths of those that defend us.
[191.960 --> 194.720] Deception is actually serious business.
[194.720 --> 199.640] This con man, Henry Oberlander, he was such an effective con man.
[199.640 --> 203.360] The British authorities say he could have undermined the entire banking system of the Western
[203.360 --> 204.360] world.
[204.360 --> 205.360] And you can't find this guy on Google.
[205.360 --> 206.360] You can't find him anywhere.
[206.360 --> 207.360] He was interviewed once.
[207.360 --> 208.360] And he said the following.
[208.360 --> 210.520] He said, look, I've got one rule.
[210.520 --> 211.680] And this was Henry's rule.
[211.680 --> 215.600] He said, look, everyone is willing to give you something.
[215.600 --> 219.480] They're ready to give you something for whatever it is they're hungry for.
[219.480 --> 220.480] And that's the crux of it.
[220.480 --> 225.360] If you don't want to be deceived, you have to know what is it that you're hungry for.
[225.360 --> 227.120] And we all kind of hate to admit it.
[227.120 --> 233.000] You know, we kind of wish we were better husbands, better wives, smarter, more powerful,
[233.000 --> 234.560] taller, richer.
[234.560 --> 237.000] The list goes on.
[237.000 --> 241.640] Lying is an attempt to bridge that gap, to connect our wishes and our fantasies about
[241.640 --> 245.720] who we wish we were, how we wish we could be, with what we're really like.
[245.720 --> 249.200] And, boy, we're willing to fill in those gaps in our lives with lies.
[249.200 --> 254.680] On a given day, studies show that you may be lied to anywhere from 10 to 200 times.
[254.680 --> 257.960] And now granted, many of those are white lies.
[257.960 --> 263.600] But in another study, it showed that strangers lied three times within the first 10 minutes
[263.600 --> 265.040] of meeting each other.
[265.040 --> 269.320] Now, when we first hear this data, we recoil.
[269.320 --> 270.760] We can't believe how prevalent lying is.
[270.760 --> 272.560] We're essentially against lying.
[272.560 --> 276.720] But if you look more closely, the plot actually thickens.
[276.720 --> 280.640] We lie more to strangers than we lie to co-workers.
[280.640 --> 284.400] Extroverts lie more than introverts.
[284.400 --> 289.480] And lie eight times more about themselves than they do other people.
[289.480 --> 292.600] Women lie more to protect other people.
[292.600 --> 297.200] If you're in an average married couple, you're going to lie to your spouse in one out of
[297.200 --> 298.600] every 10 interactions.
[298.600 --> 300.400] Now, you may think that's bad.
[300.400 --> 303.760] If you're unmarried, that number drops to three.
[303.760 --> 305.240] Lying's complex.
[305.240 --> 308.880] It's woven into the fabric of our daily and our business lives, where deeply ambivalent
[308.880 --> 310.120] about the truth.
[310.120 --> 314.240] We parse it out on an as-needed basis, sometimes for very, very good reasons, and other times
[314.640 --> 317.720] just because we don't understand the gaps in our lives.
[317.720 --> 319.360] That's truth number two about lying.
[319.360 --> 322.680] We're against lying, but we're covertly for it.
[322.680 --> 326.720] It weighs that our society has sanctioned for centuries and centuries and centuries.
[326.720 --> 328.680] It's as old as breathing.
[328.680 --> 329.680] It's part of our culture.
[329.680 --> 331.400] It's part of our history.
[331.400 --> 338.480] Think Dante, Shakespeare, the Bible, news of the world.
[338.480 --> 341.480] Lying has evolutionary value to us as a species.
[341.480 --> 346.520] Researchers have long known that the more intelligent the species, the larger the neocortex,
[346.520 --> 349.200] the more likely it is to be deceptive.
[349.200 --> 350.680] Now you might remember Coco.
[350.680 --> 354.000] Does anybody here remember Coco the Gorilla, who was taught sign language?
[354.000 --> 357.200] Coco was taught to communicate via sign language.
[357.200 --> 358.200] Here's Coco with her kitten.
[358.200 --> 361.800] It's her cute little fluffy pet kitten.
[361.800 --> 368.000] Coco once blamed her pet kitten for ripping a sink out of the wall.
[368.000 --> 370.200] We're hardwired to become leaders of the pack.
[370.200 --> 372.240] It starts really, really early.
[372.240 --> 373.240] How early?
[373.240 --> 378.480] Well, babies will fake a cry, pause, wait to see who's coming, and then go right back to
[378.480 --> 380.560] crying.
[380.560 --> 383.760] One year old's learned concealment.
[383.760 --> 386.400] Two year old's bluff.
[386.400 --> 387.760] Five year olds lie outright.
[387.760 --> 390.200] They manipulate via flattery.
[390.200 --> 392.720] Nine year olds masters of the cover up.
[392.720 --> 395.680] By the time you enter college, you're going to lie to your mom and one out of every five
[395.680 --> 397.880] interactions.
[397.880 --> 401.880] By the time we enter this work world and we're breadwinners, we enter a world that has
[401.880 --> 408.960] just cluttered with spam, fake digital friends, partisan media, ingenious identity thieves,
[408.960 --> 413.120] world-class Ponzi schemers, a deception epidemic.
[413.120 --> 417.960] In short, what one author calls a post-truth society.
[417.960 --> 424.200] It's been very confusing for a long time now.
[424.200 --> 426.440] What do you do?
[426.440 --> 430.600] Well, there are steps we can take to navigate our way through the morass.
[430.600 --> 433.120] Trained lie spotters get to the truth 90% of the time.
[433.120 --> 435.920] The rest of us are only 54% accurate.
[435.920 --> 437.520] Why is it so easy to learn?
[437.520 --> 439.520] Well, they're good liars and they're bad liars.
[439.520 --> 440.920] They're no real original liars.
[440.920 --> 442.520] We all bake the same mistakes.
[442.520 --> 444.520] We all use the same techniques.
[444.520 --> 448.400] So what I'm going to do is I'm going to show you two patterns of deception and then we're
[448.400 --> 450.600] going to look at the hot spots and see if we can find them ourselves.
[450.600 --> 454.600] We're going to start with speech.
[454.600 --> 455.760] I want you to listen to me.
[455.760 --> 458.000] I'm going to say this again.
[458.000 --> 464.520] I did not have sexual relations with that woman, Miss Lewinsky.
[464.520 --> 469.680] I never told anybody it's a lie, not a single time, never.
[469.680 --> 475.160] These allegations are false and I need to go back to work for the American people.
[475.160 --> 477.160] Thank you.
[477.160 --> 479.880] Okay.
[479.880 --> 481.400] What were the telltale signs?
[481.400 --> 486.280] Well, first we heard what's known as a non-contracted denial.
[486.280 --> 490.920] Studies show that people who are over-determined in their denial will resort to formal rather than
[490.920 --> 492.360] informal language.
[492.360 --> 495.000] We also heard distancing language that woman.
[495.000 --> 499.920] We know that liars will unconsciously distance themselves from their subject using languages
[499.920 --> 500.920] their tool.
[500.920 --> 505.080] Now, Bill Clinton has said, well, they tell you the truth.
[505.080 --> 507.480] Or Richard Nixon's favorite in all candor.
[507.480 --> 511.160] He would have been a dead giveaway for any lie spotter that knows the qualifying language
[511.240 --> 513.560] as it's called qualifying language like that.
[513.560 --> 516.240] Further discredits the subject.
[516.240 --> 521.000] Now if he had repeated the question in its entirety or if he had peppered his account with a little
[521.000 --> 525.880] too much detail and we're all really glad he didn't do that, he would have further discredited
[525.880 --> 527.640] himself.
[527.640 --> 528.640] Freud had it right.
[528.640 --> 532.520] Freud said, look, there's much more to it than speech.
[532.520 --> 534.920] No mortal can keep a secret.
[534.920 --> 538.960] If his lips are silent, he shatters with his fingertips and we all do it.
[538.960 --> 541.120] No matter how powerful you are, we all do it.
[541.120 --> 543.000] We all chatter with our fingertips.
[543.000 --> 550.400] I'm going to show you Dominique Strauss-Con with Obama, who's chattering with his fingertips.
[550.400 --> 558.280] Now this brings us to our next pattern, which is body language.
[558.280 --> 560.640] With body language, here's what you've got to do.
[560.640 --> 565.120] You really got to just throw your assumptions out the door, let the science temper your
[565.120 --> 568.840] knowledge a little bit because we think liars fidget all the time.
[568.840 --> 572.240] Well guess what, they're known to freeze their upper bodies when they're lying.
[572.240 --> 575.040] We think liars won't look in the eyes.
[575.040 --> 578.600] Well guess what, they look you in the eyes a little too much just to compensate for that
[578.600 --> 579.600] myth.
[579.600 --> 585.760] We think warmth and smiles convey honesty, sincerity, but a trained lie spotter can spot a fake
[585.760 --> 587.400] smile a mile away.
[587.400 --> 591.160] Can you all spot the fake smile here?
[591.160 --> 596.960] You can consciously contract the muscles in your cheeks, but the real smiles in the
[596.960 --> 597.960] eyes.
[597.960 --> 602.460] Those feet of the eyes, they cannot be consciously contracted, especially if you overdo the
[602.460 --> 603.460] Botox.
[603.460 --> 604.460] Don't overdo the Botox.
[604.460 --> 606.760] Nobody will think you're honest.
[606.760 --> 608.600] Now we're going to look at the hot spots.
[608.600 --> 610.400] Can you tell what's happening in a conversation?
[610.400 --> 615.280] Can you start to find the hot spots to see the discrepancies between someone's words and
[615.280 --> 616.440] someone's actions?
[616.440 --> 620.880] Now I know it seems really obvious, but when you're having a conversation with someone
[620.880 --> 625.920] that uses a speck of deception, attitude is by far the most overlooked, but telling of
[625.920 --> 627.160] indicators.
[627.160 --> 631.040] When honest person is going to be cooperative, they're going to show their own side, they're
[631.040 --> 634.200] going to be enthusiastic, they're going to be willing and helpful in getting you to
[634.200 --> 635.200] the truth.
[635.200 --> 639.480] They're going to be willing to brainstorm, name suspects, provide details.
[639.480 --> 644.920] They're going to say, hey, maybe it was those guys in payroll that forged those checks.
[644.920 --> 648.600] They're going to be infuriated if they sense the wrongly accused about the entire course
[648.600 --> 652.040] of the interview, not just in flashes, they'll be infuriated throughout the entire course
[652.040 --> 653.040] of the interview.
[653.560 --> 658.200] If you ask someone honest, what should happen to whoever did forge those checks?
[658.200 --> 663.280] An honest person is much more likely to recommend strict, rather than lenient punishment.
[663.280 --> 668.440] Now, let's say you're having that exact same conversation with someone deceptive.
[668.440 --> 674.680] That person may be withdrawn, look down, lower their voice, pause, be kind of herky jerky.
[674.680 --> 678.000] Ask a deceptive person to tell their story, they're going to pepper it with way too much
[678.000 --> 681.840] detail in all kinds of irrelevant places.
[681.840 --> 684.640] And then they're going to tell their story in strict, chronological order.
[684.640 --> 689.520] And what a trained interrogator does is they come in, and in very subtle ways, over the
[689.520 --> 694.280] course of several hours, they will ask that person to tell their story backwards.
[694.280 --> 698.600] And then they'll watch them squirm and track which questions produce the highest volume
[698.600 --> 699.600] of deceptive tells.
[699.600 --> 700.600] Why do they do that?
[700.600 --> 702.360] Well, we all do the same thing.
[702.360 --> 705.720] We rehearse our words, but we rarely rehearse our gestures.
[705.720 --> 707.920] We say, yes, we shake our heads, no.
[707.920 --> 709.280] We tell very convincing stories.
[709.280 --> 711.160] We slightly shrug our shoulders.
[711.160 --> 715.840] We commit terrible crimes, and we smile at the delight in getting away with it.
[715.840 --> 718.920] Now that smile is known in the trade as duping delight.
[718.920 --> 722.680] And we're going to see that in several videos moving forward, but we're going to start
[722.680 --> 724.080] for those of you that don't know him.
[724.080 --> 728.840] This is presidential candidate John Edwards, who shocked America by fathering a child out
[728.840 --> 729.840] of wedlock.
[729.840 --> 733.080] We're going to see him talk about getting a paternity test.
[733.080 --> 738.080] To see now, if you can spot him saying yes while shaking his head, no, slightly shrugging
[738.080 --> 739.080] his shoulders.
[739.080 --> 741.320] I'm happy to participate in one.
[741.320 --> 746.760] I know that it's not possible that this child could be mine because of the timing of events.
[746.760 --> 748.480] So I know it's not possible.
[748.480 --> 751.520] Happy to take a paternity test, and we'd love to see it happen.
[751.520 --> 752.920] Are you going to do that soon?
[752.920 --> 754.920] Is there somebody that you can't?
[754.920 --> 755.920] I'm only one side.
[755.920 --> 760.760] I can only one side of the test, but I'm happy to participate in one.
[760.760 --> 765.640] Okay, those head shakes are much easier to spot once you know to look for them.
[765.640 --> 770.320] Now there are going to be times when someone makes one expression while masking another that
[770.320 --> 773.000] just kind of leaks through in a flash.
[773.000 --> 775.680] Murderers are known to leak sadness.
[775.680 --> 778.640] Your new joint venture partner might shake your hand, celebrate, go out to dinner with
[778.640 --> 781.720] you, and then leak an expression of anger.
[781.720 --> 785.400] And we're not all going to become facial expression experts overnight here, but there's one
[785.400 --> 789.120] I can teach you that's very dangerous and that's easy to learn, and that's the expression
[789.120 --> 790.120] of contempt.
[790.120 --> 794.160] Now with anger, you've got two people in an even playing field.
[794.160 --> 798.480] It's still somewhat of a healthy relationship, but when anger turns to contempt, you've
[798.480 --> 800.080] been dismissed.
[800.080 --> 804.800] It's associated with moral superiority, and for that reason, it's very, very hard to
[804.800 --> 805.800] recover from.
[805.800 --> 807.320] Here's what it looks like.
[807.320 --> 811.360] It's marked by one lip corner pulled up and in.
[811.360 --> 814.200] It's the only asymmetrical expression.
[814.200 --> 818.640] And in the presence of contempt, whether or not deception follows, and it doesn't always
[818.640 --> 820.120] follow.
[820.120 --> 824.080] Look the other way, go the other direction, reconsider the deal, say no, thank you.
[824.080 --> 826.680] I'm not coming up for just one more night cap.
[826.680 --> 828.320] Thank you.
[828.320 --> 831.840] Science has surfaced many, many more indicators.
[831.840 --> 836.680] We know, for example, we know Lyra's will shift their blink rate, point their feet towards
[836.680 --> 838.440] an exit.
[838.440 --> 842.760] They will take barrier objects and put them between themselves and the person that's interviewing
[842.760 --> 843.760] them.
[843.760 --> 847.680] They'll alter their vocal tone, often making them make their vocal tone much lower.
[847.680 --> 850.080] Now, here's the deal.
[850.080 --> 853.480] These behaviors are just behaviors.
[853.480 --> 856.040] They're not proof of deception.
[856.040 --> 857.040] They're red flags.
[857.040 --> 858.040] We're human beings.
[858.040 --> 861.520] We make deceptive flailing gestures all over the place all day long.
[861.520 --> 863.920] They don't mean anything in and of themselves.
[863.920 --> 867.000] But when you see clusters of them, that's your signal.
[867.000 --> 868.760] Look, listen, probe.
[868.760 --> 870.080] Ask some hard questions.
[870.080 --> 872.960] Get out of that very comfortable mode of knowing.
[872.960 --> 874.600] Walk into curiosity mode.
[874.600 --> 876.480] Ask more questions.
[876.480 --> 877.480] Have a little dignity.
[877.480 --> 878.960] Treat the person you're talking to with rapport.
[878.960 --> 882.320] Don't try to be like those folks on law and order and those other TV shows that pummel
[882.320 --> 884.080] their subjects into submission.
[884.080 --> 885.080] Don't be too aggressive.
[885.080 --> 887.040] It doesn't work.
[887.040 --> 890.720] Now, we've talked a little bit about how to talk to someone who's lying.
[890.720 --> 892.040] I mean, how to spot a lie.
[892.040 --> 895.160] And as I promise, we're now going to look at what the truth looks like.
[895.160 --> 897.920] And I'm going to show you two videos.
[897.920 --> 899.280] Two mothers.
[899.280 --> 900.280] One is lying.
[900.280 --> 901.280] One is telling the truth.
[901.280 --> 904.600] And these were surfaced by researcher David Matsumoto in California.
[904.600 --> 908.680] And I think they're an excellent example of what the truth looks like.
[908.680 --> 913.640] This mother, Diane Downes, shot her kids at close range.
[913.640 --> 917.040] Draw them to the hospital while they bled all over the car.
[917.040 --> 919.520] Claimed a scraggie-haired stranger did it.
[919.520 --> 923.280] And you'll see when you see the video, she can't even pretend to be an agonizing mother.
[923.280 --> 928.340] What you want to look for here is an incredible discrepancy between horrific events that she
[928.340 --> 931.400] describes and her very, very cool demeanor.
[931.400 --> 934.320] And if you look closely, you'll see dooping delight throughout this video.
[934.320 --> 937.880] But at night, when I close my eyes, I can see Christy reaching her hand out to me while
[937.880 --> 939.480] I'm driving.
[939.480 --> 942.000] And the blood just keeps coming out of her mouth.
[942.000 --> 944.720] And that, maybe it'll fade too with time.
[944.720 --> 946.440] But I don't think so.
[946.440 --> 956.000] That haunts me the most.
[956.000 --> 960.680] Now I'm going to show you a video of an actual grieving mother, Erin Runyon, confronting
[960.680 --> 963.720] her daughter's murder and torture in court.
[963.720 --> 965.680] Here you're going to see no false emotion.
[965.680 --> 968.880] The authentic expression of a mother's agony.
[968.880 --> 973.640] I wrote the statement on the third anniversary of the night you took my baby.
[973.640 --> 974.640] And you heard her.
[974.640 --> 976.640] Then you crushed her.
[976.640 --> 978.640] You terrified her.
[978.640 --> 981.120] And tell her her heart stopped.
[981.120 --> 982.640] And she fought.
[982.640 --> 984.640] And I know she fought you.
[984.640 --> 987.840] But I know she looked at you with those amazing brown eyes.
[987.840 --> 991.600] And you still wanted to kill her.
[991.600 --> 993.040] And I don't understand it.
[993.040 --> 994.440] And they never will.
[994.440 --> 1000.000] OK, there's no doubting the veracity of those emotions.
[1000.000 --> 1005.920] Now the technology around what the truth looks like is progressing on the science of it.
[1005.920 --> 1012.120] We know, for example, that we now have specialized eye trackers, infrared brain scans, MRIs that
[1012.120 --> 1016.360] can decode the signals that our bodies send out when we're trying to be deceptive.
[1016.360 --> 1020.960] And these technologies are going to be marketed to all of us as panacea is for deceit.
[1020.960 --> 1022.960] And they will prove incredibly useful someday.
[1022.960 --> 1028.080] But you've got to ask yourself in the meantime, who do you want on your side of the meeting?
[1028.080 --> 1032.280] Someone who's trained in getting to the truth or some guy is going to drag a 400 pound
[1032.280 --> 1035.520] electosephalogram through the door.
[1035.520 --> 1039.560] Lies spotters rely on human tools.
[1039.560 --> 1043.160] They know, as someone once said, characters who you are in the dark.
[1043.160 --> 1047.920] And what's kind of interesting is that today we have so little darkness.
[1047.920 --> 1050.160] Our world has lit up 24 hours a day.
[1050.160 --> 1052.000] It's transparent.
[1052.000 --> 1056.240] With blogs and social networks broadcasting the buzz of a whole new generation of people
[1056.240 --> 1059.240] that have made a choice to live their lives in public.
[1059.240 --> 1063.520] It's a much more noisy world.
[1063.520 --> 1070.360] So one challenge we have is to remember, oversharing, that's not honesty.
[1070.360 --> 1075.360] Our manic tweeting and texting can blind us to the fact that the subtleties of human
[1075.360 --> 1078.120] decency, character, integrity, that's still what matters.
[1078.120 --> 1079.960] That's always what's going to matter.
[1079.960 --> 1086.160] So in this much noisy world, it might make sense for us to be just a little bit more explicit
[1086.160 --> 1088.840] about our moral code.
[1088.840 --> 1093.160] When you combine the science of recognizing deception with the art of looking, listening,
[1093.160 --> 1096.600] you exempt yourself from collaborating in a lie.
[1096.600 --> 1101.240] You start up that path of being just a little bit more explicit because you signal to everyone
[1101.240 --> 1102.240] around you.
[1102.240 --> 1105.280] You say, hey, my world, our world.
[1105.280 --> 1106.720] It's going to be an honest one.
[1106.720 --> 1110.720] My world is going to be one where truth is strengthened and falsehood is recognized and
[1110.720 --> 1112.200] marginalized.
[1112.200 --> 1117.560] And when you do that, the ground around you starts to shift just a little bit.
[1117.560 --> 1118.560] And that's the truth.
[1118.560 --> 1119.560] Thank you.