62 Quotes & Sayings By Eliezer Yudkowsky

Eliezer Yudkowsky is the founder of the Machine Intelligence Research Institute (MIRI) and the co-founder of the Machine Intelligence Research Society (MIRS). He was also a researcher at Google during its early days. He is known for his work on cognitive science, rationality, and artificial intelligence.

1
There is no justice in the laws of nature, no term for fairness in the equations of motion. The Universe is neither evil, nor good, it simply does not care. The stars don't care, or the Sun, or the sky. But they don't have to! WE care! There IS light in the world, and it is US! Eliezer Yudkowsky
2
One of chief pieces of advice I give to aspiring rationalists is "Don't try to be clever." And, "Listen to those quiet, nagging doubts." If you don't know, you don't know what you don't know, you don't know how much you don't know, and you don't know how much you needed to know. Eliezer Yudkowsky
If you once tell a lie, the truth is ever...
3
If you once tell a lie, the truth is ever after your enemy. Eliezer Yudkowsky
4
And someday when the descendants of humanity have spread from star to star they won’t tell the children about the history of Ancient Earth until they’re old enough to bear it and when they learn they’ll weep to hear that such a thing as Death had ever once existed Eliezer Yudkowsky
5
It's a most peculiar psychology–this business of 'Science is based on faith too, so there! ' Typically this is said by people who claim that faith is a good thing. Then why do they say 'Science is based on faith too! ' in that angry-triumphal tone, rather than as a compliment? And a rather dangerous compliment to give, one would think, from their perspective. If science is based on 'faith', then science is of the same kind as religion–directly comparable. If science is a religion, it is the religion that heals the sick and reveals the secrets of the stars. It would make sense to say, 'The priests of science can blatantly, publicly, verifiably walk on the Moon as a faith-based miracle, and your priests' faith can't do the same.' Are you sure you wish to go there, oh faithist? Perhaps, on further reflection, you would prefer to retract this whole business of 'Science is a religion too! . Eliezer Yudkowsky
6
Then you get the wrong answer and you can't go to the Moon that way! Nature isn't a person, you can't trick them into believing something else, if you try to tell the Moon it's made of cheese you can argue for days and it won't change the Moon! What you're talking about is rationalization, like starting with a sheet of paper, moving straight down to the bottom line, using ink to write 'and therefore, the Moon is made of cheese', and then moving back up to write all sorts of clever arguments above. But either the Moon is made of cheese or it isn't. The moment you wrote the bottom line, it was already true or already false. Whether or not the whole sheet of paper ends up with the right conclusion or the wrong conclusion is fixed the instant you write down the bottom line. . Eliezer Yudkowsky
By far the greatest danger of Artificial Intelligence is that...
7
By far the greatest danger of Artificial Intelligence is that people conclude too early that they understand it. Eliezer Yudkowsky
8
Most Muggles lived in a world defined by the limits of what you could do with cars and telephones. Even though Muggle physics explicitly permitted possibilities like molecular nanotechnology or the Penrose process for extracting energy from black holes, most people filed that away in the same section of their brain that stored fairy tales and history books, well away from their personal realities: Long ago and far away, ever so long ago. Eliezer Yudkowsky
I only want power so I can get books.
9
I only want power so I can get books. Eliezer Yudkowsky
10
So the next time you doubt the strangeness of the future, remember how you were born in a hunter-gatherer tribe ten thousand years ago, when no one knew of Science at all. Remember how you were shocked, to the depths of your being, when Science explained the great and terrible sacred mysteries that you once revered so highly. Remember how you once believed that you could fly by eating the right mushrooms, and then you accepted with disappointment that you would never fly, and then you flew. Remember how you had always thought that slavery was right and proper, and then you changed your mind. Don't imagine how you could have predicted the change, for that is amnesia. Remember that, in fact, you did not guess. Remember how, century after century, the world changed in ways you did not guess. Maybe then you will be less shocked by what happens next. Eliezer Yudkowsky
11
And then Harry Potter had launched in to a speech that was inspiring, yet vague. A speech to the effect that Fred and George and Lee had tremendous potential if they could just learn to be weirder. To make people's live surreal, instead of just surprising them with the equivalents of buckets of water propped above doors. (Fred and George had exchanged interested looks, they'd never thought of that one.) Harry Potter had invoked a picture of the prank they'd pulled on Neville - which, Harry had mentioned with some remorse, the Sorting Hat had chewed him out on - but which must have made Neville doubt his own sanity. For Neville it would have felt like being suddendly transported into an alternate universe. The same way everyone else had felt when they'd seen Snape apologize. That was the true power of pranking. . Eliezer Yudkowsky
12
Without thinking about it at all, Harry stepped in front of Hermione.There was an intake of breath from behind him, and then a moment later Hermione brushed past and stepped in front of him. "Run, Harry! " she said. "Boys shouldn't have to be in danger. Eliezer Yudkowsky
13
But it is cute. It's such a boy thing to do. Drop dead. Aw, you say the most romantic things. Eliezer Yudkowsky
14
Your strength as a rationalist is your ability to be more confused by fiction than by reality. If you are equally good at explaining any outcome, you have zero knowledge. Eliezer Yudkowsky
15
Since the beginningnot one unusual thinghas ever happened. Eliezer Yudkowsky
16
Because the way people are built, Hermione, the way people are built to feel inside -" Harry put a hand over his own heart, in the anatomically correct position, then paused and moved his hand up to point toward his head at around the ear level, "- is that they hurt when they see their friends hurting. Someone inside their circle of concern, a member of their own tribe. That feeling has an off-switch, an off-switch labeled 'enemy' or 'foreigner' or sometimes just 'stranger'. That's how people are, if they don't learn otherwise. . Eliezer Yudkowsky
17
People form close friendships by knowing private things about each other, and the reason most people don't make close friends is because they're too embarrassed to share anything really important about themselves. Eliezer Yudkowsky
18
Wow, " the empty air finally said. "Wow. That puts a pretty different perspective on things, I have to say. I'm going to remember this the next time I feel an impulse to blame myself for something. Neville, the term in the literature for this is 'egocentric bias', it means that you experience everything about your own life but you don't get to experience everything else that happens in the world. There was way, way more going on than you running in front of me. You're going to spend weeks remembering that thing you did there for six seconds, I can tell, but nobody else is going to bother thinking about it. Other people spend a lot less time thinking about your past mistakes than you do, just because you're not the center of their worlds. I guarantee to you that nobody except you has even considered blaming Neville Longbottom for what happened to Hermione. Not for a fraction of a second. You are being, if you will pardon the phrase, a silly-dilly. Now shut up and say goodbye. Eliezer Yudkowsky
19
The AI does not hate you, nor does it love you, but you are made out of atoms which it can use for something else. Eliezer Yudkowsky
20
What people really believe doesn't feel like a BELIEF, it feels like the way the world IS. Eliezer Yudkowsky
21
For an instant Harry imagined... Just for an instant, before his imagination blew a fuse and called an emergency shut down and told him never to imagine that again. Eliezer Yudkowsky
22
A friend isn't someone you use once and then throw away, a friend is someone you use over and over again. Eliezer Yudkowsky
23
He wanted to write someone and demand a refund on his dark side which clearly ought to have irresistible magical power but had turned out to be defective. Eliezer Yudkowsky
24
You can only arrive at mastery by practicing the techniques you have learned, facing challenges and apprehending them, using to the fullest the tools you have been taught, until they shatter in your hands and you are left in the midst of wreckage absolute.. I cannot create masters. I have never known how to create masters. Go, then, and fail.. You have been shaped into something that may emerge from the wreckage, determined to remake your Art. I cannot create masters, but if you had not been taught, your chances would be less. The higher road begins after the Art seems to fail you; though the reality will be that it was you who failed your Art. . Eliezer Yudkowsky
25
To worship a sacred mystery was just to worship your own ignorance. Eliezer Yudkowsky
26
Every mystery ever solved had been a puzzle from the dawn of the human species right up until someone solved it. Eliezer Yudkowsky
27
I don't want to rule the universe. I just think it could be more sensibly organised. Eliezer Yudkowsky
28
All right, " Harry said coldly. "I'll answer your original question, then. You asked why Dark Wizards are afraid of death. Pretend, Headmaster, that you really believed in souls. Pretend that anyone could verify the existence of souls at any time, pretend that nobody cried at funerals because they knew their loved ones were still alive. Now can you imagine destroying a soul? Ripping it to shreds so that nothing remains to go on its next great adventure? Can you imagine what a terrible thing that would be, the worst crime that had ever been committed in the history of the universe, which you would do anything to prevent from happening even once? Because that's what Death really is - the annihilation of a soul! . Eliezer Yudkowsky
29
Ending up with that gigantic outsized brain must have taken some sort of runaway evolutionary process, something that would push and push without limits. And today's scientists had a pretty good guess at what that runaway evolutionary process had been. Harry had once read a famous book called Chimpanzee Politics. The book had described how an adult chimpanzee named Luit had confronted the aging alpha, Yeroen, with the help of a young, recently matured chimpanzee named Nikkie. Nikkie had not intervened directly in the fights between Luit and Yeroen, but had prevented Yeroen's other supporters in the tribe from coming to his aid, distracting them whenever a confrontation developed between Luit and Yeroen. And in time Luit had won, and become the new alpha, with Nikkie as the second most powerful...though it hadn't taken very long after that for Nikkie to form an alliance with the defeated Yeroen, overthrow Luit, and become the new new alpha. It really made you appreciate what millions of years of hominids trying to outwit each other - an evolutionary arms race without limit - had led to in the way of increased mental capacity.' Cause, y'know, a human would have totally seen that one coming. . Eliezer Yudkowsky
30
Illusion of transparency: We always know what we mean by our words, and so we expect others to know it too. Reading our own writing, the intended interpretation falls easily into place, guided by our knowledge of what we really meant. It’s hard to empathize with someone who must interpret blindly, guided only by the words. Be not too quick to blame those who misunderstand your perfectly clear sentences, spoken or written. Chances are, your words are more ambiguous than you think. Eliezer Yudkowsky
31
What does it mean to call for a “democratic” solution if you don’t have a conflict-resolution mechanism in mind? I think it means that you have said the word “democracy, ” so the audience is supposed to cheer. It’s not so much a propositional statement, as the equivalent of the “Applause” light that tells a studio audience when to clap. Eliezer Yudkowsky
32
Many have stood their ground and faced the darkness when it comes for them. Fewer come for the darkness and force it to face them. Eliezer Yudkowsky
33
Be careful... any time you find yourself defining the "winner" as someone other than the agent who is currently smiling from on top of a giant heap of utility. Eliezer Yudkowsky
34
Be careful... any time you find yourself defining the "winner" as someone other than the agent who is currently smiling from on top of a giant heap of utility Eliezer Yudkowsky
35
You couldn't changed history. But you could get it right to start with. Do something differently the FIRST time around. This whole business with seeking Slytherin's secrets.. seemed an awful lot like the sort of thing where, years later, you would look back and say, 'And THAT was where it all started to go wrong.' And he would wish desperately for the ability to fall back through time and make a different choice. Wish granted. Now what?. Eliezer Yudkowsky
36
Okay, so either (a) I just teleported somewhere else entirely (b) they can fold space like no one's business or (c) they are simply ignoring all the rules. Eliezer Yudkowsky
37
World domination is such an ugly phrase. I prefer to call it world optimisation. Eliezer Yudkowsky
38
Logic stays true, wherever you may go, So logic never tells you where you live. Eliezer Yudkowsky
39
I'm lazy! I hate work! Hate hard work in all its forms! Clever shortcuts, that's all I'm about! Eliezer Yudkowsky
40
Only God can tell a truly plausible lie. Eliezer Yudkowsky
41
When you are older, you will learn that the first and foremost thing which any ordinary person does is nothing. Eliezer Yudkowsky
42
Boys, " said Hermione Granger, "should not be allowed to love girls without asking them first! This is true in a number of ways and especially when it comes to gluing people to the ceiling! Eliezer Yudkowsky
43
Why does any kind of cynicism appeal to people? Because it seems like a mark of maturity, of sophistication, like you’ve seen everything and know better. Or because putting something down feels like pushing yourself up. Eliezer Yudkowsky
44
Like that's the only reason anyone would ever buy a first-aid kit? Don't take this the wrong way, Professor McGonagall, but what sort of crazy children are you used to dealing with?"" Gryffindors, " spat Professor McGonagall, the word carrying a freight of bitterness and despair that fell like an eternal curse on all youthful heroism and high spirits. Eliezer Yudkowsky
45
Someday, " said the Boy-Who-Lived, "when the distant descendants of Homo sapiens are looking back over the history of the galaxy and wondering how it all went so wrong, they will conclude that the original mistake was when someone taught Hermione Granger how to read. Eliezer Yudkowsky
46
The Headmaster told Professor Flitwick that this was, indeed, a secret and delicate matter of which he had already been informed, and that he did not think pressing it at this time would help me or anyone. Professor Flitwick started to say something about the Headmaster's usual plotting going much too far, and I had to interrupt at that point and explain that it had been my own idea and not anything the Headmaster forced me into, so Professor Flitwick spun around and started lecturing me, and the Headmaster interrupted him and said that as the Boy-Who-Lived I was doomed to have weird and dangerous adventures so I was safer if I got into them on purpose instead of waiting for them to happen by accident, and that was when Professor Flitwick threw up his little hands and started shrieking in a high-pitched voice at both of us about how he didn't care what we were cooking up together, but this wasn't ever to happen again for as long as I was in Ravenclaw House or he would have me thrown out and I could go to Gryffindor which was where all this Dumbledoring belonged - . Eliezer Yudkowsky
47
Lies propagate, that's what I'm saying. You've got to tell more lies to cover them up, lie about every fact that's connected to the first lie. And if you kept on lying, and you kept on trying to cover it up, sooner or later you'd even have to start lying about the general laws of thought. Like, someone is selling you some kind of alternative medicine that doesn't work, and any double-blind experimental study will confirm that it doesn't work. So if someone wants to go on defending the lie, they've got to get you to disbelieve in the experimental method. Like, the experimental method is just for merely scientific kinds of medicine, not amazing alternative medicine like theirs. Or a good and virtuous person should believe as strongly as they can, no matter what the evidence says. Or truth doesn't exist and there's no such thing as objective reality. A lot of common wisdom like that isn't just mistaken, it's anti-epistemology, it's systematically wrong. Every rule of rationality that tells you how to find the truth, there's someone out there who needs you to believe the opposite. If you once tell a lie, the truth is ever after your enemy; and there's a lot of people out there telling lies. Eliezer Yudkowsky
48
Sometimes Harry thought the deepest split in his personality wasn't anything to do with his dark side; rather it was the divide between the altruistic and forgiving Abstract Reasoning Harry, versus the frustrated and angry Harry In The Moment. Eliezer Yudkowsky
49
In a moral dilemma where you lost something either way, making the choice would feel bad either way, so you could temporarily save yourself a little mental pain by refusing to decide. At the cost of not being able to plan anything in advance, and at the cost of incurring a huge bias toward inaction or waiting until too late... Eliezer Yudkowsky
50
Textbook science is beautiful! Textbook science is comprehensible, unlike mere fascinating words that can never be truly beautiful. Elementary science textbooks describe simple theories, and simplicity is the core of scientific beauty. Fascinating words have no power, nor yet any meaning, without the math. Eliezer Yudkowsky
51
By far, the greatest danger of Artificial Intelligence is that people conclude too early that they understand it. Eliezer Yudkowsky
52
Anything that could give rise to smarter-than-human intelligence - in the form of Artificial Intelligence, brain-computer interfaces, or neuroscience-based human intelligence enhancement - wins hands down beyond contest as doing the most to change the world. Nothing else is even in the same league. Eliezer Yudkowsky
53
There's a popular concept of 'intelligence' as book smarts, like calculus or chess, as opposed to, say, social skills. So people say that 'it takes more than intelligence to succeed in human society.' But social skills reside in the brain, not the kidneys. Eliezer Yudkowsky
54
Intelligence is the source of technology. If we can use technology to improve intelligence, that closes the loop and potentially creates a positive feedback cycle. Eliezer Yudkowsky
55
The purest case of an intelligence explosion would be an Artificial Intelligence rewriting its own source code. The key idea is that if you can improve intelligence even a little, the process accelerates. It's a tipping point. Like trying to balance a pen on one end - as soon as it tilts even a little, it quickly falls the rest of the way. Eliezer Yudkowsky
56
When you think of intelligence, don't think of a college professor; think of human beings as opposed to chimpanzees. If you don't have human intelligence, you're not even in the game. Eliezer Yudkowsky
57
We tend to see individual differences instead of human universals. Thus, when someone says the word 'intelligence, ' we think of Einstein instead of humans. Eliezer Yudkowsky
58
Since the rise of Homo sapiens, human beings have been the smartest minds around. But very shortly - on a historical scale, that is - we can expect technology to break the upper bound on intelligence that has held for the last few tens of thousands of years. Eliezer Yudkowsky
59
I am a full-time Research Fellow at the Machine Intelligence Research Institute, a small 501(c)(3) public charity supported primarily by individual donations. Eliezer Yudkowsky
60
I keep trying to explain to people that the archetype of intelligence is not Dustin Hoffman in 'The Rain Man;' it is a human being, period. It is squishy things that explode in a vacuum, leaving footprints on their moon. Eliezer Yudkowsky
61
The purpose of a moral philosophy is not to look delightfully strange and counterintuitive or to provide employment to bioethicists. The purpose is to guide our choices toward life, health, beauty, happiness, fun, laughter, challenge, and learning. Eliezer Yudkowsky