Jump to content

Sam Altman

From Wikiquote
Sam Harris Altman in 2019

Samuel Harris Altman (born 1985) is an American entrepreneur, investor, and programmer. He is the CEO of OpenAI and was the co-founder of Loopt. He previously was the president of Y Combinator and was briefly the CEO of Reddit.

Quotes

[edit]

2017

[edit]

"The Merge"

[edit]
Sam Altman (December 7, 2017). "The Merge". (origial Medium post)
  • Perhaps another reason people stopped using the word “singularity” is that it implies a single moment in time, and it now looks like the merge is going to be a gradual process. And gradual processes are hard to notice.
    I believe the merge has already started, and we are a few years in. Our phones control us and tell us what to do when; social media feeds determine how we feel; search engines decide what we think.
  • We are already in the phase of co-evolution — the AIs affect, effect, and infect us, and then we improve the AI. We build more computing power and run the AI on it, and it figures out how to build even better chips.
  • More important than that, unless we destroy ourselves first, superhuman AI is going to happen, genetic enhancement is going to happen, and brain-machine interfaces are going to happen. It is a failure of human imagination and human arrogance to assume that we will never build things smarter than ourselves.
  • The merge can take a lot of forms: We could plug electrodes into our brains, or we could all just become really close friends with a chatbot. But I think a merge is probably our best-case scenario. If two different species both want the same thing and only one can have it—in this case, to be the dominant species on the planet and beyond—they are going to have conflict. We should all want one team where all members care about the well-being of everyone else.
  • Although the merge has already begun, it’s going to get a lot weirder. We will be the first species ever to design our own descendants. My guess is that we can either be the biological bootloader for digital intelligence and then fade into an evolutionary tree branch, or we can figure out what a successful merge looks like.
  • It’s probably going to happen sooner than most people think. Hardware is improving at an exponential rate—the most surprising thing I’ve learned working on OpenAI is just how correlated increasing computing power and AI breakthroughs are—and the number of smart people working on AI is increasing exponentially as well. Double exponential functions get away from you fast.

2023

[edit]
  • I think AGI will be the best tool humanity has yet created. With it, we will be able to solve all sorts of problems. We'll be able to express ourselves in new creative ways. We'll make just incredible things for each other, for ourselves, for the world, for kind of this unfolding human story. And it's new, and anything new comes with change and change is not always all easy. But I think this will be just absolutely tremendous upside. And in nine more years if you're nice enough to invite me back, you'll roll this question and people will say, "How could we have thought we didn't want this?"
  • I think what we believe in very strongly, is that keeping the rate of change in the world relatively constant, rather than, say, go build AGI in secret and then deploy it all at once when you’re done, is much better. This idea that people relatively gradually have time to get used to this incredible new thing that is going to transform so much of the world, get a feel for it, have time to update. You know, institutions and people do not update very well overnight. They need to be part of its evolution, to provide critical feedback, to tell us when we’re doing dumb mistakes, to find the areas of great benefit and potential harm, to make our mistakes and learn our lessons when the stakes are lower than they will be in the future. Although we still would like to avoid them as much as we can, of course. And I don’t just mean we, I mean the field as a whole, sort of understanding, as with any new technology, where the tricky parts are going to be..
  • In a well functioning society, governments would be doing the AGI project and [nuclear] fusion and a whole bunch of things — and yet they’re not. So we either sit around and watch the gradual decline of state capacity and say ‘that’s a bummer’ and we’re just not going to have any more technical progress . . . or you do the next best thing and just build great companies.
  • Is [AI] gonna be like the printing press that diffused knowledge, power, and learning widely across the landscape that empowered ordinary, everyday individuals that led to greater flourishing, that led above all two greater liberty? Or is it gonna be more like the atom bomb – huge technological breakthrough, but the consequences (severe, terrible) continue to haunt us to this day?
  • We face serious risk. We face existential risk. The challenge that the world has is how we’re going to manage those risks and make sure we still get to enjoy those tremendous benefits. No one wants to destroy the world. Let's make sure we come together as a globe — and I hope this place can play a real role in this. We talk about the IAEA as a model where the world has said 'OK, very dangerous technology, let's all put some guard rails.' And I think we can do both. I think in this case, it's a nuanced message 'cause it's saying it's not that dangerous today but it can get dangerous fast. But we can thread that needle.
  • Trust the exponential. Flat looking backwards, vertical looking forwards.
    • Conversation on Twitter with Elon Musk (December 3, 2022)

2024

[edit]
  • One area that I'm particularly interested personally in open source for is I want an open source model that is as good as it can be, that runs on my phone, and that I think is going to, you know...the world doesn't quite have the technology for a good version of that yet, but that seems like a really important thing to go do at some point.

2025

[edit]
  • When bubbles occur, smart people get overly excited about a grain of truth. If you look at most bubbles in history, like the technology bubble, there was something real there. Technology was really important. The internet was really a big thing. People got too excited about it.
  • All of these claims are utterly untrue.
    • statement issued by Sam Altman and other immediate family members in response to his sister's allegations of abuse, as reported in The Indianapolis Star, CNBC, BBC and other secondary sources

Quotes about Altman

[edit]
[edit]
Wikipedia
Wikipedia
Wikipedia has an article about:

Media related to Sam Altman on Wikimedia Commons