Interview notes – Sam Altman on OpenAI, ChatGPT, Helion, Hermeus – StrictlyVC

He doesn’t read the news
Likes trolling on Twitter…”Twitter’s fun”
“Twitter’s gonna be fine”

HIS INVESTMENTS

400 personal investments, a few thousand including YC
All the companies he’s added value to are those he thinks about in his free time – while hiking, texting the founder an idea

Most successful investment? Stripe on a multiples basis

Helion – fusion energy
Personally invested $375M (!)
Other thing besides OpenAI that he spends a lot of time on
New energy system that works on super low cost
Hardest challenge is how to replace all (current) generative capacity on Earth really quickly
“Who can deliver energy the cheapest, and enough of it”
Simple machine, affordable cost, reasonable size
If fusion works…will change dynamics of what’s possible – enables more downstream (eg, more powerful planes)

Hermeus – supersonic jet company
Led $100M round
Was also involved with a competitor Boom – but different tech and approach
Huge market with multiple needs

Worldcoin
He’s a cofounder, on the board, but not day to day involved
Will tell its story soon – believes it will go over well (unlike earlier negative media coverage)
We give up more privacy to Facebook than Worldcoin
Phenomenal team
Launch in months
Interested in any tech to experiment with global UBI (versus what one country can do)

Re: crypto — “honestly not super interested”
“Love spirit of web3, but don’t intuitively feel why we need it”

Inception Fertility
In-vitro gametogenesis
In shadow of AI
Next 5-7 years of biotech will be remarkable
Human life extension – “yeah maybe that’s gonna work”

Investing for 20 years, president of YC for 5-6 years
Garry (new YC president) will do a lot of things differently and be wildly successful
Last few years were really hard for YC
YC can remake itself now – tourists are leaving now

ARTIFICIAL INTELLIGENCE

OpenAI has pulled together “most talent dense” AI team
“Gonna be tremendously good”

Why did ChatGPT and DALL-E so surprise people?
“Don’t know…reflected on it a lot”

If you make a good UX on top of something – believed users wanted to interact via dialogue
Pieces were there for awhile

Standard belief was AI would take over low skill / truck driving / generic white collar
Going exact opposite direction – taking over creativity where we thought humans might have special sauce
It’s not an intuitive finding

Released GPT-3 three years ago – thought ChatGPT would be incremental, was surprised by public reaction

ChatGPT will cause societal changes – eg, academic integrity
“Stakes are still relatively low”
Covid did show us society can update to massive changes faster than he expected

Given expected economic impact – “more gradual is better”

GPT-4 will come out when we’re confident we can do it safely and responsibly
Will release tech much more slowly than people will like
GPT-4 rumor mill is a ridiculous thing

Re: ChatGPT – built a thing, couldn’t figure out how to monetize it, put it out via an API, and users figured out how to use it

Would like to see AI super democratized, have several AGIs in the world
Cost of intelligence and energy trends down and down
“Massive surplus…benefits all of us”
Believes in capitalism – best service at lowest price

Society will need to agree on what AGI should never do
Broad absolute rules of the system
Within that, AI can do different things – safe for work one, edgier creative one – different values they enforce
A user can write up a spec of what they want, and AI will act according to it – “should be your AI…to serve you”

Microsoft – only tech company he’d be excited to partner with this deeply
Satya, Kevin Scott, Mary McHale
Values-aligned company

“We’re very much here to build AGI”
“We wanna be useful to people”

Re: Google’s AI – hasn’t seen it, assume they’re a competent org

We’re in a new world – generated text is something we all need to adapt to, like we adapted to calculators
“I’d much rather have ChatGPT teach me…than read a textbook”

Anthropic – rival AI, stressing an ethical layer
Very talented team
Multiple AGIs in the world is better than one

Society decided free speech is not quite absolute – in similar ways AI / LLMs will need to have bounds too

Video is coming… no confident prediction about when
Legitimate research project – could take awhile

AUDIENCE Q&A

When fusion online?
By 2028, could be plugging fusion generators into grid (pending regulators)

Re: AI worst and best case?
“best case is so unbelievably good that it’s hard to imagine, discovering new knowledge in a year instead of 70K years”
“Bad case is lights out for all of us”
More worried about accidental mis-use in short term, less about the AI itself being evil

How far away is AGI?
Much blurrier and gradual transition than people think

Re: state of San Francisco
Real shame we treat people like this
How elected leaders don’t fix the problem
Tech has some responsibility for it
But other cities do better than this
Super long in-person work and Bay Area

Re: ChatGPT reaction
Expected one order magnitude less hype, users, of everything
Less hype is probably better
The tech is impressive, but not robust
Use it 100x, see the weaknesses

How Sam uses ChatGPT?
Summarize super long documents, emails
For translation

Re: Google code red, threat to search
When people talk about new tech being end of a giant company, they’re usually long
Change is coming (for Google), but not as dramatically as people think

Before Google, memorizing facts was important – and now we’ll change again – and we’ll adapt faster than most people think

Prefers hybrid work, like YC – few days at home, few days in office
Skeptical that fully remote is thing everyone does
Most important companies will still be heavily in-person

Safety engineering for AI is different from standard safety engineering – consequences are greater, deserves more study

Raising capital now is hard, especially later stages – but other things easier – easier to rise above noise, hire, get customers
What he’d do now — recommends for founders – “do AI for some vertical”

Advice for AI startups
Differentiate by building deep relationships with users, some moat like network effect
Plan for AI models continually improving
OpenAI is a platform, but also wants to do a killer app (platform + killer app) to show people what’s possible

Podcast notes – ChatGPT goes prime time! – Practical AI (Daniel and Chris)

Practical AI 206: ChatGPT goes prime time! – Listen on Changelog.com

Hosts: Daniel Whitenack (Data scientist), Chris Benson (Lockheed Martin)

All about ChatGPT

Chris – feels collaborative, like having a partner

Lots of structuring in the output – bulleted lists, paragraphs

Humans get things wrong / incomplete all the time – yet we’re holding AI to a higher standard

Shifting to more open access – maybe in response to open source AI products like Stable Diffusion

Chris – expect to see more “fast followers” to ChatGPT soon

TECHNICALS

GPT language models – it’s a “causal language model” not a “mass language model”
Trained to predict next word in sequence of words, but based on all previous words
Auto-regressive – predicts next thing based on all previous things, and so forth
That’s why the text develops as if a human were typing
Few shot learning – can change answer style based on your questions and prompts

Zero shot = input that a model has never seen before
Few shot = provide a small number of inputs / prompts to guide the model

ChatGPT – trained with “reinforcement learning from human feedback” (RLHF)
Human preference is key part of this

How does it scale?
Human feedback is expensive
3 steps:
1. Pre-train a language model (aka a “policy”) – not based on human feedback
2. Gather human preference data to train a reward model – outputs prediction of human preference
3. Fine tune (1) based on (2)

eg for ChatGPT,
(1) is GPT3.5 (for ChatGPT)
(2) outputs data based on (1), add human labels of preference, and train a “reward model”

GPT3 is 100B+ parameters
ChatGPT reward model is 6B parameters

For (2), goal is to reduce harm by adding human feedback into the loop

For (3), will penalize if it strays too far from (1), and score output according to (2)
Try only to make small iterative changes (adiabatic)

What’s next?
Open research questions – (2) architecture and process hasn’t been fully optimized, lot to explore there
Will be new language models coming (eg, GPT4, Microsoft, Google) – trying different (1) and (2)

Chris Albon tweet:

Sci-fi got it wrong.

We assumed AI would be super logical and humans would provide creativity.

But in reality it’s the opposite. Generative AI is good at getting an approximately correct output, but if you need precision and accuracy you need a human.

The Venn of blockchain and AI

I’ve been thinking about the relationship between blockchains and AI lately. Both are emerging foundational technologies and I think it’s no accident they are both coming of age at the same time.

Multiple writers have already expressed this view:

AIs can be used to generate “deep fakes” while cryptographic techniques can be used to reliably authenticate things against such fakery. Flipping it around, crypto is a target-rich environment for scammers and hackers, and machine learning can be used to audit crypto code for vulnerabilities. I am convinced there is something deeper going on here. This reeks of real yin-yangery that extends to the roots of computing somehow

From Venkatesh Rao: https://studio.ribbonfarm.com/p/the-dawn-of-mediocre-computing

I think AI and Web3 are two sides of the same coin. As machines increasingly do the work that humans used to do, we will need tools to manage our identity and our humanity. Web3 is producing those tools and some of us are already using them to write, tweet/cast, make and collect art, and do a host of other things that machines can also do. Web3 will be the human place to do these things when machines start corrupting the traditional places we do/did these things.

From Fred Wilson: https://avc.com/2022/12/sign-everything/

In both writers’ examples, blockchain helps solve some of the problems that AI creates, and vice-versa. I’m reminded of Kevin Kelly who said, Each new technology creates more problems than it solves.

Blockchains and AI have a sort of weird and emergent technological symbiosis and I’m here for it.

So the brain flatulence below is just my way to think aloud, using the writing process to work through the question(s).

*Note: when I say “blockchain”, I include what Fred Wilson calls web3 and Venkatesh calls crypto; there are just a few canonical applications that we’re all familiar with (namely, bitcoin and ethereum); and when I say “AI”, I am thinking about the most popular machine learning models like GPT3 and Stable Diffusion

*Note also: I am just a humble user of these new and powerful AI tools, and can barely understand the abstract of a typical machine learning research paper; so part of the reason why I’m writing this is to find out where I’m wrong a la Cunningham’s law

A blockchain is a tool for individual sovereignty; while an AI is a tool for individual creativity

A blockchain operates at maximum transparency; while an AI operates largely as a black box

A blockchain clearly shows the chain of ownership and history; while an AI… (does something like the opposite in the way it aggregates and melds and mutates as much data as possible?)

A blockchain is “trustless”, in the sense that what you see on-chain is the agreed upon “truth” of all its users; while an AI is (?), in the sense that what it generates is more or less unique to the specific prompt / question / user (and even this can change as the model is updated, or new data is added]

An AI is much easier to use than a blockchain

An AI can create vast quantities of content, very cheaply; while a (truly “decentralized”) blockchain is limited by scalability and cost

An AI is centralized (to a specific company, or model, or data set) in the sense that decision making rests with a team or company; while a blockchain is decentralized and decision making is distributed

A surprising user experience – as in, an unexpected but delightful output – is typically net positive for a user of AI, while seeing something happen on a blockchain that you don’t expect would generally be pretty bad (yes, of course there are airdrops)

Blockchains are a competitive threat to industries with a high degree of centralization (such as fiat currency issuance, and payment networks); AI is a competitive threat to many individual online workers (such as language translators, and freelance writers, and basic QA/QC employees)

Both blockchains and AI have multiple open source products that can be forked by developers

Both blockchains and AI are platforms upon which many other products and services can be built

Both blockchains and AI are technologies that exploded into the popular consciousness in the last 10 years

Both “blockchain” and “AI” are very broad suitcase words, in part because they are both the product of many technologies combined in innovative ways: for blockchain that is everything from cryptography to smart contract programming to PoW mining to distributed consensus mechanisms; for AI that’s, uh…well everything listed here and more, I suppose

I’ll end here for now, but let me know what I got wrong, what I’m missing, and what questions or ideas this might inspire

Addendum #1: I asked ChatGPT

Addendum #2: This NYT article notes that SBF (“Sam Bankscam Fraud”) donated at least $500M to organizations researching AI alignment and AI safety. Not exactly the kind of symbiosis I want to explore, but worth noting.

Addendum #3:

Blockchains can only give precise answers, while AI can give approximate answers or even fabricate answers

Blockchains are censorship resistant, while AI is centralized (most are created by small doxxed teams) and have implemented restrictions on usage (most have rules against, for example, CSAM or nudity)