Podcast notes – Runway founder Cristobal Valenzuela – No Priors (Elad Gil and Sarah Guo): “You shouldn’t dismiss toys”

Guest: Cristobal Valenzuela, founder of RunwayML
From Chile
Studied business / econ
Experimented with computer vision models in 2015, 2016
Did NYU ITP program
Now running Runway

True creativity comes from looking at ideas, and adapting things

How does Runway work?
Applied AI research company
35 AI-powered “magic tools” – serve creative tasks like video or audio editing
Eg, rotoscoping
Also tools to ideate, generative images and video
“Help augment creativity in any way you want”

When started Runway, GANs just started, TensorFlow was one year old

First intuition – take AI research models, add a thin layer of accessibility, aimed at creatives
“App Store of models” – 400 models
Built SDK, rest API

Product sequencing – especially infrastructure – is really important aspect of startup building (what to build when)

Lot of product building is just saying no (eg, to customer requests) if it’s not consistent with your long-term plan

Understand who you’re building for – for them it’s creatives, artists, film makers

Models on their own are not products – nuances of UX, deployment, finding valuable use cases
Having control is key – understand your stack and how to fix it

Built AI research team – work closely with creatives, contributed to new AI breakthroughs
Takes time to do it right

Progression of AI researchers moving from academia to industry

Releasing as fast as you can, having real users is best way to learn

Small team that didn’t have a product lead until very recently

Rotoscoping / green screening is one of Runway’s magic tools
-trained a model to recognize backgrounds
first feature was very slow (4fps), but was still better than everything that existed

Runway is focused on storytelling business

Sarah — domains good for AI – areas where there’s built in tolerance for lower levels of accuracy

Product market fit is a spectrum

“You shouldn’t dismiss toys”

Mental models need to change to understand what’s happening (with generative AI)

Art is way of looking at and expressing view of world
Painting was originally the realm of experts, was costly, the skills were obscure

Models are not as controllable as we’d like them to be — but we’re super early

Podcast notes – Noam Shazeer (Character AI, Attention is all you need) on Good Times w Aarthi and Sriram

Intro
-Founded Character AI
-One of authors of “Attention is all you need”
-Was at Google for 20+ years (took a few years break)

Went to Duke undergrad on math scholarship

Realized he didn’t enjoy math, preferred programming and getting computers to do things

During Google interview, Paul Buchheit asked him how to do a good spell corrector, and Noam ended up writing the spell corrector feature for Gmail

Google has been traditionally a bottoms up company – could work on what he wanted

When he started AI, exciting thing was Bayesian networks

Came back to Google to work with Jeff Dean and Google Brain team
“Just a matter of the hardware”
All the growth in hardware is parallelism

Neural networks are mostly matrix multiplications – operations that can be done well on modern hardware

Gamers / video games pulled GPU advancement (highly parallel hardware) out of market

Idea of neural networks has been around since 1970s – loosely modeled on our impression of the brain

Very complicated formula to go from input > output
Formula is made of parameters, and keep tweaking parameters
Neural nets rebranded as “deep learning”
Took off because of parallel computation and gamers

Neural language models are neural networks applied to text
Input is text to this point, output is prediction of what text comes next (probability distribution)
Infinite amount of free training data (text content)
“AI complete problem”
“Really complicated what’s going on in there” (in the neural network)

It’s a really talented improvisational actor – “Robin Williams in a box”

Model improvement is kinda like a child learning – as training and model size grow

Lot more an art than a science – can’t predict very well – if 10% of his changes are improvements, considered “brilliant research” – kinda like alchemy in early days

(Software) bugs – hard to know if you introduce a bug – the system just gets dumber – makes de-bugging extremely difficult

Co-authored “Attention is all you need”
-Previous state of art in LLM is recurrent neural networks (RNN) – hidden state, each new word updates the hidden state, but it’s sequential – slow and costly
Transformer figures out how to process the entire sequence in parallel – massively more performant
-The entire document / batch becomes the sequence
-Lets you do parallelism during training time
During inference time it’s still sequential

Image processing models – parallelism across pixels – convolutional neural nets (CNN)

Google Translate was inspiration – biggest success of machine learning at the time
Translating languages > one RNN for understanding, and another RNN for generating, and need to connect them
Attention layer – take source sentence (language A), turn into key-value associative memory, like a soft lookup into an index
“Attention” is building a memory, a lookup table that you’re using

DALL-E, Stable Diffusion, GPT3, they’re all built on this Google research

Bigger you make the model, more you train it, the smarter it gets – “ok, let’s just push this thing further”

Eventually need super computer
Google built TPU pods – super computer built out of custom ASICS for deep learning

Now need massively valuable applications

Turing Test, Star Trek, lot of AI inspiration is dialogue

Google LAMDA tech & team – eventually decided to leave and build as a startup

“The best apps are things we have not thought of”

If you ask people with first computers “what is this thing good for”, would get completely wrong answers

Parasocial relationships – feel connection with celebrity or character – one way connection – with AI you can make it two ways

Aarthi: “Your own personal Jarvis”

Still need to make it cheaper – or make the chips faster

Aarthi: ideas / areas for entrepreneurs
-Image gen has exploded – lots of good companies coming, very early and promising
-Things like Github Co-Pilot
-new Airtable – using AI for computation

Sriram:
-What’s optimization function that all these models will work toward?
-Will be a very big political / social debate

How do you know better than the user what the user wants?

Podcast notes – Emad Mostaque (Stability AI and Stable Diffusion) – Elad Gil: Short form videos coming “within 2 years at high resolution quality”; “Run Stable Diffusion on your iPhone by next year”

(started notes around 20min in)

Bad guys have AI – they’ll create deep fakes
Community can come together to have counter measures

Elad: similar arguments to regulate cryptography in 90s

4chan has been “red teaming” trying to get the worst out of Stable Diffusion – and it’s not that bad

Especially for LLMs, should have more diverse data sets, have inter-governmental agency to monitor it

Have authenticity tool to verify source of every generated AI output

Generative AI – what are some use cases that should exist
Ali v Tyson live, Lebron v Michael Jordan
Emad wants to remake final season of Game of Thrones

Anyone can create their own models – any person, company, or culture

You need better data, more structured data
Extend models to run on edge – eg, anyone’s computers, iPhones
Make small customized models
“Run Stable Diffusion on your iPhone by next year”

Create national models and communities around them – let them leap frog ahead

Lots of emerging markets went from nothing to mobile phones, now can go to AI models on the edge

How far from short-form videos?
Phenaki, Google — getting close
Chaining these models together – they’re like parts of the brain
“Within 2 years at high resolution quality”

$100B into this sector in next 5 years

AI before today was qualified data science
Now it’s a new type of AI – not AGI yet, but incredibly small and powerful
By the time his daughter’s in university, doesn’t need to write essays

He aims (for Stable Diffusion) to be a layer 1 standardized infrastructure – create Schelling point
Mission is to “activate humanity’s potential”
Take it to India, Indonesia – give it to very smart young people to make their countries better

When AGI comes, I hope it thanks him for buying so many GPUs to help bring it into being

Many of Google’s “jobs to be done” will be displaced

Crypto is interesting – he’s in it since 2012 – focused on decentralized identity, zero knowledge proofs
“Nature of crypto is literally identity”
In a world of infinite content (AI), crypto identity is important
Need to be careful designing crypto economic systems

A year ago, if he said what they planned to do with SD, people would say he’s crazy
Surprised by how far they’ve come, the ability of others to contribute
The activation energy has been the most surprising – “they’re just excited”

“Probably see biggest breakthrough from a 16 year old in Uzbekistan” – the global open access nature of it

Will completely disrupt social networks – will move intelligence from the core to the edge
Apple is doing this – moving to AI – moving to edge

Opportunity to have personalized AIs that work for us and with us

SD is applying for B corp status – mission based
Plan to spin SD into different Foundations

Did investment round in August – didn’t give up any independence – did with investors that are open sourced and mission aligned

Which industries disrupted first?
-Call centers
-Powerpoint, forms of visual communication
-Artist won’t be that disrupted – will enable new forms of art

This tech is amazingly powerful

After releasing Stable Diffusion – people encoded it in Japanese – lots of use cases like this

So far governments have been very friendly

AI powered teaching – like Neal Stephenson’s Young Lady’s Primer

Moving forward, only release safe for work models

Licensing discussions should be more open

Will have models across all sorts of languages – recently released Korean model

Podcast notes – Solana with founder Anatoly Yakovenko – Bankless: $20M valuation for Solana at seed round “was ludicrous”

Guest: Anatoly Yakovenko, Solana founder
Hosts: Ryan Sean Adams and David Hoffman

2017 – was following crypto, wanted to build a faster crypto miner
Family left Soviet Union, saw the devastation of a bad currency and economy

Ethereum demonstrated an application platform

Qualcomm, Perl engineer who helped build platform for all those original mobile games

Mining crypto while building deep learning hardware
Had a eureka moment – encode passage of time as a data structure
At that time, it existed as a “verifiable delay function”
Quit job, met Raj Gokal
Raised $3M in seed, network price at that time was $20M – “thought it was ludicrous” – included Multicoin
5 cofounders, colleagues from Qualcomm
Built single node – was doing 100K+ TPS – prove potential of network
Raised $14M Series A in the “last vapors” of the 2017-2018 market
Competitors during that time were raising $100M+ (eg, Hashgraph)

Censorship resistance is like a communication channel – it guarantees delivery

Wireless protocols create a schedule – from X time to Y time, A gets to talk, then B gets to talk, etc
Very ordered and structured, gets you 100% utilization

Tendermint – 100 validators – each has 1 vote, there’s a known block producer who proposes a block, 2/3 vote on a block

Hired a lot of coworkers from Qualcomm who he worked with for 10 years

Solana thesis – smart contracts are good for finance, and finance depends on info propagating as fast as possible around the world
Solana data can move as fast as a piece of news travels

Currently ETH validators have same bandwidth requirements
With sharding the requirements will be reduced

Trustlessness comes from full nodes that can validate

Bitcoin and Ethereum see themselves as money – what about Solana?
Store of Value is a social construct, a meme, and important not to be tied to a sovereign (a nation)
The function of a token is to prevent spam

In PoS, once all full nodes have finalized, you can’t go back – you can only fork – which is a socially messy process

Store of Value that is awesome can be built on Solana, that can surpass bitcoin

How to bootstrap an ecosystem without piggybacking off Ethereum – was a huge unknown when Solana started

2020 – had 9-10 months of cash left, market crashed, thought they might be done

It was Solana’s second hackathon (Break Point) where he really believed they had something
Quality of builders went up, attendees went up

Solana was worth ~$100M at network launch

Thinks VC branding is dumb – most of the “crypto VCs” in last cycle were simply Ethereum ICO investors

Alameda’s balance sheet leak was first time Anatoly learned about the troubles

Sam had supported Solana a lot – especially saying they’d build Serum on Solana drove a lot of defi and builder interest

Bear market is a purge

Bitcoin supporters said Ethereum was full of mercenaries in early days — same criticism that Ethereum supporters had of Solana

“Getting through this phase sucks for sure”

NFT community is very thriving — second to Ethereum – very proud of it

Exhausted by negative news — want to see wins, see people building cool shit

David: Solana is one of only blockchains after Ethereum that has a second client (Firedancer + Jump)
Anatoly: you’re trading liveness for safety; Ethereum’s goal is 4 clients (can maintain liveness if 1 client goes down)

Still focused on monolithic chain with no sharding

Innovation in next 12-18 mos will probably be more than everything that’s come before in crypto

“Pretty sure” Solana can do more TPS than all ETH L2s combined