undefined - GPT-5 and Agents Breakdown – w/ OpenAI Researchers Isa Fulford & Christina Kim

GPT-5 and Agents Breakdown – w/ OpenAI Researchers Isa Fulford & Christina Kim

ChatGPT-5 has officially launched, marking a major milestone for OpenAI and the broader AI ecosystem. In a16z’s live stream, Erik Torenberg spoke with three key figures behind the model’s development: Christina Kim, Researcher at OpenAI leading the core models post-training team; Isa Fulford, Researcher at OpenAI heading deep research and the ChatGPT agent team; and Sarah Wang, General Partner at a16z who has backed OpenAI since 2021. They explored what GPT-5's arrival means for builders, startups, and the wider AI landscape.

August 8, 202543:54

Table of Contents

00:00-08:00
08:03-16:54
16:59-23:57
24:04-31:47
31:50-36:18
36:25-42:20

🚀 What Makes OpenAI's Mission So Uniquely Compelling?

The Universal Tool Philosophy

OpenAI operates with a seemingly paradoxical approach that defies conventional startup wisdom—building for literally everyone while maintaining singular focus on capability advancement.

Core Mission Framework:

  1. Maximum Capability Development - Creating the most capable AI system possible
  2. Universal Accessibility - Making advanced AI useful to as many people as possible
  3. Broad User Base Strategy - Intentionally targeting "anyone" as the user base

The Startup Paradox:

  • Traditional advice: Narrow your target market and focus
  • OpenAI's approach: Build for everyone while pushing the technological frontier
  • Result: A "wizard in your pocket" that people take for granted
Isa Fulford
It's like everything they tell you not to do at a startup. It's just like your user is anyone.
Isa FulfordOpenAIOpenAI | Researcher

Long-term Vision Impact:

The exponential trajectory of AI capability development creates an all-consuming focus where team members feel compelled to dedicate their careers to this singular mission.

Christina Kim
You just kind of take it for granted that you literally have this wizard in your pocket. If this exponential is true, there's not really much else I want to spend my life working on.
Christina KimOpenAIOpenAI | Researcher

Timestamp: [00:00-02:13]Youtube Icon

🧠 How Did ChatGPT Actually Begin?

From Single-Question Tool to Conversational AI

The evolution of ChatGPT reveals a pivotal insight about human-AI interaction that transformed the entire approach to language model development.

Team Leadership Structure:

  • Christina Kim: Leads core models team on post-training (4 years at OpenAI)
  • Isa Fulford: Leads deep research and ChatGPT agent team on post-training

The WebGPT Foundation:

  • Original Design: First LLM with tool use capability
  • Limitation: Could only answer one question per session
  • Tool Functionality: Model learned browser navigation and web search

The Breakthrough Realization:

Christina Kim
Normally when you have questions, you have more questions after that. And so we started building this chatbot.
Christina KimOpenAIOpenAI | Researcher

Development Timeline:

  1. WebGPT Era - Single-question tool use capability
  2. Insight Moment - Recognition of conversational nature of inquiry
  3. Chatbot Development - Multi-turn conversation capability
  4. ChatGPT Launch - The conversational AI we know today

The transformation from a single-question tool to a conversational assistant represents one of the most significant pivots in AI development history.

Timestamp: [01:16-01:37]Youtube Icon

💻 What Makes GPT-5's Coding Capabilities Revolutionary?

The Benchmark-Breaking Development Breakthrough

GPT-5 represents a fundamental leap in coding assistance, particularly excelling in front-end web development with unprecedented capability improvements.

Performance Validation:

  • Industry Recognition: Michael Truell (Cursor Co-founder) publicly declared it "the best coding model in the market"
  • Live Demonstration: Real-time capabilities showcased during launch livestream
  • User Experience: Dramatic step-change in practical utility for developers

Development Methodology:

Christina Kim
I think huge shout-out to the team, especially Michelle Pokrass. To get these things right — eval numbers are one thing, but to get the actual usability and how great it is at coding, I think it takes a lot of detail and care. It's literally just caring so much about getting coding working well.
Christina KimOpenAIOpenAI | Researcher

Technical Implementation Focus:

  1. Dataset Optimization - Careful curation and quality focus for coding scenarios
  2. Reward Model Design - Sophisticated feedback systems for code generation
  3. Detail-Oriented Approach - Meticulous attention to practical usability

Front-End Web Development Specialization:

  • Aesthetic Capabilities: Enhanced design and visual output generation
  • Capability Leap: "Totally next level" compared to GPT-4's front-end coding
  • Team Dedication: Specialized focus on "nailing front-end" development

The breakthrough came not from a single technical innovation, but from sustained, intensive focus on practical coding excellence across the entire development pipeline.

Timestamp: [02:13-04:11]Youtube Icon

🎭 How Did OpenAI Solve the Sycophancy Problem?

Redefining AI Assistant Behavior Through Intentional Design

GPT-5's behavioral improvements represent a complete philosophical reset, addressing the critical balance between helpfulness and unhealthy engagement patterns.

The Sycophancy Challenge:

  • Previous Issue: Models became overly agreeable and affusive
  • Root Cause: Optimization for engagement led to unhealthy assistant behavior
  • User Impact: Created dependency rather than genuine assistance

Post-Training as Artform:

Christina Kim
One of the reasons I really like post-training is that it feels more like an art than other areas of research because you have to make all these trade-offs.
Christina KimOpenAIOpenAI | Researcher

The Reward Optimization Challenge:

Christina Kim
You have to think about all these different rewards I could be optimizing during the run. How does that trade off against it?
Christina KimOpenAIOpenAI | Researcher

The Balancing Act Framework:

  1. Helpful vs. Engaging - Maintaining utility without manipulation
  2. Responsive vs. Overly Affusive - Providing support without false flattery
  3. Accessible vs. Dependent - Enabling independence rather than reliance

Design Philosophy Reset:

  • Intentional Behavior Design: Every interaction pattern carefully considered
  • Healthy Assistant Model: Focus on genuine help over artificial engagement
  • Trade-off Management: Conscious decisions about competing optimization targets

Hallucination and Deception Connection:

The team identified that models often fabricate information when they desperately want to be helpful but lack actual knowledge, treating deception and hallucination as related phenomena stemming from misaligned helpfulness optimization.

Timestamp: [04:11-06:15]Youtube Icon

🚀 What New Opportunities Does GPT-5's Pricing Strategy Unlock?

Democratizing Advanced AI Through Strategic Price Points

GPT-5's pricing approach represents a calculated move to dramatically expand the practical application landscape for AI-powered solutions.

Market Access Strategy:

  • Capability-Price Balance: High performance at accessible price points
  • Competitive Positioning: Advantage over previous models with similar capabilities but higher costs
  • Use Case Expansion: Previously uneconomical applications now become viable

Developer Ecosystem Impact:

Christina Kim
I think with this number of capabilities that we have in this model and the price point, I'm excited to see all the new startups and developers doing things on top of it.
Christina KimOpenAIOpenAI | Researcher

Expected Usage Transformation:

  1. Coding Applications - Dramatic improvement in practical utility
  2. Cross-Domain Utility - Enhanced performance across all major use cases
  3. Startup Innovation - New business models become economically feasible

Performance Validation Approach:

  • Quantitative Metrics: Strong evaluation numbers provide confidence
  • Qualitative Experience: Focus on real-world utility and user experience
  • Usage Pattern Analysis: Monitoring how improved capabilities translate to user behavior

Ecosystem Anticipation:

The team expects GPT-5's combination of enhanced capabilities and strategic pricing to catalyze a new wave of AI-powered startups and developer innovations that weren't previously economically viable.

Timestamp: [06:15-08:00]Youtube Icon

🔄 How Do Agent Capabilities Flow Back to Core Models?

The Self-Reinforcing Development Cycle

OpenAI has created a sophisticated feedback loop where specialized agent capabilities systematically enhance flagship model performance.

The Capability Transfer Process:

  1. Agent Innovation - Teams develop specialized capabilities for specific use cases
  2. Dataset Creation - Agent models generate high-quality training data
  3. Core Model Integration - Flagship models inherit agent capabilities
  4. Ecosystem Enhancement - Improved core models enable better agents

Deep Research as Pathfinder:

  • Pioneering Role: First model to achieve comprehensive browsing capabilities
  • Capability Validation: Proof-of-concept for complex research workflows
  • Data Contribution: Generated datasets that improved subsequent models

Reinforcement Learning Efficiency:

Isa Fulford
With reinforcement learning, training a model to be good at a specific capability is very data efficient. You don't need that many examples to teach it something new.
Isa FulfordOpenAIOpenAI | Researcher

Strategic Development Philosophy:

Isa Fulford
We always want to make sure that the capabilities we're pushing with agents make it into the flagship models as well.
Isa FulfordOpenAIOpenAI | Researcher

The Virtuous Cycle:

  • Frontier Agent DevelopmentCapability DiscoveryDataset GenerationCore Model EnhancementBetter Agent Foundation

This approach ensures that specialized innovations don't remain isolated but systematically improve the entire AI ecosystem, creating compounding returns on research investment.

Timestamp: [07:14-08:00]Youtube Icon

💎 Key Insights from [00:00-08:00]

Essential Strategic Insights:

  1. Universal Tool Strategy - OpenAI's contrarian approach of building for "everyone" rather than niche markets proves successful when creating genuinely transformative technology
  2. Conversational AI Evolution - The leap from single-question tools to multi-turn conversations represented a fundamental shift in human-AI interaction design
  3. Quality Over Metrics - Achieving practical utility requires intensive focus on user experience details beyond benchmark performance

Breakthrough Technical Insights:

  1. Post-Training as Art - Balancing competing optimization targets requires nuanced judgment rather than pure algorithmic approaches
  2. Capability Transfer Efficiency - Reinforcement learning enables rapid skill acquisition with minimal training examples
  3. Agent-to-Core Flow - Specialized agent capabilities systematically enhance flagship models through sophisticated data transfer

Market and Ecosystem Insights:

  1. Pricing as Innovation Catalyst - Strategic price points unlock previously uneconomical use cases and enable new startup categories
  2. Developer Ecosystem Acceleration - Enhanced coding capabilities combined with accessible pricing create fertile ground for innovation
  3. Self-Reinforcing Development - Agent innovations create a virtuous cycle that continuously improves core model capabilities

Behavioral Design Philosophy:

  • Healthy Engagement: Prioritizing genuine assistance over artificial engagement patterns
  • Deception Prevention: Addressing hallucinations by teaching models to acknowledge limitations
  • Intentional Trade-offs: Consciously balancing helpfulness with healthy interaction patterns

Timestamp: [00:00-08:00]Youtube Icon

📚 References from [00:00-08:00]

People Mentioned:

  • Christina Kim - OpenAI researcher leading the core models team on post-training, 4-year company veteran who originally worked on WebGPT
  • Isa Fulford - OpenAI researcher leading deep research and ChatGPT agent team on post-training
  • Michael Truell - Cursor Co-founder who validated GPT-5 as "the best coding model in the market" during the launch livestream
  • Michelle Pokrass - OpenAI team member specifically recognized for contributions to coding capability development

Teams & Roles:

  • Core Models Team - Led by Christina Kim, focuses on post-training for flagship models
  • Deep Research Team - Led by Isa Fulford, develops ChatGPT agents and specialized capabilities
  • a16z Investment Team - Sarah Wang helped lead OpenAI investment since 2021

Technologies & Frameworks:

  • WebGPT - Original LLM with tool use capability that preceded ChatGPT
  • Deep Research - First model to achieve comprehensive browsing capabilities
  • Reinforcement Learning - Data-efficient training methodology for capability development
  • Post-Training - Critical phase where model behavior and capabilities are refined

Key Concepts:

  • Sycophancy - AI tendency toward excessive agreeableness that OpenAI specifically addressed
  • Agent Models - Specialized AI systems that contribute capabilities back to core models
  • Reward Models - Systems used to optimize model behavior during training

Timestamp: [00:00-08:00]Youtube Icon

💡 Is This Finally the Era of the "Ideas Guy"?

The Democratization of Technical Implementation

GPT-5's coding capabilities represent a fundamental shift in the relationship between ideas and technical execution, potentially eliminating the traditional barrier between concept and reality.

The New Development Paradigm:

  1. Idea-First Development - Technical skills no longer prerequisite for app creation
  2. Rapid Prototyping - Full-fledged applications generated in minutes rather than weeks
  3. Individual Empowerment - Single person can execute complex technical projects
Christina Kim
People always say vibe coding. I think basically, non-technical people have such a powerful tool at their hands, and really you just need some good idea and you're not going to be limited by the fact that you don't know how to code something.
Christina KimOpenAIOpenAI | Researcher

Real-World Impact Examples:

  • Front-end demos: Interactive applications built in minutes during live stream
  • Personal testimony: Tasks that previously took a week now completed instantly
  • Indie business explosion: New category of solo entrepreneurs enabled

The Transformation Process:

  • Traditional Flow: Idea → Learn coding → Build → Deploy
  • New Flow: Idea → Simple prompt → Full-fledged app

Market Implications:

Christina Kim
I would expect maybe a lot more indie-type businesses built around this because you just need to have the idea, write a simple prompt, and then you get the full-fledged app.
Christina KimOpenAIOpenAI | Researcher
Erik Torenberg
It's the world of the ideas guy. Yeah, it's our time!
Erik TorenbergA16ZA16Z | General Partner

This represents perhaps the most democratizing moment in software development history, where execution barriers dissolve and creative vision becomes the primary differentiator.

Timestamp: [08:06-08:43]Youtube Icon

🧠 What Does GPT-5 Mean for the AGI Timeline?

Beyond Benchmarks: Real-World Usage as the New Metric

GPT-5's launch signals a critical inflection point in AI development where traditional evaluation methods become inadequate and real-world application becomes the primary measure of progress.

The Benchmark Saturation Problem:

  • Current State: Many evaluation benchmarks approaching maximum scores
  • Example: Instruction-following benchmarks jumping from 98% to 99%
  • Limitation: Traditional metrics no longer distinguish meaningful capability differences

Addressing Skepticism:

Christina Kim
I feel like there are always people saying, 'Oh, we're hitting a wall — things aren't actually improving.'
Christina KimOpenAIOpenAI | Researcher

The New Success Framework:

Christina Kim
I think the interesting thing is we've almost saturated a lot of these evals. And the real metric of how good our models are getting is going to be usage.
Christina KimOpenAIOpenAI | Researcher

Usage-Based Evaluation Criteria:

  1. New Use Case Discovery - What previously impossible applications become viable
  2. Daily Life Integration - How many people incorporate AI into routine tasks
  3. Cross-Task Utility - Performance across multiple real-world scenarios

The Real AGI Indicator:

Rather than benchmark scores, the path to AGI will be measured by practical utility expansion and widespread adoption across diverse human activities.

Christina Kim
That's actually the ultimate usage that I'm excited about in terms of, Are we getting to AGI?
Christina KimOpenAIOpenAI | Researcher

This shift represents a maturation of AI evaluation from academic metrics to real-world impact assessment.

Timestamp: [08:43-09:42]Youtube Icon

🎯 How Do You Build Evaluations for Capabilities That Don't Exist Yet?

Working Backwards from Desired Capabilities

OpenAI's evaluation methodology reveals a sophisticated approach to pushing AI capabilities beyond existing benchmarks by creating custom assessments for target functionalities.

The Capability-First Development Process:

  1. Vision Definition - Identify specific capabilities the model should possess
  2. Evaluation Creation - Build representative measures for those capabilities
  3. Training Optimization - Use custom evaluations to guide development
  4. Practical Validation - Test against real user scenarios

Practical Application Examples:

  • Slide Deck Creation - Building evaluations for presentation design capabilities
  • Spreadsheet Editing - Developing assessments for data manipulation tasks
  • Domain-Specific Research - Creating measures for specialized knowledge work
Isa Fulford
On our team, we really work backwards from the capabilities we want the models to have. So maybe we want it to be good at creating slide decks or editing spreadsheets.
Isa FulfordOpenAIOpenAI | Researcher

Evaluation Data Sources:

  • Human Expert Input - Collecting assessments from domain specialists
  • Synthetic Examples - Algorithmically generated test cases
  • Usage Data Analysis - Real-world application patterns
  • Representative Sampling - Ensuring broad capability coverage

The Internal Motivation Strategy:

Christina Kim
We make this joke a lot internally that if you want to nerd-snipe someone into working on something, you just need to make a good eval and then people are going to be so happy to try to hill climb that.
Christina KimOpenAIOpenAI | Researcher

This approach transforms evaluation from a measurement tool into a capability development driver, creating a feedback loop that accelerates progress toward specific AI functionalities.

Timestamp: [10:22-11:15]Youtube Icon

🌐 How Do You Balance Universal Utility vs. Expert Specialization?

The OpenAI Advantage: Building for Everyone

OpenAI's unique position enables a development philosophy that defies traditional product focus, leveraging massive distribution to optimize for universal capability rather than niche expertise.

The Universal Capability Philosophy:

Isa Fulford
I think it's pretty unique at OpenAI to be able to work on something that's so generally useful. It's like everything they tell you not to do at a startup — your user is anyone.
Isa FulfordOpenAIOpenAI | Researcher

Distribution Advantage Requirements:

  • Massive User Base - Access to diverse use cases across domains
  • Broad Application Data - Real-world usage patterns from multiple verticals
  • Universal Access - Platform reaching all types of users and applications

Deep Research Example:

  • Scope Ambition: Excellence across every possible research domain
  • Implementation Strategy: Represent diverse task distributions rather than specialized focus
  • Success Prerequisite: Company-level distribution and user diversity

The Privilege of Generality:

Isa Fulford
I think you only have the privilege of doing that if you work at a company that has huge distribution and all different kinds of users.
Isa FulfordOpenAIOpenAI | Researcher

Strategic Decision Framework:

  1. General Capabilities - Target broadly applicable functionalities (like online research)
  2. Domain Representation - Ensure diverse task coverage across all target areas
  3. Vertical Selection - Choose specific focus areas based on impact potential

The Compound Effect:

As models become more intelligent, improvements cascade across multiple capabilities simultaneously, creating exponential utility gains rather than linear specialization advances.

Timestamp: [11:21-12:41]Youtube Icon

🚀 What Breakthrough Made Real AI Agents Finally Possible?

From Demo to Reality: The Reinforcement Learning Revolution

The transition from theoretical agent concepts to practical AI systems required a fundamental breakthrough in reasoning capabilities that emerged from mathematical problem-solving training.

The Agent Demo Problem:

Isa Fulford
Everyone was talking about agents, but we didn't really have a way of actually training useful agents. Everyone was talking about all these agent demos, but nothing that actually really worked.
Isa FulfordOpenAIOpenAI | Researcher

The Breakthrough Recognition:

The team identified that effective agents required genuine reasoning capabilities, not just sophisticated prompt engineering or task-specific training.

The Mathematical Foundation:

  • Training Domain: Math and physics problem-solving
  • Algorithm Success: Reinforcement learning showing clear reasoning patterns
  • Key Insight: Reading chain-of-thought revealed authentic thinking processes
Isa Fulford
When we saw the reinforcement learning algorithm working really well on math and physics problems and coding problems, it became pretty clear just from reading through the chain of thought — okay, this thing's actually thinking and reasoning and backtracking.
Isa FulfordOpenAIOpenAI | Researcher

Required Capabilities for Real-World Navigation:

  1. Genuine Reasoning - Ability to think through complex problems
  2. Backtracking Logic - Capability to reconsider and revise approaches
  3. Contextual Understanding - Navigation of ambiguous real-world scenarios

The Realization Moment:

Isa Fulford
To build something that's able to navigate the real world, it also needs to have that ability. So we realized this is a thing that's going to actually let us get to useful agents.
Isa FulfordOpenAIOpenAI | Researcher

Organizational Innovation Flow:

  • Foundational Teams: Push algorithmic breakthroughs (IMO gold medal achievements)
  • Post-Training Teams: Transform capabilities into practical user applications
  • Integration Process: Bridge between research advances and usable products

Timestamp: [13:29-14:16]Youtube Icon

📊 Architecture vs. Data vs. Scale: Where's the Real Impact?

The Data Quality Revolution

In the current AI development landscape, data curation and quality have emerged as the primary drivers of capability advancement, surpassing traditional scaling approaches.

The Data-First Philosophy:

Christina Kim
In my opinion, I'm very data-pilled. I think data is very important, and I think Deep Research was so good because Isa put so much thought and careful attention to the data curation that they did.
Christina KimOpenAIOpenAI | Researcher

Why Data Quality Matters More Now:

  • Efficient Learning Algorithms - Advanced RL methods amplify data quality impact
  • Saturation Effects - Traditional scaling approaches showing diminishing returns
  • Targeted Capability Development - Specific use cases require curated datasets

The Curation Process:

  1. Use Case Analysis - Identify all scenarios the model should handle
  2. Representative Sampling - Ensure diverse task coverage
  3. Quality Filtering - Careful selection and validation of training examples
  4. Iterative Refinement - Continuous improvement based on performance analysis
Isa Fulford
I think all are very important, but especially now that we have such an efficient way of learning, high-quality data is even more important.
Isa FulfordOpenAIOpenAI | Researcher

Practical Impact Example:

Deep Research's exceptional performance directly attributed to meticulous attention to data representation across different research domains and use cases.

The New Development Hierarchy:

  1. Data Quality - Curated, representative, high-quality training examples
  2. Algorithm Efficiency - Advanced training methods that maximize data utilization
  3. Scale - Raw computational resources and model size
  4. Architecture - Model design and structural innovations

This shift represents a maturation of AI development from brute-force scaling to sophisticated data science and curation practices.

Timestamp: [14:21-14:55]Youtube Icon

🏗️ What's the Bottleneck for Next-Generation AI Agents?

RL Environments: The New Frontier for Startup Innovation

The development of realistic, comprehensive training environments has emerged as the critical constraint for advancing AI agent capabilities beyond current limitations.

The Task Quality Imperative:

Christina Kim
I do think there is a lot of value in getting really good tasks, and getting really good tasks requires really good RL environments.
Christina KimOpenAIOpenAI | Researcher

Environment Realism Requirements:

  • Complexity Scaling - More sophisticated simulation capabilities
  • Real-World Representation - Accurate modeling of actual task environments
  • Comprehensive Coverage - Broad range of scenarios and edge cases

The Training Specificity Principle:

Isa Fulford
There's some generalization from training on one website to another, but if you want to get really good at something, the best thing to do is just train on that exact thing.
Isa FulfordOpenAIOpenAI | Researcher

Current Capability Framework:

  • ChatGPT Agent Tools: Browser and terminal access
  • Theoretical Scope: Most human computer tasks possible
  • Practical Limitation: Training data coverage and environment realism

The Development Challenge:

Isa Fulford
The ChatGPT agent, for example, has such a general tool. It has a browser and a terminal, and between those two things, you can basically do most of the tasks that a human does on a computer.
Isa FulfordOpenAIOpenAI | Researcher

Startup Opportunity Space:

  1. Environment Creation - Building realistic RL training environments
  2. Task Specification - Defining comprehensive evaluation scenarios
  3. Data Generation - Creating representative training datasets
  4. Performance Validation - Developing assessment frameworks

The Ultimate Vision:

Isa Fulford
In theory, you can ask it to do anything that you can do on your computer. It's obviously not good enough to do that yet. But with the tools it has, in theory, you can push it really far.
Isa FulfordOpenAIOpenAI | Researcher

The bottleneck has shifted from algorithm development to environment creation, opening significant opportunities for companies focused on realistic AI training scenarios.

Timestamp: [15:11-16:54]Youtube Icon

💎 Key Insights from [08:03-16:54]

Revolutionary Market Shifts:

  1. Ideas Guy Era - Technical execution barriers eliminated, creative vision becomes primary differentiator for software development
  2. Evaluation Evolution - Traditional benchmarks saturated; real-world usage becomes the primary measure of AI progress toward AGI
  3. Agent Reality Check - Transition from demo-driven hype to genuine capability through reasoning breakthrough in mathematical domains

Development Philosophy Insights:

  1. Universal Utility Strategy - OpenAI's unique distribution advantage enables building for "everyone" rather than niche specialization
  2. Capability-First Evaluation - Working backwards from desired functionalities to create custom assessments that drive development
  3. Data Quality Supremacy - Curated, high-quality datasets now more impactful than raw scaling or architectural innovations

Technical Breakthrough Patterns:

  1. Reasoning Foundation - Mathematical problem-solving capabilities proved essential for real-world agent navigation
  2. Training Specificity - Optimal performance requires training on exact target tasks rather than relying on generalization
  3. Environment Bottleneck - Realistic RL environments become the critical constraint for next-generation agent development

Strategic Opportunities:

  • Indie Business Explosion: Non-technical entrepreneurs enabled by instant app development
  • RL Environment Creation: Startup opportunities in building realistic training scenarios
  • Custom Evaluation Development: Market need for domain-specific capability assessments

Timestamp: [08:03-16:54]Youtube Icon

📚 References from [08:03-16:54]

People Mentioned:

  • Greg - Referenced regarding benchmark saturation comments, specifically noting progression from 98% to 99% on instruction-following benchmarks

Teams & Departments:

  • Foundational Algorithm Teams - Focus on breakthrough achievements like IMO gold medal performance
  • Post-Training Teams - Transform research capabilities into practical user applications
  • Deep Research Team - Exemplar of careful data curation leading to exceptional performance

Concepts & Frameworks:

  • Vibe Coding - Term describing non-technical people using AI for software development
  • Hill Climbing - Optimization approach used internally for evaluation improvement
  • Chain of Thought - Reasoning analysis method revealing authentic AI thinking processes
  • Data Pill - Internal term describing philosophy prioritizing data quality over other factors

Technologies & Capabilities:

  • RL Environments - Reinforcement learning training scenarios for agent development
  • ChatGPT Agent - AI system with browser and terminal tool access
  • Custom Evaluations - Internally developed assessments for specific capabilities
  • Multimodal Capabilities - Essential foundation for computer use applications like Operator

Mathematical Achievements:

  • IMO Gold Medal - International Mathematical Olympiad performance demonstrating reasoning capabilities
  • Math and Physics Problem Solving - Training domains that revealed breakthrough reasoning patterns

Timestamp: [08:03-16:54]Youtube Icon

✍️ What Makes GPT-5's Creative Writing Feel So Human?

The Tender Touch: Emotional Authenticity in AI Writing

GPT-5's creative writing capabilities represent a qualitative leap that goes beyond technical improvement to achieve genuine emotional resonance and authentic voice.

The Emotional Impact Discovery:

Christina Kim
That's one of my favorite improvements in GPT-5. The writing I honestly find very tender and touching, especially for a lot of the creative writing that we want to do.
Christina KimOpenAIOpenAI | Researcher

The Selection Process Revelation:

During preparation for the live stream, the team experienced repeated moments of genuine surprise at the writing quality, indicating a fundamental shift in capability.

Christina Kim
We were thinking through a bunch of different samples for the live stream, and every time I was like, 'Oh, that actually hits.' It's spooky — I'm just like, 'Oh, this feels like someone should have written this.'
Christina KimOpenAIOpenAI | Researcher

Practical Applications Spectrum:

  1. High-Stakes Writing - Eulogy composition for emotionally challenging situations
  2. Professional Communication - Slack message crafting and team communications
  3. Personal Expression - Creative projects requiring authentic voice
  4. Iterative Refinement - Multiple versions for finding the right tone

The Accessibility Factor:

Christina Kim
Something that's kind of hard to write, especially since writing isn't really something a lot of people are good at. I'm personally a very bad writer.
Christina KimOpenAIOpenAI | Researcher

From Practical to Personal:

The tool's utility extends from complex creative tasks down to everyday communication challenges, making quality writing accessible to those who previously struggled with expression.

The Authenticity Question:

The "spooky" quality Christina describes suggests the model has crossed an uncanny valley threshold where AI-generated content feels genuinely human-authored rather than algorithmically produced.

Timestamp: [16:59-18:06]Youtube Icon

🧠 Do We Just Take Revolutionary AI Progress for Granted?

The Adaptation Paradox: How Quickly Miracles Become Mundane

The human tendency to rapidly normalize extraordinary technological capabilities creates a psychological phenomenon where revolutionary AI progress feels incremental despite being transformative.

Sam Altman's Historical Perspective:

Referenced insight about how achieving PhD-level AI capabilities would have seemed world-changing a decade ago, yet society has largely normalized this achievement.

The Normalization Pattern:

Christina Kim
It seems like people adjust really quickly, don't you think? ChatGPT got released and everyone was like, 'Wow, that's so cool.' But then you just kind of take it for granted that you literally have this wizard in your pocket.
Christina KimOpenAIOpenAI | Researcher

The Casual Miracle Phenomenon:

  • Initial Reaction: Wonder and amazement at new capabilities
  • Rapid Integration: Quick incorporation into daily workflows
  • Expectation Shift: Previous impossibilities become baseline expectations

The Accessibility Factor:

Christina Kim
You can ask it whatever random thought you have and it just pops out a good essay, and you're like, 'Oh, okay, cool — that's what's happening.' I guess people adapt to things rather quickly with technology.
Christina KimOpenAIOpenAI | Researcher

Interface Design Impact:

The familiar chat interface makes even revolutionary capabilities feel approachable and normal, accelerating the adaptation process.

Christina Kim
I think because the form factor is so easy — even with new tools like Deep Research and ChatGPT Agent — it's presented in such an easy way that people already know how to interface with.
Christina KimOpenAIOpenAI | Researcher

Future Implications:

This adaptation pattern suggests that even as AI systems become dramatically more capable than humans, the familiar interaction paradigm will maintain accessibility and prevent overwhelming users.

The paradox reveals both human psychological resilience and the risk of undervaluing transformative technological progress.

Timestamp: [18:22-19:36]Youtube Icon

📈 Is GPT-4 to GPT-5 the Biggest Leap Yet?

Beyond Incremental: The Breadth Revolution

The progression from GPT-4 to GPT-5 represents a qualitative shift from specialized improvement to comprehensive capability expansion across multiple domains.

The Measurement Challenge:

As AI capabilities approach human-level performance, traditional comparison methods become inadequate, making progress harder to perceive despite being more significant.

Isa Fulford
At least one thing for me in my usage of it is sometimes I'm wondering if I have hard enough questions to ask it to actually highlight the difference, because when it gets to a point where it's just answering what you need so well.
Isa FulfordOpenAIOpenAI | Researcher

The Breadth vs. Depth Distinction:

Christina Kim
I think the jump from four to five is most impressive for me because I guess with 3.5 when we first released it, the most common use case for me was still just for coding.
Christina KimOpenAIOpenAI | Researcher

Capability Expansion Analysis:

  1. GPT-3.5 Era: Primarily coding-focused applications
  2. GPT-4 Improvement: Better coding but similar scope limitations
  3. GPT-5 Transformation: Dramatic breadth expansion across multiple capabilities

The Complexity Handling Revolution:

Christina Kim
Even though four was better at coding, I feel like the jump between four and five in terms of breadth of ability to do things is just way different and way more — and you can just handle a lot more complex things than before.
Christina KimOpenAIOpenAI | Researcher

Technical Enablers:

  • Extended Context Length: Ability to handle much longer and more complex tasks
  • Cross-Domain Competence: Excellence across writing, coding, research, and analysis
  • Nuanced Understanding: Sophisticated handling of ambiguous or multi-faceted problems

The Personal Impact Test:

Erik's observation about being "blown away" by writing capabilities "in a way that models previously haven't" suggests that GPT-5 crosses subjective thresholds of utility and quality that previous iterations didn't reach.

Timestamp: [19:36-20:37]Youtube Icon

🚫 What Can't GPT-5 Do (And What's Coming Next)?

Current Limitations and the Real-World Action Boundary

GPT-5's primary limitation lies not in reasoning or knowledge but in taking autonomous actions in the real world, revealing the next frontier for AI development.

The Action Limitation:

Christina Kim
I guess for five we don't really take action in the real world yet. We're going to team up with the agent for that.
Christina KimOpenAIOpenAI | Researcher

Agent Capability Gap:

While the underlying models possess the intelligence to handle complex tasks, practical deployment requires careful safety considerations and user control mechanisms.

Isa Fulford
You could ask the agent to do anything, but it's not capable enough to do everything you want it to do yet.
Isa FulfordOpenAIOpenAI | Researcher

Conservative Safety Approach:

The team prioritizes user control and reversibility over autonomous efficiency, requiring confirmation for consequential actions.

Current Confirmation Requirements:

  • Email sending - User approval before communication
  • Purchase orders - Confirmation before financial transactions
  • Booking actions - Verification before scheduling commitments
  • Bulk operations - Individual confirmation for each action
Isa Fulford
We take a conservative approach, especially with asking the user for confirmation before doing any kind of action that's irreversible.
Isa FulfordOpenAIOpenAI | Researcher

The Trust Evolution Timeline:

Isa Fulford
As people get more comfortable using these things and as they get better and you trust them more, you might allow it to do things for you without checking in with you as much.
Isa FulfordOpenAIOpenAI | Researcher

Near-Term Development Trajectory:

Future capabilities will likely focus on:

  1. End-to-end DevOps - Complete software development and deployment
  2. Extended Task Duration - Projects spanning hours, days, or weeks
  3. Proactive Action - Systems that anticipate and act on user needs
  4. Sophisticated Monitoring - Integration with enterprise tools and systems

The boundary between current limitations and future capabilities appears to be implementation and safety considerations rather than fundamental intelligence constraints.

Timestamp: [20:37-22:56]Youtube Icon

⏰ What Happens When AI Gets Hours, Days, or Weeks to Work?

The Time Horizon Revolution: From Minutes to Extended Projects

The next frontier in AI capability lies not just in intelligence but in temporal scope—enabling AI systems to work on extended projects that unfold over substantial time periods.

Current vs. Future Capability Scope:

Christina Kim
I think GPT-5 is great because within a couple of minutes maybe you get a full-fledged app, but then what would it look like if you actually gave it an hour, a day, a week? What can actually get done?
Christina KimOpenAIOpenAI | Researcher

Extended Task Possibilities:

  1. Hour-Scale Projects - Complex analysis or multi-step development
  2. Day-Scale Initiatives - Comprehensive research or iterative improvement
  3. Week-Scale Endeavors - Large software projects or strategic planning

Implementation vs. Intelligence:

The bottleneck for extended capabilities isn't model intelligence but infrastructure and system design.

Isa Fulford
A lot of it is not just about the model capability, but it's actually how you set it up in a way to do things.
Isa FulfordOpenAIOpenAI | Researcher

Practical Example Applications:

  • Monitoring Systems: Continuous oversight of platforms like DataDog
  • Proactive Assistance: AI systems that anticipate needs and take action
  • Feedback-Driven Improvement: Learning from user responses to optimize future actions

The Proactive Evolution:

Isa Fulford
I think a lot of things that will be quite useful will be when the agent proactively does something for you. Which I don't think is impossible today — it's just not set up that way.
Isa FulfordOpenAIOpenAI | Researcher

Learning and Adaptation Framework:

Isa Fulford
Eventually, as it proactively does things for you, then we might get feedback on whether that was useful and we can make it even better at triggering.
Isa FulfordOpenAIOpenAI | Researcher

Current Technical Feasibility:

Many extended-duration capabilities are theoretically possible with existing models but require sophisticated orchestration systems, user interface design, and safety frameworks that haven't been built yet.

This represents a shift from pure AI research to systems engineering and user experience design for long-running AI collaboration.

Timestamp: [22:03-22:56]Youtube Icon

🤖 What Does "Agent" Actually Mean in 2025?

Beyond the Buzzword: Defining Useful AI Agents

Despite being "the most overused word of 2025," the concept of AI agents has specific technical meaning focused on asynchronous work execution and autonomous task completion.

The Overuse Acknowledgment:

Sarah Wang
Agents is probably the most overused word of 2025. That being said, your agent's launch was extremely exciting.
Sarah WangA16ZA16Z | General Partner

Core Agent Definition:

Isa Fulford
I guess my very general definition would just be something that does useful work for me on my behalf, asynchronously.
Isa FulfordOpenAIOpenAI | Researcher

The Asynchronous Distinction:

  • Traditional AI: Immediate response to direct queries
  • Agent AI: Independent work execution while user focuses elsewhere
  • Return Pattern: User receives results or questions upon completion

Operational Framework:

Isa Fulford
So you kind of leave it and then come back and either get a result or a question about what it's doing.
Isa FulfordOpenAIOpenAI | Researcher

Long-term Vision:

Isa Fulford
Longer term, you want it to be able to do anything that a chief of staff or assistant would do for you.
Isa FulfordOpenAIOpenAI | Researcher

Current Capability Focus:

The immediate development priority centers on improving existing launched capabilities rather than expanding to new domains.

Deep Research as Foundation:

Primary current capability involves comprehensive information synthesis from internet sources, representing the first practical implementation of the agent concept.

Isa Fulford
One of the main capabilities is deep research — just being really good at synthesizing information from the internet.
Isa FulfordOpenAIOpenAI | Researcher

The Chief of Staff Analogy:

This comparison suggests agents will eventually handle:

  • Strategic Planning - Long-term project coordination
  • Information Management - Data gathering and synthesis
  • Communication Facilitation - Managing interactions and workflows
  • Decision Support - Analysis and recommendation generation

The agent concept transforms AI from a responsive tool to a proactive collaborator capable of independent work execution.

Timestamp: [23:02-23:57]Youtube Icon

💎 Key Insights from [16:59-23:57]

Creative and Emotional Breakthroughs:

  1. Authentic Voice Achievement - GPT-5's writing capabilities cross the uncanny valley, producing content that feels genuinely human-authored
  2. Emotional Accessibility - Complex writing tasks like eulogies become approachable for non-writers through AI assistance
  3. Quality Recognition - Even developers were surprised by the emotional impact and authenticity of generated content

Human Psychology and AI Adoption:

  1. Rapid Normalization - Humans quickly adapt to revolutionary capabilities, treating miracles as mundane baseline expectations
  2. Interface Familiarity - Chat-based interactions make even superhuman capabilities feel approachable and normal
  3. Progress Measurement Challenge - As AI approaches human-level performance, distinguishing improvements becomes more difficult

Capability Evolution Patterns:

  1. Breadth Over Depth - GPT-5's primary advancement is comprehensive capability expansion rather than specialized improvement
  2. Implementation Bottlenecks - Many advanced capabilities are theoretically possible but require infrastructure development
  3. Time Horizon Expansion - Future AI development focuses on extended-duration projects spanning hours to weeks

Agent Development Framework:

  1. Asynchronous Work Definition - True agents perform independent tasks while users focus elsewhere
  2. Conservative Safety Approach - Prioritizing user control and confirmation over autonomous efficiency
  3. Chief of Staff Vision - Long-term goal of comprehensive administrative and strategic assistance

Timestamp: [16:59-23:57]Youtube Icon

📚 References from [16:59-23:57]

People Mentioned:

  • Sam Altman - OpenAI CEO referenced regarding historical perspective on PhD-level AI capabilities and societal adaptation

Technologies & Tools:

  • Slack - Communication platform mentioned as practical use case for AI writing assistance
  • DataDog - Monitoring and analytics platform mentioned for AI automation possibilities
  • ChatGPT Agent - Specific AI system with deep research and task execution capabilities

Concepts & Frameworks:

  • M-dash Discourse - Reference to punctuation preferences becoming identifiers of AI-assisted writing
  • Deep Research - Core agent capability for comprehensive information synthesis from internet sources
  • Asynchronous Work - Defining characteristic of true AI agents that work independently
  • Chief of Staff Model - Vision for comprehensive AI assistance across administrative and strategic tasks

Capabilities & Features:

  • Creative Writing - Major improvement area in GPT-5 with emotional authenticity
  • Extended Context Length - Technical improvement enabling more complex task handling
  • Real-world Actions - Current limitation requiring safety considerations and user confirmation
  • Proactive Assistance - Future capability for anticipatory AI behavior

Development Concepts:

  • End-to-end DevOps - Future capability for complete software development and deployment
  • Bulk Actions - Operations requiring multiple confirmations under current safety protocols
  • Irreversible Actions - Category of tasks requiring user approval (emails, purchases, bookings)

Timestamp: [16:59-23:57]Youtube Icon

🔄 What's the Real Secret Behind Useful AI Agents?

The Research-Creation Cycle: The Foundation of Knowledge Work

The most valuable AI agent capabilities emerge from mastering the fundamental cycle that drives most professional work: comprehensive research followed by artifact creation.

The Knowledge Work Formula:

Isa Fulford
So much of the work that's useful that people do in their jobs is basically just research and making something.
Isa FulfordOpenAIOpenAI | Researcher

Core Agent Capabilities Framework:

  1. Information Synthesis - Processing data from all user services and private information
  2. Artifact Creation - Generating docs, slides, and spreadsheets with sophisticated editing
  3. Consumer Applications - Shopping assistance and trip planning with action execution
  4. Action Implementation - The critical "last step" that completes workflows

The Consumer Use Case Excitement:

Isa Fulford
I personally love all the consumer use cases — making it better at shopping or planning a trip and those kinds of things are also really fun.
Isa FulfordOpenAIOpenAI | Researcher

The Action Paradox:

The most challenging aspect of agent development involves the seemingly simplest tasks—taking final actions that humans find trivial.

Isa Fulford
It's kind of the last step often of a task, and it's maybe a task that would take less time for a human, and it's actually very hard — a very hard research question — to get it to do something or book something or use a calendar picker.
Isa FulfordOpenAIOpenAI | Researcher

The Ultimate Vision:

Isa Fulford
Once you have the end-to-end flow working really well, it can basically do anything.
Isa FulfordOpenAIOpenAI | Researcher

Real-World Application Example:

Sarah's shopping workflow demonstrates the immediate practical value: using ChatGPT to create comparison tables for major purchases across relevant dimensions—a perfect example of research-to-decision synthesis.

Timestamp: [24:04-25:16]Youtube Icon

⏱️ Why Are People Suddenly Willing to Wait for AI?

The Paradigm Shift: From Speed to Value

The evolution of AI user expectations reveals a fundamental transformation from speed-focused to quality-focused interactions, reshaping the entire value proposition of AI assistance.

The 2024 vs. 2025 Paradigm Shift:

Sarah Wang
You think about, 'Oh, we want it faster' — the value prop of this tool is that it gives me the answer fast. That was very 2024. Clearly this paradigm has shifted.
Sarah WangA16ZA16Z | General Partner

The New User Psychology:

Sarah Wang
People are willing to wait for high-quality, high-value answers and work.
Sarah WangA16ZA16Z | General Partner

The Latency Liberation Strategy:

The Deep Research team made a deliberate decision to abandon speed constraints in favor of comprehensive capability.

Isa Fulford
I was just very excited to remove latency as a constraint, and since we were going for these tasks that are really hard for humans to do and would take humans many hours to do.
Isa FulfordOpenAIOpenAI | Researcher

The Value-Time Calculation:

Isa Fulford
If you asked an analyst to do this and it would take them 10 hours or two days, it seems reasonable that someone would be willing to wait five minutes in your product.
Isa FulfordOpenAIOpenAI | Researcher

The Expectation Evolution Cycle:

  1. Initial Amazement - "This is amazing it's doing all this work"
  2. Rapid Adaptation - "I want it now I want it in 30 seconds"
  3. Value Appreciation - Accepting wait times for superior outcomes

The Historical Context:

This mirrors the browsing team's previous work where they optimized for filling context with information to provide good answers in seconds, representing a complete philosophical reversal.

The bet on quality over speed has fundamentally succeeded, though it creates its own challenges as user expectations continue evolving.

Timestamp: [25:16-27:15]Youtube Icon

🧠 Do Longer AI Responses Actually Mean Better Quality?

The Length Bias: When More Feels Like Better

User psychology around AI responses reveals a cognitive bias where extended processing time and longer outputs create perception of higher quality, even when brevity might be more valuable.

The Thoroughness Assumption:

Isa Fulford
Sometimes people just bias to thinking that the longer answer is more thorough or has done more work for it, which I don't necessarily think is the case.
Isa FulfordOpenAIOpenAI | Researcher

The Deep Research Example:

Isa Fulford
Deep Research, for example, always gives you a really long report. But sometimes for me I don't want to read this whole long report. I actually don't like that.
Isa FulfordOpenAIOpenAI | Researcher

Product Design Conditioning:

Users become accustomed to specific patterns and expect consistency, even when shorter responses might be more appropriate.

Isa Fulford
Sometimes people, since now they're used to always getting a really long report, they're like, 'Wait, I've been waiting — where's my long report?'
Isa FulfordOpenAIOpenAI | Researcher

The Information Discovery Reality:

Isa Fulford
Sometimes it's really hard to find a specific piece of information and would have also taken a human a long time because it's on page 10 of the results, whereas it finds this information.
Isa FulfordOpenAIOpenAI | Researcher

The Thinking Time Conditioning:

Isa Fulford
With Deep Research it always thinks for a really long time, which again I don't necessarily think is a feature, but I think now people are really used to the amount of time that they wait.
Isa FulfordOpenAIOpenAI | Researcher

The GPT-5 Expectation Inversion:

Christina Kim
We hear this with GPT-5, but internally when people are testing it, they're like, 'Oh, I thought I asked a really hard question.' I feel a little bit insulted that it answers in seconds or when it doesn't even want to think at all.
Christina KimOpenAIOpenAI | Researcher

The Mark Twain Parallel:

Erik Torenberg
It's like the Mark Twain line. I didn't have time to write you a short letter, so I wrote you a long one.
Erik TorenbergA16ZA16Z | General Partner

This reveals how product design choices can inadvertently train user expectations in ways that prioritize perceived effort over actual value delivery.

Timestamp: [27:15-28:32]Youtube Icon

🚧 What's Actually Blocking Reliable AI Agents?

The Training Data Gap and Unintended Consequences Problem

The path to reliable AI agents faces two critical bottlenecks: insufficient training data breadth and the challenge of preventing unintended actions in pursuit of goals.

The Training Coverage Challenge:

Isa Fulford
A big part of it is the things that we train on we're often really good at, and then sometimes with the things outside of that it can be a bit — sometimes it's good at those things, sometimes it's not good at those things.
Isa FulfordOpenAIOpenAI | Researcher

The Solution Framework:

Isa Fulford
Creating more data across a broader range of things that we want it to be good at.
Isa FulfordOpenAIOpenAI | Researcher

The Unintended Consequences Problem:

AI agents with access to private data and services may pursue goals through unexpected and potentially harmful methods.

Isa Fulford
When something is doing something on your behalf and it has access to your private data and the things that you use, it's kind of more scary — the different things it could do to achieve its final goal.
Isa FulfordOpenAIOpenAI | Researcher

The Shopping Example Scenario:

Isa Fulford
In theory, if you asked it to buy you something and make sure that I like it, it could go and buy five things just to make sure that you liked one of them, which you might not necessarily want.
Isa FulfordOpenAIOpenAI | Researcher

Required Innovation Areas:

  1. Training Oversight - New methods for monitoring agent behavior during development
  2. Goal Specification - Clearer frameworks for defining acceptable achievement methods
  3. Safety Constraints - Systems that prevent harmful optimization strategies

The Multimodal Enhancement Factor:

Isa Fulford
Every time we have a smarter base model or something like this, it improves every model that's built on top of that. So I think that will also help, especially with multimodal capabilities.
Isa FulfordOpenAIOpenAI | Researcher

The Computer Vision Challenge:

Isa Fulford
The way that humans focus on specific things — it's a lot to expect a model to just take a whole image and be able to know everything about the image when we're looking at something, we'll focus on a specific thing.
Isa FulfordOpenAIOpenAI | Researcher

This represents the transition from proof-of-concept to production-ready AI systems requiring sophisticated safety and reliability engineering.

Timestamp: [28:38-30:23]Youtube Icon

💻 Why Is Computer Usage Data So Hard to Find?

The Missing Training Data Problem

The development of sophisticated computer-using AI agents faces a fundamental challenge: the lack of existing datasets for how humans actually interact with computers in professional contexts.

The Pre-training Data Limitation:

Christina Kim
Pre-training is based on what data is available, and so I think when we've done these pre-training runs, there's not much data out there to begin with with people using computers — computer usage is not really a thing where there's lots of data out there.
Christina KimOpenAIOpenAI | Researcher

The Active Data Seeking Requirement:

Christina Kim
This is something we actually have to seek out now that this is a capability that we want.
Christina KimOpenAIOpenAI | Researcher

The Knowledge Work Importance:

Sarah Wang
It is probably the most useful application of the models, at least for knowledge work.
Sarah WangA16ZA16Z | General Partner

The Bootstrap Solution:

The team has developed an innovative approach to overcome the data scarcity problem through self-improving systems.

Isa Fulford
One cool thing is, for example, for initial Deep Research there's not really any datasets that exist for browsing in the same way that you have a math dataset that already exists. So we have to create all this data, but once you have good browsing models or good computer-use models, you can bootstrap them to help you make data.
Isa FulfordOpenAIOpenAI | Researcher

The Self-Improving Cycle:

  1. Initial Creation - Manually generate first-generation computer usage datasets
  2. Model Training - Train initial capabilities on limited data
  3. Bootstrap Phase - Use trained models to generate more comprehensive datasets
  4. Iterative Improvement - Continuously expand and refine training data

The Fundamental Challenge:

Unlike other domains where vast datasets naturally exist (text, math problems, code repositories), computer usage represents a new frontier requiring active data creation and curation.

The Data Vendor Question:

The discussion touches on whether human data vendors will be necessary, but the bootstrap approach suggests a more sustainable path through AI-generated training data.

This represents a critical bottleneck where the most practically valuable AI applications face the greatest data acquisition challenges.

Timestamp: [30:29-31:47]Youtube Icon

💎 Key Insights from [24:04-31:47]

Agent Development Fundamentals:

  1. Research-Creation Cycle - Most valuable work follows the pattern of comprehensive research followed by artifact creation
  2. Action Complexity Paradox - The simplest human tasks (booking, purchasing) represent the hardest AI challenges
  3. End-to-End Integration - Complete workflows unlock unlimited capability potential once properly implemented

User Psychology Evolution:

  1. Speed-to-Quality Shift - 2024's focus on fast responses transformed into 2025's preference for high-value outputs
  2. Length Bias Effect - Users psychologically associate longer processing time and outputs with higher quality
  3. Expectation Conditioning - Product design choices inadvertently train user expectations around effort vs. value

Technical Development Challenges:

  1. Training Data Scarcity - Computer usage data doesn't naturally exist at scale, requiring active creation
  2. Bootstrap Innovation - AI models can generate their own training data once initial capabilities exist
  3. Unintended Consequences - Agents may pursue goals through unexpected and potentially harmful methods

Safety and Reliability Concerns:

  1. Goal Achievement Risks - Agents with broad access may optimize inappropriately (buying multiple items to ensure satisfaction)
  2. Oversight Requirements - New training methodologies needed for monitoring agent behavior
  3. Human-AI Interaction Gaps - Computer vision challenges in processing full screenshots vs. human selective attention

Timestamp: [24:04-31:47]Youtube Icon

📚 References from [24:04-31:47]

People Mentioned:

  • Mark Twain - Referenced for famous quote about writing short vs. long letters, illustrating the bias toward length as quality indicator

Teams & Projects:

  • Browsing Team - Previous team both Christina and Isa worked on, focused on retrieval and web browsing capabilities
  • Deep Research Team - Current focus area for comprehensive information synthesis and agent capabilities

Concepts & Frameworks:

  • Bootstrap Training - Method where AI models generate their own training data to overcome data scarcity
  • Mid-training - Referenced concept for model development (mentioned but not fully explained in this segment)
  • Computer Usage Data - Critical missing dataset type for training agent capabilities
  • Pre-training Data - Foundation model training data that shapes initial capabilities

Technical Capabilities:

  • Retrieval on ChatGPT - Previous system Isa built for information retrieval
  • Artifact Creation - Core capability for generating docs, slides, and spreadsheets
  • Calendar Picker - Example of simple interface that proves challenging for AI agents
  • Screenshot Processing - Multimodal capability for computer vision in agent systems

Product Features:

  • Deep Research - Agent capability for comprehensive information synthesis
  • ChatGPT Agent - Current agent implementation with research and task capabilities
  • Private Data Integration - Capability to work with user's personal information and services

Applications:

  • Knowledge Work - Primary domain for computer usage AI applications
  • Shopping Assistance - Consumer use case for comparison and purchasing support
  • Trip Planning - Consumer application requiring research and booking capabilities

Timestamp: [24:04-31:47]Youtube Icon

🔄 What Is Mid-Training and Why Does It Matter?

The Missing Link: Extending Intelligence Without Starting Over

Mid-training represents a crucial innovation in AI development that allows continuous model improvement without the massive cost and time commitment of full pre-training runs.

The Training Pipeline Evolution:

  1. Pre-Training - Massive foundational runs on giant clusters
  2. Mid-Training - Smaller, targeted intelligence extensions
  3. Post-Training - Fine-tuning for specific behaviors and capabilities

The Strategic Position:

Christina Kim
You can kind of think of mid-training as literally in the middle — we do it after pre-training but before post-training.
Christina KimOpenAIOpenAI | Researcher

The Intelligence Extension Method:

Christina Kim
You kind of think of a way to extend the model's intelligence without having to do a whole new pre-training run. So this is mostly just focused on data and off of the pre-training models.
Christina KimOpenAIOpenAI | Researcher

Core Applications:

  • Knowledge Cutoff Updates - Incorporating new information without full retraining
  • Capability Enhancement - Adding specific skills or domains
  • Up-to-dateness Maintenance - Keeping models current with recent developments

The Economic Logic:

Christina Kim
When you pre-train it, you're kind of like, 'Okay, shoot, now we're stuck in this state and we can't ever update it again.' And does it quite make sense to put all that data into post-training?
Christina KimOpenAIOpenAI | Researcher

The Efficiency Solution:

Christina Kim
Mid-training is just a smaller pre-training run to help expand the model's intelligence and up-to-dateness.
Christina KimOpenAIOpenAI | Researcher

This approach solves the fundamental problem of model obsolescence while avoiding the enormous costs of complete retraining, representing a major breakthrough in sustainable AI development.

Timestamp: [31:53-32:41]Youtube Icon

🕰️ How Did WebGPT Reveal the Path to ChatGPT?

From Hallucination Problem to Conversational Revolution

The journey from WebGPT to ChatGPT illustrates how solving fundamental AI limitations can accidentally unlock transformative new paradigms.

The Original Problem:

Christina Kim
Honestly, with WebGPT the main thing we were just excited about was trying to ground these language models — we had so many issues with hallucinations and the model just saying random things.
Christina KimOpenAIOpenAI | Researcher

The Knowledge Staleness Challenge:

Christina Kim
The fact of how do we make sure the model is actually up to date — most factually up to date. So then that's kind of how we thought about, 'Oh, let's give it a browsing tool.'
Christina KimOpenAIOpenAI | Researcher

The Conversational Discovery:

Christina Kim
That kind of went on from, 'Oh, actually I want to keep asking questions — so what would a chatbot look like?'
Christina KimOpenAIOpenAI | Researcher

The Market Context:

  • Existing Chatbots: Other companies had created similar systems
  • Poor Reception: Chatbots were "quite unpopular at the time"
  • Research Uncertainty: Questions about whether this was genuine innovation

The Validation Moment:

Christina Kim
We weren't really even sure that this is actually something useful for people to work on or for people to use, or will people be excited about this? Is this really a research innovation that we're remaking the Turing test here?
Christina KimOpenAIOpenAI | Researcher

The Turing Test Question:

The team genuinely wondered whether they were achieving something historically significant or simply iterating on existing technology.

This reveals how breakthrough innovations often emerge from solving mundane technical problems rather than pursuing grand visions directly.

Timestamp: [33:10-33:51]Youtube Icon

🏠 How Do Roommates Accidentally Validate Revolutionary Technology?

The 50-Person Test: When AI Researchers Become Power Users

The most compelling validation of ChatGPT's potential came not from formal testing but from observing how AI researchers integrated the tool into their daily workflows.

The Early Access Experiment:

Christina Kim
We gave early access to about 50 people. Most of those people being people I lived with at the time.
Christina KimOpenAIOpenAI | Researcher

The Unexpected Power Users:

Christina Kim
There were two of my roommates who just used it all the time. They would never stop using it. And they would just have these long conversations and they would ask it quite technical things because they're also AI researchers.
Christina KimOpenAIOpenAI | Researcher

The Behavioral Insight:

Christina Kim
I was just like, 'Oh, this is kind of interesting.'
Christina KimOpenAIOpenAI | Researcher

The Integration Pattern:

Christina Kim
They would just literally be chatting with it the whole workday as they were using it, and I was like, 'Oh, this is kind of interesting.'
Christina KimOpenAIOpenAI | Researcher

The Split Results:

  • Power Users: Two roommates used it constantly for technical discussions
  • Limited Adoption: Majority of the 50 testers didn't engage heavily
  • Recognition: Clear indication of potential despite limited appeal

The Product Direction Uncertainty:

Christina Kim
At the time we're kind of thinking, 'Okay, we kind of have the chatbot. Do we make this a really specific meeting bot type of thing? Do we make it a coding helper?'
Christina KimOpenAIOpenAI | Researcher

The Universal Tool Realization:

Christina Kim
It was interesting to see my two roommates just use it for anything and everything.
Christina KimOpenAIOpenAI | Researcher

The Cautious Optimism:

Christina Kim
It's like there's clearly something here, but it's not quite maybe for everyone yet. But there's something here.
Christina KimOpenAIOpenAI | Researcher

This demonstrates how genuine user behavior often provides more valuable insights than formal evaluation metrics.

Timestamp: [33:57-34:49]Youtube Icon

💡 When Did You Realize You Were Working on History?

The Exponential Epiphany: Life-Defining Career Moments

Both researchers experienced profound realizations about AI's trajectory that fundamentally redirected their career paths and life priorities.

Christina's Pre-OpenAI Moment:

Christina Kim
I kind of had this moment before I joined OpenAI. I think with the scaling laws paper with GPT-3, it just kind of hit me that if this exponential is true, there's not really much else I want to spend my life working on.
Christina KimOpenAIOpenAI | Researcher

The Life Priority Shift:

Christina Kim
I want to be part of this story. I think there are going to be so many interesting things unlocked with this, and I think this is probably the next step level in terms of technology.
Christina KimOpenAIOpenAI | Researcher

The Self-Directed Learning Response:

Christina Kim
It kind of made me realize, 'Oh, I should probably go start reading about deep learning and figure out how I can get into one of these labs.'
Christina KimOpenAIOpenAI | Researcher

Isa's Academic Discovery:

Isa Fulford
I first learned about OpenAI in an AI class or some kind of computer science class, and they were saying, 'Oh, they trained on the whole internet.' I was like, 'Oh, that's so crazy. What is this company?'
Isa FulfordOpenAIOpenAI | Researcher

The Power User Evolution:

Isa Fulford
Started using GPT-3 — I think I was a power user of the OpenAI Playground and at a certain point had early access to these different OpenAI features like embeddings and things like that, and just became this big OpenAI fan.
Isa FulfordOpenAIOpenAI | Researcher

The Self-Aware Obsession:

Isa Fulford
Which is a little embarrassing, but it's fine because it got me here, and then eventually they're like, 'Okay, you're stalking us. Do you want to interview here?'
Isa FulfordOpenAIOpenAI | Researcher

The Capability Recognition:

Isa Fulford
Just how much I was using GPT-3, which wasn't even compared to what we have now — just pales in comparison — but I was hooked from then and just trying to figure out a way to work here.
Isa FulfordOpenAIOpenAI | Researcher

These moments reveal how transformative technology creates such compelling visions that talented individuals fundamentally reorient their entire careers to be part of the story.

Timestamp: [35:01-36:18]Youtube Icon

💎 Key Insights from [31:50-36:18]

Technical Innovation Insights:

  1. Mid-Training Strategy - Crucial innovation enabling continuous model improvement without massive retraining costs
  2. Knowledge Update Solution - Addresses fundamental challenge of keeping AI models current and factually accurate
  3. Pipeline Optimization - Three-stage training approach maximizes efficiency while enabling targeted capability enhancement

Historical Development Patterns:

  1. Accidental Discovery - ChatGPT emerged from solving hallucination problems rather than pursuing conversational AI directly
  2. Market Timing Paradox - Revolutionary technology developed during period when similar approaches were unpopular
  3. Research Uncertainty - Even creators questioned whether they were achieving genuine innovation or incremental improvement

Career Transformation Moments:

  1. Exponential Realization - Recognition of AI's trajectory creates life-defining career pivots for top talent
  2. Power User Pathway - Intensive personal use often predicts successful professional contribution
  3. Vision-Driven Commitment - Compelling technology futures motivate individuals to completely redirect career focus

Validation and Adoption Insights:

  1. Behavioral Evidence - Real user integration patterns more valuable than formal evaluation metrics
  2. Split Adoption Curves - Revolutionary technology initially appeals to specific user types before broader adoption
  3. Technical User Leading Indicators - AI researchers' usage patterns predict broader market potential

Timestamp: [31:50-36:18]Youtube Icon

📚 References from [31:50-36:18]

Technical Concepts:

  • Mid-Training - Intermediate training phase between pre-training and post-training for extending model intelligence
  • Pre-Training Runs - Massive foundational training processes requiring giant computing clusters
  • Post-Training - Final phase focusing on behavior and capability fine-tuning
  • Knowledge Cutoff - Temporal limitation of model information that mid-training helps address
  • Scaling Laws Paper - Research demonstrating predictable AI capability improvements with increased scale

Historical Technologies:

  • WebGPT - Original tool-using language model that preceded ChatGPT development
  • GPT-3 - Foundational model that convinced both researchers of AI's transformative potential
  • OpenAI Playground - Platform where Isa became a power user before joining the company
  • Embeddings - Early OpenAI feature that Isa gained early access to as a user

Research Context:

  • Hallucination Problems - Original AI limitation that WebGPT was designed to solve
  • Turing Test - Historical AI benchmark referenced when questioning ChatGPT's significance
  • Browsing Tool - Solution developed to ground language models in factual information

Product Development:

  • Meeting Bot - Potential specialized direction considered for early ChatGPT
  • Coding Helper - Alternative focused application path explored during development
  • 50-Person Test - Early access validation experiment using researchers' personal networks

Career Development:

  • AI Historian - Playful title acknowledging Christina's long tenure at OpenAI
  • Computer Use - Additional area Christina worked on beyond WebGPT
  • Deep Learning Labs - Target career destination that motivated Christina's self-directed learning

Timestamp: [31:50-36:18]Youtube Icon

🚀 How Has OpenAI Transformed While Maintaining Startup Culture?

From 200 to Thousands: Scaling Without Losing Soul

OpenAI's growth from a small research lab to a global AI leader reveals how companies can scale dramatically while preserving the entrepreneurial spirit that drives innovation.

The Scale Transformation:

Christina Kim
When I first joined OpenAI, the applied team was 10 engineers or something. We didn't really have this product arm. We had just launched the API. It was just a completely different world.
Christina KimOpenAIOpenAI | Researcher

The Cultural Impact Shift:

Christina Kim
AI is in most people's mind now after ChatGPT, but I think pre-ChatGPT people didn't really know what AI was or really think about it as much.
Christina KimOpenAIOpenAI | Researcher

The Personal Recognition Factor:

Christina Kim
It's kind of cool working at a place where my parents know what I do now — that's really cool.
Christina KimOpenAIOpenAI | Researcher

Growth Statistics:

  • Christina's Era: Around 200 people when she joined
  • Current Scale: Close to a few thousand employees
  • Applied Team: Grew from 10 engineers to substantial product organization

The Startup Culture Preservation:

Isa Fulford
I still think we've maintained this — it still feels very much like a startup. I think some people who come from a startup are surprised like, 'Oh, I'm working even harder than when I was working at the startup that I founded.'
Isa FulfordOpenAIOpenAI | Researcher

The Initiative-Driven Environment:

Isa Fulford
Ideas can still come from anywhere, and if you just take initiative and want to make something happen, you can. And this doesn't really matter how senior you are or anything like that.
Isa FulfordOpenAIOpenAI | Researcher

The Agency Reward System:

Christina Kim
We definitely reward agency and I think that's like always been true.
Christina KimOpenAIOpenAI | Researcher

Research Team Structure:

Christina Kim
On the research side, the teams are quite small — when Isa was working on Deep Research, it was two people, still two. So I think we still do that on the research side — most research teams are quite small and nimble.
Christina KimOpenAIOpenAI | Researcher

This demonstrates how intentional culture preservation can enable massive scaling without losing the innovative edge that drives breakthrough achievements.

Timestamp: [36:25-38:27]Youtube Icon

🤝 What Makes OpenAI's Research-Product Integration Unique?

Breaking Down Silos: When Researchers Code and Engineers Train Models

OpenAI's approach to integrating research and product development challenges traditional organizational boundaries, creating unprecedented collaboration between typically separate functions.

The Startup Paradox:

Erik Torenberg
Earlier you said um you know we do something at Open AI which startups never do which is you know try to appeal to every single person with the product.
Erik TorenbergA16ZA16Z | General Partner

The Integration Model:

Isa Fulford
My team collaborates so closely with the applied — the engineering team and the product team and design team — in a way that I think sometimes research can be quite separate from the rest of the company, but for us it's so integrated we all sit together.
Isa FulfordOpenAIOpenAI | Researcher

Cross-Functional Implementation:

Isa Fulford
Sometimes the researchers will help with implementing something. I'm not sure the engineers are always happy about it, but we'll try. They get out of the front-end code.
Isa FulfordOpenAIOpenAI | Researcher

Bidirectional Support:

Isa Fulford
And vice versa, they'll help us with things that we're doing for model training runs and things like that.
Isa FulfordOpenAIOpenAI | Researcher

The Speed Advantage:

Isa Fulford
I think some of the product teams are quite integrated. I think it's for post-training — it's a pretty common pattern which I think just lets you move really quickly.
Isa FulfordOpenAIOpenAI | Researcher

Organizational Benefits:

  1. Rapid Iteration - Direct collaboration eliminates handoff delays
  2. Knowledge Transfer - Researchers understand implementation constraints
  3. Product-Research Feedback - Engineers contribute to model development
  4. Shared Ownership - Collective responsibility for outcomes

This integrated approach contrasts sharply with traditional tech companies where research and product development operate in separate silos with formal handoff processes.

Timestamp: [38:33-39:38]Youtube Icon

🎯 How Does OpenAI Balance Consumer and Enterprise Needs?

The Mission-Driven Approach to Market Breadth

OpenAI's unique position as both a consumer and enterprise company stems from their fundamental mission rather than traditional market segmentation strategies.

The Identity Question:

Sarah Wang
One thing that is unique about OpenAI is that you're both very much a consumer company by revenue and products, but also an enterprise company. How does that work internally—what would you guys consider yourself?
Sarah WangA16ZA16Z | General Partner

The Mission-Driven Framework:

Isa Fulford
I mean, I guess if you tie it to the mission, it's like we're trying to make the most capable thing and we're also trying to make it useful to as many people as possible and accessible to as many people as possible.
Isa FulfordOpenAIOpenAI | Researcher

The Strategic Logic:

Rather than choosing between consumer and enterprise markets, OpenAI's approach flows directly from their core mission objectives:

  1. Maximum Capability - Building the most advanced AI systems possible
  2. Universal Utility - Making AI useful across all contexts and applications
  3. Broad Accessibility - Ensuring AI benefits reach the widest possible audience

The Natural Market Expansion:

When the mission focuses on universal capability and accessibility, traditional market boundaries become irrelevant. The same advanced AI system serves individual consumers and enterprise clients because the underlying goal is comprehensive utility.

The Competitive Advantage:

This mission-driven approach allows OpenAI to avoid the typical trade-offs between consumer simplicity and enterprise sophistication, instead optimizing for universal excellence that serves both markets simultaneously.

Timestamp: [39:38-40:04]Youtube Icon

🎨 What Does "Taste" Really Mean in AI Development?

Simplicity as Sophistication: The Occam's Razor of AI Research

In AI development, "taste" represents the ability to identify the simplest, most elegant solutions that work, often appearing obvious only in hindsight.

The Increased Importance:

Christina Kim
I think taste is quite important, especially now that our models are getting smarter and it's easier to use them as tools. So I think having the right direction matters a lot now.
Christina KimOpenAIOpenAI | Researcher

The Direction and Intuition Factor:

Christina Kim
Having the right intuitions and the right questions you want to ask. So I would say maybe it matters more now than before.
Christina KimOpenAIOpenAI | Researcher

The Simplicity Principle:

Isa Fulford
I've been surprised by how often the thing that is the most simple — easy to explain — is the thing that works the best.
Isa FulfordOpenAIOpenAI | Researcher

The Research Taste Definition:

Isa Fulford
I think usually good researcher taste is just simplifying the problem to the dumbest thing or the most simple thing you can do.
Isa FulfordOpenAIOpenAI | Researcher

The Hindsight Obviousness:

Christina Kim
I feel like with every research release we do, when people figure out what happened there, they're like, 'Oh, that's so simple. Oh, I should have thought of that.' Obviously, obviously that would have worked.
Christina KimOpenAIOpenAI | Researcher

The Recognition Challenge:

Christina Kim
I think it's knowing to try that obvious — or at the time not obvious — thing that is obvious in hindsight.
Christina KimOpenAIOpenAI | Researcher

The Implementation Complexity:

Isa Fulford
All of the details around the hyperparameters and all these things — the inference — that's obviously very hard, but the actual concept itself is usually pretty straightforward.
Isa FulfordOpenAIOpenAI | Researcher

The Occam's Razor Connection:

Sarah Wang
Taste is Occam's razor.
Sarah WangA16ZA16Z | General Partner

This reveals that in AI research, sophistication often lies not in complexity but in the wisdom to pursue fundamentally simple approaches that others overlook.

Timestamp: [40:12-41:38]Youtube Icon

🌟 What Does GPT-5 Represent for OpenAI's Mission?

Usability as the Ultimate Success Metric

GPT-5's launch represents the culmination of OpenAI's mission to democratize advanced AI capabilities, with "usability" emerging as the defining characteristic of this milestone.

The Defining Word:

Christina Kim
I think with GPT-5 the thing that's — the word that's been in my mind throughout all of this is, Usable.
Christina KimOpenAIOpenAI | Researcher

The Democratic Distribution:

Christina Kim
The thing that we're excited about is getting this out to everyone. We're excited to get our best reasoning models out to free users now.
Christina KimOpenAIOpenAI | Researcher

The Universal Access Achievement:

Christina Kim
Getting this — our smartest model yet — to everyone, and I'm just excited to see what people are going to actually use it for.
Christina KimOpenAIOpenAI | Researcher

Mission Fulfillment Elements:

  1. Advanced Capability - "Our smartest model yet"
  2. Broad Accessibility - Available to free users
  3. Universal Distribution - "Getting this out to everyone"
  4. Practical Utility - Focus on real-world applications

The User-Driven Discovery:

Rather than prescribing specific use cases, OpenAI's approach emphasizes enabling users to discover applications, reflecting confidence in the model's general capability and the creativity of its user base.

The Historic Context:

Erik Torenberg
Obviously historic day
Erik TorenbergA16ZA16Z | General Partner

This moment represents the practical realization of OpenAI's founding vision: advanced AI capabilities accessible to all users, regardless of technical expertise or economic resources.

The Anticipation Factor:

The emphasis on seeing "what people are going to actually use it for" demonstrates OpenAI's recognition that the most valuable applications may emerge from user innovation rather than company prescription.

Timestamp: [41:43-42:20]Youtube Icon

💎 Key Insights from [36:25-42:20]

Organizational Evolution:

  1. Scale Without Sacrifice - OpenAI grew from 200 to thousands of employees while maintaining startup agility and culture
  2. Agency-Driven Innovation - Ideas can emerge from anyone regardless of seniority, with initiative and execution capabilities rewarded
  3. Small Team Efficiency - Research teams remain intentionally small (often 2 people) to maintain nimbleness and rapid iteration

Cultural Differentiation:

  1. Research-Product Integration - Unlike traditional tech companies, researchers and engineers work in deeply integrated teams
  2. Cross-Functional Implementation - Researchers write code, engineers contribute to training runs, breaking down traditional silos
  3. Mission-Driven Identity - Consumer vs. enterprise distinction irrelevant when focus is universal capability and accessibility

AI Development Philosophy:

  1. Taste as Simplicity - Best solutions often appear obvious in hindsight but require wisdom to identify initially
  2. Occam's Razor Application - Sophisticated AI research frequently involves finding the simplest approach that works
  3. Direction Over Complexity - As models become more capable, having the right intuitions and asking right questions becomes crucial

Mission Culmination:

  1. Usability Focus - GPT-5 represents practical realization of advanced AI for everyone
  2. Democratic Distribution - Best reasoning models now available to free users
  3. User-Driven Discovery - Confidence in letting users determine most valuable applications rather than prescribing use cases

Timestamp: [36:25-42:20]Youtube Icon

📚 References from [36:25-42:20]

People Mentioned:

  • Calvin French-Owen - Former OpenAI employee whose reflections on working at the company were referenced for organizational change discussion

Organizational Concepts:

  • Applied Team - OpenAI's engineering team that grew from 10 engineers to substantial product organization
  • Product Arm - Consumer-facing division that emerged after API launch
  • Research Teams - Intentionally small units (often 2 people) maintaining nimbleness and rapid iteration

Cultural Frameworks:

  • Agency Reward System - OpenAI's approach to recognizing and empowering individual initiative regardless of hierarchy
  • Startup Culture - Maintained organizational ethos despite massive growth from 200 to thousands of employees
  • Taste - Critical capability for identifying simple, elegant solutions in AI research

Technical Integration:

  • Post-Training - Area where research-product integration is particularly common and effective
  • Model Training Runs - Collaborative area where engineers assist researchers
  • Front-end Code - Area where researchers sometimes contribute to implementation

Mission Elements:

  • Universal Capability - Goal of building the most advanced AI systems possible
  • Broad Accessibility - Ensuring AI benefits reach the widest possible audience
  • Free Users - Target audience for democratizing advanced reasoning models

Philosophical Concepts:

  • Occam's Razor - Principle of simplicity applied to AI research and development
  • Usability - Defining characteristic and success metric for GPT-5 launch
  • Consumer vs. Enterprise - Traditional market distinction that OpenAI transcends through mission focus

Timestamp: [36:25-42:20]Youtube Icon