undefined - 20VC: Cohere Founder on How Cohere Compete with OpenAI and Anthropic $BNs | Why Countries Should Fund Their Own Models & the Need for Model Sovereignty | How Sam Altman Has Done a Disservice to AI with Nick Frosst

20VC: Cohere Founder on How Cohere Compete with OpenAI and Anthropic $BNs | Why Countries Should Fund Their Own Models & the Need for Model Sovereignty | How Sam Altman Has Done a Disservice to AI with Nick Frosst

Nick Frosst is a Canadian AI researcher and entrepreneur, best known as co-founder of Cohere, the enterprise-focused LLM. Cohere has raised over $900 million, most recently a $500 million round, bringing its valuation to $6.8 billion. Under his leadership, Cohere hit $100M in ARR. Prior to founding Cohere, Nick was a researcher at Google Brain and a protégé of Geoffrey Hinton.

September 1, 202567:50

Table of Contents

0:37-7:54
8:00-15:54
16:00-23:58
24:03-31:59
32:04-39:59
40:05-47:54
48:02-55:58
56:03-1:03:58
1:04:03-1:14:23

🧠 What did Geoffrey Hinton teach Nick Frosst at Google Brain?

Learning Research Through Physical Intuition

Nick Frosst's experience as Geoffrey Hinton's first hire at Google Brain revealed an unexpected approach to AI research that fundamentally shaped his understanding of the field.

Key Learning Approach:

  1. Physical Analogies Over Equations - Hinton approached complex algorithms through natural world metaphors rather than pure mathematical formulations
  2. Creative Problem-Solving - Research discussions involved imagining balls, elastic bands, pulleys, and surfaces to understand loss functions and optimizers
  3. Curiosity-Driven Exploration - The approach was characterized by "what would happen if..." thinking rather than rigid mathematical derivations

Surprising Research Philosophy:

  • Playful Methodology: Research was approached with creativity and play rather than strict academic formality
  • Intuition-Based Understanding: Complex AI concepts were grounded in physical intuition before mathematical implementation
  • Natural Language Processing: Algorithms were discussed through descriptive, physical world analogies

This mentorship experience over three to four years taught Frosst that breakthrough AI research often comes from intuitive understanding rather than purely analytical approaches.

Timestamp: [0:48-2:12]Youtube Icon

🚀 Did Google miss the ChatGPT opportunity despite inventing transformers?

The Transformer Paradox at Google

Despite Google's foundational role in creating the transformer architecture, the company's failure to capitalize on this breakthrough raises important questions about innovation execution.

The Innovation Timeline:

  1. 2017: Google Brain published the transformer architecture with Aidan Gomez and other brilliant researchers
  2. Years Later: The technology was scaled and commercialized by external companies rather than Google
  3. Mass Exodus: All the people who worked on the original transformer left Google to continue developing the technology elsewhere

Systemic Challenges:

  • Slow Internal Commercialization: The transformer wasn't quickly scaled up or commercialized within Google's ecosystem
  • External Development: Much of the scaling work had to be done by other organizations years later
  • Talent Migration: Key researchers departed to pursue transformer development at other companies

Current State:

  • DeepMind Integration: Google has now consolidated AI efforts under DeepMind, which continues producing quality work
  • Ongoing Innovation: The company still maintains brilliant researchers and develops good products
  • Competitive Position: Despite early advantages, Google wasn't at the forefront of the consumer AI revolution

Timestamp: [2:12-3:23]Youtube Icon

🏢 How does Cohere differentiate from OpenAI and Anthropic?

Enterprise-First AI Strategy

Cohere positions itself uniquely in the foundational model landscape through singular focus on enterprise applications rather than general consumer use cases.

Market Position:

  • Exclusive Club: One of fewer than 20 companies globally building large language models
  • Geographic Distribution: Most companies are in America, a handful in China, Cohere in Canada, and one in France
  • Foundational Model Company: Builds language models similar to OpenAI and Anthropic but with different focus

Enterprise Specialization:

  1. Tool Integration: Models trained specifically for enterprise tool use and API integration
  2. Business Data Access: Designed to work with proprietary business data and internal systems
  3. Workplace Augmentation: Focused on helping users accomplish work tasks rather than general conversation

Training Methodology:

  • Synthetic Business Environments: Creates fake companies, emails, and APIs for training purposes
  • Workplace-Specific Data: Uses data types specifically relevant to business contexts
  • Task-Oriented Learning: Trains models to augment workplace productivity rather than engagement metrics

This approach differentiates Cohere from consumer-focused models that prioritize conversational ability and user engagement.

Timestamp: [3:23-5:39]Youtube Icon

📊 Is data or compute the bigger bottleneck for AI progress?

The Data Quality Challenge

Despite advances in synthetic data generation, real-world data remains a critical constraint in AI development, with quality being more important than quantity.

Data Bottleneck Reality:

  • Still Constrained: Data remains a bottleneck despite synthetic data capabilities
  • Real Data Foundation: High-quality real-world data is essential to start synthetic data processes
  • Quality Over Quantity: Access to high-quality data is still a major consideration for AI companies

Synthetic Data Benefits:

  1. Model Improvement: Synthetic data makes models better than they would be without it
  2. Scaling Capability: Provides ability to generate large amounts of training data
  3. Controlled Environments: Allows creation of specific scenarios for targeted training

Continued Real Data Needs:

  • In-House Annotation: Companies still employ human annotators to create real, non-synthetic data
  • Quality Standards: Real data provides the quality benchmark for synthetic data generation
  • Foundation Requirement: Synthetic data processes require real-world data as a starting point

The combination of real and synthetic data appears to be the current optimal approach for training enterprise-focused AI models.

Timestamp: [5:39-6:09]Youtube Icon

⚙️ Which is the biggest constraint: compute, algorithms, or data?

Algorithms Haven't Changed Much

An analysis of the three pillars of AI development reveals that algorithmic innovation has been surprisingly stable, while data quality remains the primary constraint.

Algorithm Stability:

  • Minimal Changes: Algorithms haven't evolved significantly since the transformer architecture
  • Transformer Dominance: The industry continues using the same model architecture approaching 10 years
  • Incremental Improvements: Changes have been small rather than revolutionary

Training Evolution:

  1. Base Models: Originally just called "large language models," trained to complete sentences
  2. Human Feedback Integration: Added reinforcement learning from human feedback (RLHF) with SFT data
  3. Multiple Steps: Now involves base modeling, reinforcement learning, and various other RL techniques

Current Bottleneck Assessment:

  • Not Algorithms: Algorithms are not the constraint for making models more useful
  • Data Quality Focus: The primary bottleneck is obtaining good quality data
  • Synthetic Data Generation: Creating high-quality synthetic data from real data remains challenging

The evolution from simple sentence completion to conversational AI required training methodology changes rather than fundamental algorithmic breakthroughs.

Timestamp: [6:09-7:16]Youtube Icon

📈 Why does Nick Frosst think GPT-5 was worse than GPT-4?

Questioning the Scaling Laws Narrative

Frosst's assessment that GPT-5 was actually worse than GPT-4 challenges the prevailing belief that more compute automatically leads to better AI performance.

Scaling Laws Skepticism:

  • Compute ≠ Progress: The assumption that more compute leads to exponential progress may be flawed
  • Performance Regression: GPT-5's perceived inferior performance suggests scaling limitations
  • 12-24 Month Timeline: Questions whether scaling laws will continue delivering benefits in the near term

Industry Narrative Challenge:

  1. Popular Belief: The industry widely believes more compute will drive continuous exponential progress
  2. Evidence Against: GPT-5's performance suggests this narrative may be incorrect
  3. Scaling Plateau: Indicates potential plateauing in model capabilities despite increased resources

Implications for AI Development:

  • Resource Allocation: Questions whether throwing more compute at problems is the optimal strategy
  • Alternative Approaches: Suggests need for different methodologies beyond pure scaling
  • Performance Evaluation: Highlights importance of objective assessment over marketing narratives

This perspective suggests that the AI industry may need to reconsider its fundamental assumptions about the relationship between computational resources and model performance.

Timestamp: [7:16-7:54]Youtube Icon

💎 Summary from [0:37-7:54]

Essential Insights:

  1. Geoffrey Hinton's Research Philosophy - AI breakthroughs come from creative, intuitive thinking using physical analogies rather than pure mathematical approaches
  2. Google's Transformer Paradox - Despite inventing the transformer architecture, Google failed to capitalize on it, leading to mass talent exodus and external commercialization
  3. Enterprise AI Differentiation - Cohere's singular focus on enterprise applications through specialized training data and synthetic business environments sets it apart from consumer-focused competitors

Actionable Insights:

  • Research Methodology: Approach complex problems through physical intuition and creative analogies before diving into mathematical formulations
  • Innovation Execution: Having breakthrough technology isn't enough; organizations need systems to rapidly commercialize and scale innovations
  • Market Positioning: Specialized focus on specific use cases (enterprise vs. consumer) can create competitive advantages in crowded markets
  • Scaling Skepticism: Question industry narratives about scaling laws and evaluate model performance objectively rather than accepting marketing claims

Timestamp: [0:37-7:54]Youtube Icon

📚 References from [0:37-7:54]

People Mentioned:

  • Geoffrey Hinton - Nick Frosst's mentor at Google Brain, legendary AI researcher who taught research through physical analogies and creative intuition
  • Aidan Gomez - Co-author of the transformer paper at Google Brain in 2017, now CEO of Cohere

Companies & Products:

  • Google Brain - Google's AI research division where the transformer architecture was invented
  • DeepMind - Google's AI subsidiary that has now subsumed other Google AI efforts
  • OpenAI - Competitor in the foundational model space, known for ChatGPT
  • Anthropic - Another competitor building foundational AI models
  • Cohere - Enterprise-focused foundational model company co-founded by Nick Frosst

Technologies & Tools:

  • Transformer Architecture - The foundational neural network architecture published by Google Brain in 2017 that revolutionized AI
  • ChatGPT - OpenAI's consumer AI product that popularized conversational AI
  • GPT-4 and GPT-5 - OpenAI's language models discussed in performance comparison

Concepts & Frameworks:

  • Reinforcement Learning from Human Feedback (RLHF) - Training methodology that improved AI models' conversational abilities
  • Scaling Laws - The belief that more computational resources lead to exponentially better AI performance
  • Synthetic Data Generation - Creating artificial training data to supplement real-world data for model training
  • SFT Data - Supervised Fine-Tuning data used in the reinforcement learning process

Timestamp: [0:37-7:54]Youtube Icon

🤖 Why does Cohere's Nick Frosst think ChatGPT got worse?

Product Experience and Model Selection Issues

Nick Frosst believes ChatGPT's user experience has deteriorated due to their model auto-selection feature:

Key Problems Identified:

  1. Slower Response Times - The model selection process adds unnecessary delays
  2. Overcomplicated Responses - Simple queries trigger deep research mode when users just want quick answers
  3. User Frustration - The system often misinterprets user intent, providing PhD-level analysis for basic questions

The Core Issue:

  • Delayed Implementation: OpenAI waited 1-1.5 years to implement model auto-selection
  • Poor User Control: Users can't easily specify they want simple, direct answers
  • Mismatched Expectations: The system assumes complexity when simplicity is preferred

Timestamp: [8:00-8:27]Youtube Icon

🏢 Why does Cohere focus on enterprise over consumer AI?

Strategic Business Focus and Personal vs Professional Use Cases

Nick Frosst explains Cohere's enterprise-first approach based on fundamental differences between personal and professional AI needs:

Personal Life Limitations:

  • Authentic Communication: Doesn't want to automate personal responses like texting his mom
  • Human Connection: Values writing personal messages himself rather than delegating to AI
  • Limited Automation Desire: Most personal tasks don't benefit from AI intervention

Enterprise Opportunities:

  1. Complex Workflow Automation - Multi-step processes like expense filing
  2. Document Processing - Analyzing emails, receipts, and internal documentation
  3. System Integration - Connecting with company APIs and approval workflows
  4. Compliance Management - Cross-referencing expenses with company policies

The Vision:

A comprehensive system that can handle commands like "file my expenses" by automatically:

  • Scanning through emails and receipt photos
  • Cross-referencing with internal expense policies
  • Interfacing with company expense APIs
  • Obtaining necessary approvals

Timestamp: [8:49-9:39]Youtube Icon

🎯 What is Nick Frosst's definition of AGI?

Artificial General Intelligence as Human-Like Computer Interaction

Nick Frosst provides a clear, behavioral definition of AGI that differs from typical technical specifications:

His AGI Definition:

"A computer that you treat like a person" - When users naturally expect and interact with a computer as they would with another human being.

Current Reality Check:

  • Not There Yet: People don't currently treat language models like they treat people
  • Behavioral Gap: There's still a clear distinction in how users approach AI vs human interaction
  • Expectation Management: Users still adjust their communication style for AI systems

Industry Context:

  • Shifting Definitions: OpenAI and Microsoft have changed their AGI definitions multiple times
  • Lack of Consensus: The AI community has "many years of people discussing AGI and not many definitions thereof, next to none"
  • Practical Approach: Focuses on user behavior rather than technical capabilities

Timestamp: [10:10-10:48]Youtube Icon

📈 Does Nick Frosst believe scaling laws still work for AI?

Skepticism About Compute-Driven Progress and Public Perception

Nick Frosst challenges the prevailing narrative around scaling laws and infinite compute leading to AGI:

His Position on Scaling:

  • Technology Won't Reach AGI: Current language model technology doesn't get us to true AGI
  • Ongoing Work Required: Progress needs more modeling work, product development, and better integrations
  • Non-Plateauing Areas: Building safer data integrations and connectors still has room for improvement

Public Opinion Reality:

  • Academic Skepticism: Most university computer science students don't believe throwing more compute will achieve AGI
  • Broader Doubt: "A lot of the world does not think scaling laws are super prevalent"
  • Industry vs Reality: The hype around scaling laws may not reflect actual expert consensus

Focus Areas for Real Progress:

  1. Product Development - Better user interfaces and experiences
  2. Integration Work - Connecting models with existing systems
  3. Safety Improvements - Trustworthy data handling and permissions
  4. Specialized Applications - Enterprise-focused model refinements

Timestamp: [8:27-11:18]Youtube Icon

🔧 How does Cohere view AI model specialization vs generalization?

The Spectrum Between Task-Specific and Universal Models

Nick Frosst explains how modern language models occupy a middle ground between extreme specialization and complete generalization:

Historical Context - 2015 Machine Learning:

  • Task-Specific Training: Every neural network task required training a model from scratch
  • Example: Cat identification models were trained specifically on cat images
  • Data-Centric Approach: Create dataset → train new model → deploy for single use case

Modern Language Model Reality:

  • Foundation + Refinement: Train on all language, then refine for specific tasks
  • Cannot Train on Single Tasks: Summarization models can't just train on summarization data
  • Spectrum Positioning: Not at either extreme - somewhere in the middle

Cohere's Enterprise Approach:

  1. General Language Competency - Models good at all language tasks
  2. Enterprise Refinement - Specialized for business environments
  3. Specific Capabilities:
  • Using internal company tools
  • Reading massive documentation
  • Secure deployment requirements
  • Customer customization needs

Industry Examples:

  • Anthropic's Code Models: Good at code generally, not just specific coding tasks like debugging or refactoring
  • Specialization Sweet Spot: Broad capability within a domain rather than narrow task focus

Timestamp: [12:42-14:52]Youtube Icon

🎯 Why doesn't Cohere care about AI benchmark scores?

Customer Success Over Competitive Metrics

Nick Frosst explains why Cohere prioritizes practical customer outcomes over industry benchmark performance:

Cohere's Philosophy:

  • Customer-First Optimization: Focus on whether customers can easily accomplish their goals
  • Practical Success Metrics: "If a customer uses our model, gets a copy of our model from us, and they try to do something with it, we care that it works as easy as possible"
  • Real-World Performance: Optimize for actual use cases rather than theoretical benchmarks

Problems with Current Benchmarks:

  1. Misaligned Metrics: Benchmarks don't reflect real customer needs
  2. Annual Cycling: Benchmarks change yearly without consistency
  3. Hype-Driven: More focused on discourse and marketing than practical utility
  4. Limited Scope: Don't capture enterprise-specific requirements

What Matters Instead:

  • Ease of Implementation - How quickly customers can deploy and use the model
  • Task-Specific Performance - Success in actual customer workflows
  • Integration Quality - How well models work with existing enterprise systems
  • Customization Capability - Ability to adapt to specific customer needs

Industry Criticism Response:

When critics point to benchmark scores, Cohere acknowledges the "whole long conversation to be had about evals" but maintains focus on customer satisfaction over competitive rankings.

Timestamp: [14:52-15:49]Youtube Icon

💎 Summary from [8:00-15:54]

Essential Insights:

  1. Product Experience Matters - ChatGPT's model auto-selection feature made the user experience worse by adding delays and complexity
  2. Enterprise vs Consumer Focus - AI has more practical value in professional workflows than personal life automation
  3. AGI Definition Clarity - True AGI means treating computers like people, which we haven't achieved yet

Actionable Insights:

  • Benchmark Skepticism: Focus on customer success metrics rather than industry benchmark scores
  • Specialization Strategy: Modern AI models work best when trained broadly then refined for specific domains
  • Scaling Law Reality: Most experts don't believe infinite compute alone will achieve AGI

Timestamp: [8:00-15:54]Youtube Icon

📚 References from [8:00-15:54]

People Mentioned:

  • Sam Altman - OpenAI CEO referenced regarding AGI definitions and their changes over time
  • Geoffrey Hinton - Mentioned as Nick Frosst's mentor during his time at Google Brain

Companies & Products:

  • OpenAI - Discussed regarding ChatGPT's model selection issues and consumer product focus
  • Microsoft - Referenced as OpenAI's partner in defining AGI milestones
  • Anthropic - Mentioned for their Claude model and code-focused capabilities
  • Cursor - AI coding tool referenced as being challenged by Anthropic's offerings
  • Cohere - Nick Frosst's company focused on enterprise AI solutions
  • Eleven Labs - Referenced as example of specialized voice AI use case

Technologies & Tools:

  • ChatGPT - Discussed regarding user experience degradation with model auto-selection
  • Claude - Anthropic's AI model mentioned for code capabilities
  • Language Models/LLMs - Core technology discussed throughout the segment

Concepts & Frameworks:

  • Scaling Laws - The theory that more compute leads to better AI performance, which Frosst questions
  • AGI (Artificial General Intelligence) - Defined by Frosst as computers treated like people
  • Model Specialization vs Generalization - The spectrum between task-specific and universal AI models
  • Enterprise AI Applications - Focus on business workflow automation rather than consumer use cases

Timestamp: [8:00-15:54]Youtube Icon

🎯 Why AI benchmarks don't reflect real customer needs according to Cohere's Nick Frosst?

The Disconnect Between Academic Benchmarks and Enterprise Reality

Evolution of AI Benchmarks:

  1. LM1B (Early Era) - Completing newspaper articles based on first part
  2. HellaSwag - Common sense reasoning tasks
  3. Current Focus - Math reasoning (AIME) and pixel manipulation (ARC-AGI)

The Customer Reality Check:

  • Math reasoning benchmarks: None of Cohere's customers ask models to do math reasoning in workplace settings
  • ARC AGI challenge: Pixel manipulation tasks that no enterprise customer has ever requested
  • Workplace applications: Don't align with academic benchmark priorities

The Gamification Problem:

  • Benchmarks can be easily gamed by training specifically on them
  • Results reflect how much models were trained on those benchmarks rather than true utility
  • Big players may gamify: Though Nick admits uncertainty about industry practices

Enterprise vs Consumer Perspectives:

  • Consumer space: Benchmarks create excitement and drive user engagement with latest models
  • Enterprise space: Customers care about production deployment and ROI, not leaderboard rankings
  • Real success metrics: "Did I get to production? Did I buy LLMs, deploy them and get ROI?"

Timestamp: [16:00-18:06]Youtube Icon

⚡ How fast is AI model evolution really progressing compared to hardware

The Paradox of Speed in AI Development

The Perception vs Reality:

  • Rapid model releases: Creating impression of exponential progress
  • Fundamental architecture: Still using transformers invented in 2017
  • Core mechanism: Models still "take in words and predict the next word"

Training Time Evolution:

  • 2011 neural nets: Hours to days of training time
  • Current LLMs: Months of training required for state-of-the-art models
  • Hardware constraint: Still training on H100s and Nvidia chips from 18 months ago

The Training Pipeline Additions:

  1. Base modeling step - Initial transformer training
  2. Supervised Fine-Tuning (SFT) - Human-written sentence-response pairs
  3. Reinforcement Learning from Human Feedback - Model generates responses, receives good/bad feedback

The Dichotomy:

  • Surface level: Constant new model releases and iterations
  • Fundamental level: Same core technology for years
  • Innovation reality: Improved training methods, not revolutionary architecture changes

Timestamp: [18:13-20:08]Youtube Icon

💰 Would Cohere spend $5M on a single AI researcher like Meta and Anthropic

The Reality of AI Talent Economics

The Compensation Landscape:

  • Industry headlines: Stories of $10-20 million packages for top AI researchers
  • Value creation: Many researchers genuinely bringing equivalent value to companies
  • Market dynamics: Super impactful industry justifying high compensation

What Drives High Compensation:

  • Demanding work: Requires extensive experience and ingenuity
  • Dedication required: Significant time investment and specialized skills
  • Industry impact: Transformative technology creating massive value

Cohere's Approach to Talent:

  • People-first philosophy: "The company is the people who are there"
  • Comprehensive consideration: Financial perspective plus workplace environment
  • Equity participation: Many employees already own significant equity stakes
  • Value-based decisions: "If they were bringing in the right value, yeah" - willing to spend $5M

The Retention Challenge:

  • Stability matters: People want workplace stability beyond just compensation
  • Purpose alignment: Employees seek value-aligned organizations
  • Work satisfaction: Feeling good about the work they're doing
  • Mixed signals: Stories of huge hires followed by quick departures at major companies

Timestamp: [20:13-22:54]Youtube Icon

🎪 What AI hype and misleading rhetoric does Nick Frosst find most damaging

The Beautiful Technology vs Dangerous Hype

The Technology Reality:

  • Transformative impact: "Most beautiful technology I've ever seen"
  • Personal transformation: Already fundamentally changing how Nick works
  • Future certainty: Will fundamentally change how everyone works soon
  • Genuine innovation: Represents real technological breakthrough

The Problematic Hype:

  • AGI assumptions: Most damaging and confusing aspect of current discourse
  • Work displacement fears: "We will all have no work to do"
  • UBI inevitability: Assumption that Universal Basic Income becomes necessary
  • Misleading rhetoric: Creates confusion rather than clarity

The Information Problem:

  • Misinformation spread: Lots of false information circulating
  • Truth obstruction: Hype is not helpful for getting to the truth
  • Industry confusion: Makes it harder to understand real capabilities and limitations

The Balanced Perspective:

  • Acknowledge power: Technology is genuinely transformative
  • Reject extremes: Neither dismissive nor apocalyptic views are helpful
  • Focus on reality: Practical applications and limitations matter more than speculation

Timestamp: [22:54-23:58]Youtube Icon

💎 Summary from [16:00-23:58]

Essential Insights:

  1. Benchmark Reality Gap - Academic AI benchmarks don't reflect real enterprise needs; customers care about production ROI, not leaderboard rankings
  2. Technology Evolution Paradox - Despite rapid model releases, AI still uses 2017 transformer architecture with improved training methods rather than fundamental breakthroughs
  3. Talent War Economics - High AI researcher compensation reflects genuine value creation, but workplace culture and stability matter beyond just money

Actionable Insights:

  • Focus on practical deployment metrics rather than academic benchmark performance when evaluating AI solutions
  • Understand that AI progress is iterative improvement on existing architecture, not revolutionary leaps
  • Consider comprehensive talent retention strategies including purpose, stability, and equity alongside competitive compensation
  • Distinguish between transformative AI capabilities and harmful hype around job displacement and AGI timelines

Timestamp: [16:00-23:58]Youtube Icon

📚 References from [16:00-23:58]

People Mentioned:

  • Joëlle Pineau - Recently hired by Cohere from Facebook/Meta, mentioned in context of talent acquisition
  • Mark Zuckerberg - Referenced for willingness to pay billion-dollar compensation for chief scientists
  • Aidan - Mentioned as part of Cohere's leadership team in compensation discussions

Companies & Products:

  • Meta - Referenced for high-profile AI researcher hiring and compensation packages
  • Anthropic - Mentioned as paying $10-20 million packages for top AI researchers
  • Nvidia - H100 chips referenced as current training hardware standard

Technologies & Tools:

  • Transformers - Core AI architecture invented in 2017, still fundamental to current models
  • H100s - Nvidia's current generation training chips mentioned as 18 months old
  • Supervised Fine-Tuning (SFT) - Training method using human-written sentence-response pairs
  • Reinforcement Learning from Human Feedback - Training approach where models receive good/bad feedback

Concepts & Frameworks:

  • LM1B - Early AI benchmark focused on completing newspaper articles
  • Hello Swag - 2022-era AI benchmark for common sense reasoning
  • AIM (Math Reasoning) - Current benchmark focused on mathematical problem solving
  • ARC AGI Challenge - Pixel manipulation benchmark for testing AI reasoning
  • Universal Basic Income (UBI) - Economic concept mentioned in context of AI job displacement hype

Timestamp: [16:00-23:58]Youtube Icon

🤖 Does AI pose an existential threat to humanity?

Debunking the AGI Doom Narrative

The existential threat discourse around AI has significantly diminished this year, and for good reason - it was fundamentally incorrect. This narrative was not only wrong but actively harmful in several ways:

Why the Existential Threat Narrative Was Wrong:

  1. No Imminent Danger: The technology doesn't pose an immediate existential threat to humanity at large
  2. Distraction from Real Issues: It prevented meaningful discussion about actual ways AI could be damaging
  3. Poor Public Understanding: It made it harder for people to understand what the technology actually is and does

Real AI Risks We Should Focus On:

  • Workforce Disruption: Rapid changes to employment and job markets
  • Income Inequality: Potential widening of economic gaps without proper policy
  • System Disruption: Significant changes to how organizations and societies function

Current State of the Discourse:

The remnants of this fear-based narrative still persist, particularly visible in how employees at large organizations respond to AI implementation. Many remain nervous and scared rather than embracing the technology's potential benefits.

Timestamp: [24:03-24:54]Youtube Icon

💼 How do employees really react to AI in the workplace?

The Reality of AI Adoption in Enterprises

There's a notable difference in how AI adoption is perceived versus experienced in practice, with geographic and organizational variations playing a significant role.

Employee Reactions Vary by Region:

  • European Organizations: More prevalent nervousness and resistance to AI introduction
  • North American Enterprises: Generally more interest and excitement about using LLMs
  • Overall Pattern: Many employees do not welcome AI introduction wholeheartedly in large companies

Why Some Employees Embrace AI:

  1. Augmentative Nature: They recognize LLMs enhance rather than replace their capabilities
  2. Task Elimination: AI allows them to avoid doing mundane tasks they don't enjoy
  3. Productivity Enhancement: Provides tools to focus on more meaningful work

The Cohere Experience:

When working with enterprise customers, the company has found that most people are interested and excited about using LLMs, primarily because they understand the technology's augmentative potential rather than viewing it as a replacement threat.

Timestamp: [24:54-25:34]Youtube Icon

🎯 Will AI agents replace young marketing professionals?

The Limits of AI Replacement in Creative Roles

A heated debate emerges around whether AI agents will replace entry-level marketing professionals, revealing fundamental disagreements about AI capabilities and limitations.

The Replacement Argument:

  • Target Demographics: 25-26 year old marketing managers and SDRs
  • Skill Assessment: Many lack exceptional craft mastery or deep passion
  • Timeline Prediction: Phenomenal agents expected to surpass them within 12 months
  • Inevitability Claim: These roles will be replaced entirely

The Augmentation Counter-Argument:

What LLMs Excel At:

  • Tasks involving text manipulation and information synthesis
  • Combining existing documents and data into new formats
  • Providing starting points for creative work

What Humans Still Do Better:

  1. Interpersonal Communication: Talking to people and building relationships
  2. Cultural Understanding: Grasping zeitgeist, trends, and what will resonate
  3. Intuitive Decision-Making: Using human experience to determine relevance and impact
  4. Creative Insight: Understanding context that isn't captured in internet text datasets

The Reality of Current AI Usage:

Modern marketing professionals use AI as a starting point - prompting for campaign storylines, getting initial ideas, then applying human judgment to refine, select, and develop the most promising concepts. This represents augmentation rather than replacement.

Timestamp: [25:34-28:47]Youtube Icon

🔬 Why haven't LLMs made independent breakthroughs yet?

Understanding the Fundamental Limitations of Current AI

Despite years of LLM usage and incredible capabilities, there's a notable absence of independent breakthroughs made by AI systems, revealing important insights about their nature.

The Breakthrough Gap:

  • No Independent Discoveries: LLMs haven't solved problems that no human has solved before
  • Human-Driven Innovation: All major breakthroughs still originate from human researchers
  • Pattern Recognition: No instances of LLMs being asked to solve unsolved problems and succeeding

Why This Isn't Just a Matter of Time:

Fundamental Technical Limitations:

  1. Statistical Nature: LLMs are statistical models trained on text data
  2. Sequence Modeling: They work by predicting patterns in existing information
  3. Generalization Boundaries: While capable of generalizing across unseen tasks, they're limited by training data

What This Means for AI Capabilities:

  • Phenomenal at Known Patterns: Excellent at tasks involving existing information and established patterns
  • Limited Novel Discovery: Cannot generate truly original insights beyond their training scope
  • Augmentative Role: Best suited for enhancing human capabilities rather than replacing human creativity and problem-solving

This limitation explains why LLMs excel at reformatting information, combining existing knowledge, and providing starting points, but struggle with genuine innovation and breakthrough discoveries.

Timestamp: [26:41-27:22]Youtube Icon

🏢 What will companies look like in 5-10 years with AI?

The Future Workplace: Language-First Computing

The workplace transformation will mirror historical technological shifts, fundamentally changing how people interact with computers and organize their work.

The Daily Work Experience:

Primary Interface Changes:

  • Language-Based Computing: Predominantly using natural language to interact with computers
  • Automated Mundane Tasks: AI handles boring, information-processing work that doesn't require creativity
  • Human-Focused Activities: More time spent on interpersonal communication and strategic thinking

Workflow Transformation:

  1. Arrive at Work: Sit in front of a computer as usual
  2. Speak to Systems: Use language commands instead of traditional interfaces
  3. Delegate Routine Work: AI handles tasks with available information that don't need insight
  4. Focus on Value-Add: Spend time talking to people, evaluating AI output, and strategic planning

Historical Context for Change:

Previous Workforce Transformations:

  • Computer Introduction: Fundamental changes to how work gets done
  • Personal Computer Era: Shifted individual productivity capabilities
  • Internet Revolution: Connected global information and communication
  • Industrial Revolution: Massive workforce restructuring (90%+ farm workers to <5%)

Managing the Transition:

Critical Considerations:

  • Workforce Resilience: Ensuring labor force can adapt to changes
  • Income Inequality: Preventing AI from widening economic gaps
  • Human-Centric Design: Ensuring AI enables people to do work they're good at and enjoy
  • Policy Framework: Developing supportive labor policies for smooth transition

Timestamp: [29:25-30:42]Youtube Icon

⚖️ Will AI increase or decrease income inequality?

Policy Determines AI's Impact on Economic Disparity

The relationship between AI and income inequality isn't predetermined - it depends entirely on the policy frameworks societies choose to implement.

Policy-Dependent Outcomes:

  • Good Labor Policy: AI could help reduce income inequality
  • Bad Labor Policy: AI could significantly worsen economic disparities
  • Historical Precedent: Previous technological revolutions show both positive and negative outcomes based on policy responses

Lessons from Industrial Revolution:

What We Did Right:

  1. Universal Benefit Recognition: Everyone now agrees industrial automation was beneficial overall
  2. Agricultural Transformation: Moved from 90%+ farm workers to less than 5% efficiently
  3. Labor Protection Development: Created unions, workers' rights, and protective policies

What We Did Wrong Initially:

  • Child Labor: Kids working in dangerous conditions like coal mines
  • Poor Working Conditions: Exploitative practices that were quickly recognized as problematic
  • Inadequate Protections: Initial lack of worker safety and rights frameworks

Key Policy Areas for AI Era:

Essential Focus Areas:

  • Labor Force Resilience: Ensuring workers can adapt and transition
  • Income Distribution: Preventing concentration of AI benefits among few
  • Worker Rights: Developing new protections for AI-augmented workplace
  • Transition Support: Helping displaced workers find new opportunities

The historical pattern suggests that while technological revolutions create temporary disruption and inequality, proper policy responses can ensure broad-based benefits and improved living standards for everyone.

Timestamp: [30:54-31:59]Youtube Icon

💎 Summary from [24:03-31:59]

Essential Insights:

  1. AI Existential Threat Narrative Debunked - The discourse around AI posing imminent existential threats has diminished because it was fundamentally incorrect and harmful to productive discussion
  2. Workplace AI Adoption Varies Geographically - European organizations show more resistance while North American enterprises demonstrate greater enthusiasm for LLM implementation
  3. AI Augmentation vs Replacement Debate - Fundamental disagreement exists about whether AI will replace workers or augment their capabilities, with evidence supporting augmentation in creative roles

Actionable Insights:

  • Focus on Real AI Risks: Address workforce disruption, income inequality, and system changes rather than existential threats
  • Understand AI Limitations: LLMs excel at pattern recognition and information synthesis but cannot make independent breakthroughs or discoveries
  • Prepare for Language-First Computing: Future workplaces will primarily use natural language interfaces with AI handling routine tasks while humans focus on interpersonal and strategic work
  • Policy Determines Outcomes: AI's impact on income inequality depends entirely on labor policy choices, with historical precedent showing both positive and negative possibilities

Timestamp: [24:03-31:59]Youtube Icon

📚 References from [24:03-31:59]

People Mentioned:

  • Benny - Salesforce executive who appeared on the show two days prior, discussed "human plus agent" concept

Companies & Products:

  • Salesforce - Referenced in context of AI agent discussion and human augmentation philosophy
  • Evian - Used as example brand in marketing campaign scenario for AI usage

Historical Events & Concepts:

  • Industrial Revolution - Historical precedent for workforce transformation and policy development
  • Agricultural Revolution - Transformation from 90%+ farm workers to less than 5% of workforce
  • Computer Era - Previous technological shift that changed workplace dynamics
  • Personal Computer Revolution - Individual productivity transformation period
  • Internet Creation - Global connectivity and information access revolution
  • Printing Press - Historical example of technology disrupting existing systems

Technologies & Frameworks:

  • Large Language Models (LLMs) - Core technology discussed for workplace augmentation and limitations
  • Statistical Models of Text - Technical description of how current AI systems function
  • Sequence Models - Underlying architecture explaining AI capabilities and limitations

Policy Concepts:

  • Labor Policy - Critical factor determining AI's impact on income inequality
  • Workers' Rights - Historical development from industrial revolution
  • Unions - Labor organization structure that emerged from previous technological disruption
  • Income Inequality - Economic disparity concern related to AI implementation

Timestamp: [24:03-31:59]Youtube Icon

🏛️ Why does Cohere's Nick Frosst believe AI needs significant policy changes?

Economic Impact and Policy Necessity

Nick Frosst emphasizes that while AI has the potential to augment human productivity and create a better economy, significant policy intervention is crucial to prevent exacerbating existing problems.

Current Economic Concerns:

  • Income inequality has been rising over several years, even before AI and language models became popular
  • Technology deployment without proper policy frameworks could worsen these existing disparities
  • Employment policies are essential to ensure AI benefits are distributed fairly

Historical Precedent:

  • Many past technological advances that led to better productivity and economic outcomes required coordination between businesses and governments
  • Public policy played a crucial role in ensuring technology served broader societal interests
  • Unison approach between private and public sectors has been key to successful technology adoption

The concern isn't that AI will replace humans, but that without proper policy frameworks, the benefits may not be equitably distributed, potentially widening the gap between different economic groups.

Timestamp: [32:04-32:56]Youtube Icon

🔓 What is Cohere's unique approach to open versus closed AI models?

The Middle Ground Strategy

Cohere has developed a distinctive approach that balances openness with commercial viability, positioning itself between fully open and completely closed AI models.

Cohere's Hybrid Model:

  • Non-commercial release: Foundational model weights are available for scientific research and personal use
  • Commercial licensing: Businesses must establish commercial relationships for enterprise usage
  • Community credibility: Allows researchers and developers to validate and test models before committing
  • Business sustainability: Maintains revenue streams while supporting the research community

Strategic Benefits:

  1. Validation opportunity - Potential customers can test models on their specific problems
  2. Community trust - Transparency builds credibility within the AI research community
  3. Commercial protection - Revenue model remains intact for business operations
  4. Competitive advantage - Few foundational model companies take this balanced approach

Industry Perspective:

Nick notes surprise that more AI companies haven't adopted this middle-ground strategy, especially given that many companies that started as "open" have moved away from releasing model weights entirely.

Timestamp: [33:01-34:15]Youtube Icon

🎯 How should founders balance competitive awareness with company focus?

The Middle Ground Philosophy

Nick Frosst advocates for a balanced approach to competitive intelligence, warning against both extremes of obsessive competitor monitoring and complete market ignorance.

The Two Dangerous Extremes:

  1. Over-analysis paralysis:
  • Spending entire time analyzing competitors' every move
  • Obsessing over minute details like "2% better performance on specific benchmarks"
  • Getting distracted by constant small changes in other businesses
  1. Willful ignorance:
  • Operating with "head in the sand" mentality
  • Only focusing internally without market awareness
  • Missing important industry developments

The Balanced Approach:

  • Natural awareness: AI discourse is inescapable - "every other headline" covers AI developments
  • Customer-focused questions: Be prepared to answer customer concerns about competitive positioning
  • Grounded priorities: Stay focused on core questions:
  • What are you actually doing?
  • Who are you actually helping?
  • How is this making things better for your customers?

Industry Reality:

Most AI industry professionals suffer from information overload rather than lack of competitive intelligence, making selective attention more valuable than comprehensive monitoring.

Timestamp: [34:33-35:57]Youtube Icon

🔮 Will prompting still be the primary AI interaction method in 5 years?

The Evolution of Human-AI Interaction

Nick Frosst predicts that while the fundamental concept of prompting will persist, the skill-based aspect of "prompt engineering" will largely disappear as models become more intuitive.

Historical Evolution of Prompting:

  • Early days: Required "tricking" models with specific formatting

  • Example: Writing "paragraph + 'In summary:' + newline" to get summaries

  • Models only trained on web text, not human feedback

  • Success required understanding model limitations and workarounds

  • Current trajectory: Models increasingly trained to match human expectations

  • Less need for specialized prompting techniques

  • More natural language interaction patterns

  • Reduced importance of "prompting as a skill"

Future Prediction:

  1. Basic interaction remains: Writing/saying something to a model and getting responses back will continue
  2. Iteration persists: Refining requests when initial responses aren't satisfactory
  3. Skill requirement changes: From "prompt engineering" to understanding how language models work

Essential Knowledge Framework:

Instead of prompting skills, professionals will need to understand:

  • Core functionality: How language models operate and are trained
  • Capability boundaries: What models can and cannot do
  • Emergent behaviors: Which capabilities arise naturally vs. which don't
  • Technology limitations: Avoiding "digital god" or "magic spell" thinking

This mirrors how people learned to use computers, telephones, and other technologies - understanding capabilities and limitations rather than mastering arcane interaction methods.

Timestamp: [36:02-38:40]Youtube Icon

💰 What was Nick Frosst's experience with Cohere's major fundraising round?

Fundraising Evolution and Personal Experience

Nick Frosst shares insights from Cohere's significant fundraising experience, highlighting how investor conversations have matured alongside the AI industry.

Personal Role and Approach:

  • Collaborative effort: Multiple team members involved, not solely led by Nick
  • Technical focus: Primarily discussed technology and product development
  • Genuine enthusiasm: Enjoyed conversations with VCs and pension funds due to passion for Cohere's mission
  • Natural fit: Combined love for the company with natural inclination for discussion

Industry Maturation - Then vs. Now:

2-3 Years Ago:

  • Questions focused on basic understanding: "What is this? How does that work?"
  • Significant time spent explaining fundamental concepts
  • Educational component dominated conversations

Current Environment:

  • Investors understand the technology and its applications
  • Conversations focus on specific customer implementations:
  • RBC's usage patterns
  • Fujitsu partnerships
  • LG integration examples
  • More interesting discussions about practical applications rather than theoretical concepts

Key Insight:

The fundraising process has become more sophisticated and productive as the industry has matured, allowing for deeper conversations about actual business impact rather than basic technology education.

Timestamp: [38:45-39:59]Youtube Icon

💎 Summary from [32:04-39:59]

Essential Insights:

  1. Policy intervention necessity - AI requires significant policy changes not because it replaces humans, but to prevent exacerbating existing income inequality that predates AI adoption
  2. Hybrid business model success - Cohere's approach of releasing weights for non-commercial use while requiring commercial licensing creates community credibility and business sustainability
  3. Competitive intelligence balance - Founders should avoid both obsessive competitor monitoring and market ignorance, focusing instead on customer value creation

Actionable Insights:

  • Consider hybrid open/closed models that balance community engagement with commercial viability
  • Understand that prompting as a specialized skill will diminish, but deep knowledge of AI capabilities and limitations will become essential
  • Leverage industry maturation in fundraising by focusing on specific customer use cases rather than basic technology education
  • Maintain balanced competitive awareness without losing focus on core customer problems and solutions

Timestamp: [32:04-39:59]Youtube Icon

📚 References from [32:04-39:59]

People Mentioned:

  • Nick Frosst - Co-founder of Cohere, discussing AI policy, business models, and fundraising experiences

Companies & Products:

  • Cohere - Enterprise-focused AI company with hybrid open/closed model approach
  • Meta - Referenced regarding potential shift from open to closed AI model approach
  • RBC - Royal Bank of Canada, mentioned as Cohere customer implementation example
  • Fujitsu - Technology company partnering with Cohere for AI solutions
  • LG - Electronics company using Cohere's AI technology

Technologies & Tools:

  • Language Models - Core technology discussed throughout, including training methodologies and capability evolution
  • Foundational Models - AI models that serve as base for various applications, central to Cohere's business strategy

Concepts & Frameworks:

  • Open vs Closed AI Models - Business model spectrum from fully open-source to completely proprietary approaches
  • Prompt Engineering - Specialized skill for interacting with AI models, predicted to become less relevant
  • Income Inequality - Economic concern that AI deployment could exacerbate without proper policy frameworks
  • Model Sovereignty - Concept of countries maintaining control over their AI infrastructure and capabilities

Timestamp: [32:04-39:59]Youtube Icon

🏗️ How does Cohere train AI models efficiently on just two GPUs?

Training Strategy & Resource Efficiency

Core Training Philosophy:

  1. Two-GPU Architecture - All models (Command A, Command A reasoning, Command A vision) designed to fit on just two GPUs
  2. Strategic Sweet Spot - Balances performance, cost, and actual GPU availability for enterprise customers
  3. Orders of Magnitude Less Spending - Significantly lower foundational model training costs compared to competitors

Early Innovation Approach:

  • Scrappy Beginnings: Started with minimal funding by linking together scattered GPU resources across data centers
  • Published Research: Demonstrated feasibility of training models with fragmented compute resources
  • Practical Reality: While possible, it's much slower than renting dedicated data center capacity

Business Impact:

  • Customer Deployment: Many companies were bottlenecked on production deployment due to GPU scarcity
  • Competitive Advantage: Efficient training translates to efficient customer solutions
  • Resource Allocation: Compute spending has shifted over the years but remains a significant portion of funding

Timestamp: [40:05-42:09]Youtube Icon

💰 How does Cohere compete with OpenAI and Anthropic's billions?

Strategic Focus & Competitive Positioning

Singular Enterprise Focus:

  1. No Consumer Distractions - Unlike competitors, Cohere doesn't build consumer apps or chase $200/month personal subscriptions
  2. ROI Over AGI - Constantly emphasizes "ROI not AGI" to customers and stakeholders
  3. Production-First Mindset - Dedicated to getting enterprises successfully deployed with AI solutions

Different Model Requirements:

  • Enterprise vs Consumer Models: Enterprise needs are fundamentally different from consumer applications
  • No Image Generation: Workforce users don't typically need image creation capabilities, unlike consumer users who find it "very fun"
  • Deployment Constraints: Enterprise models must work within existing infrastructure limitations

Competitive Differentiation:

  • Smaller Funding Rounds: Acknowledges having smaller funding compared to competitors
  • Specialized Expertise: Focus on enterprise-specific challenges that larger competitors may not prioritize
  • Custom Solutions: Willingness to create bespoke models for specific customer needs

Timestamp: [42:09-44:49]Youtube Icon

🛠️ What role do forward deployed engineers play at Cohere?

Customer Success & Implementation Strategy

Forward Deployed Engineer Model:

  1. Crucial Component - Essential for getting companies up and running in production
  2. Customer-Specific Setup - Engineers help match technology to specific business configurations
  3. Value Delivery Focus - Ensure solutions actually deliver measurable value to customers

Pricing & Customization:

  • Customer-Dependent Pricing - Entirely based on what the customer wants to accomplish
  • Custom Model Creation - Some customers receive completely bespoke models
  • Flexible Engagement - Pricing varies significantly based on implementation complexity

Industry Perspective:

  • Not Poor Technology - Rejects notion that forward deployed engineers indicate inferior products
  • Enterprise Reality - Enterprise technology requires human engagement unlike consumer products
  • Business Matching - Technology must be adapted to existing business processes and infrastructure

Timestamp: [44:54-46:15]Youtube Icon

📈 Do enterprise companies trade at lower multiples than consumer companies?

Valuation Dynamics & Personal Perspective

Enterprise Investment Characteristics:

  1. Higher Revenue Quality - More stable and predictable than consumer revenue
  2. Greater Stickiness - Enterprise customers tend to stay longer once implemented
  3. Slower Growth - Working with large enterprises naturally limits growth velocity

Valuation Reality:

  • $6.8 Billion Valuation - Recent funding round achieved significant enterprise valuation
  • Potential Enterprise Discount - Acknowledges possibility of valuation penalty for slower growth
  • Staggering Numbers - Describes valuations as "impossible for an individual to conceive of"

Personal Money Philosophy:

  • Humble Origins - First job was working as a cook making burgers
  • Money Motivation - Believes "everybody cares about money" and is "motivated by money"
  • M&A Interest - Confirms receiving acquisition offers over the company's 5-year journey
  • Perspective on Wealth - Views current valuations as "well beyond the realm" of what regular people can engage with

Timestamp: [46:15-47:54]Youtube Icon

💎 Summary from [40:05-47:54]

Essential Insights:

  1. Efficiency Strategy - Cohere trains all models to fit on two GPUs, spending orders of magnitude less than competitors while maintaining performance
  2. Enterprise Focus - Singular dedication to enterprise customers with "ROI not AGI" philosophy, avoiding consumer distractions that competitors pursue
  3. Implementation Reality - Forward deployed engineers are crucial for enterprise success, helping match technology to specific business needs rather than indicating poor products

Actionable Insights:

  • Enterprise AI deployment requires human expertise and customization, not just plug-and-play solutions
  • Efficient model training can be a significant competitive advantage when targeting resource-constrained enterprise customers
  • Focused market positioning (enterprise-only) can compete effectively against larger, more diversified competitors

Timestamp: [40:05-47:54]Youtube Icon

📚 References from [40:05-47:54]

People Mentioned:

  • Palantir - Referenced for glamorizing the forward deployed engineer model in enterprise technology

Companies & Products:

  • OpenAI - Mentioned as a competitor with billions in funding, focusing on consumer applications
  • Anthropic - Referenced as another well-funded competitor in the foundational model space
  • Command A - Cohere's flagship model designed for two-GPU deployment
  • Command A Reasoning - Recently released reasoning-focused model variant
  • Command A Vision - Recently released vision-capable model variant
  • North - Cohere's agentic framework for enterprise knowledge workers

Technologies & Tools:

  • GPUs - Graphics Processing Units, the primary compute resource for AI model training and deployment
  • Forward Deployed Engineers - Customer-facing technical specialists who help implement enterprise solutions

Concepts & Frameworks:

  • ROI vs AGI - Cohere's philosophy prioritizing Return on Investment over Artificial General Intelligence
  • Enterprise Discount - Potential valuation penalty for companies focused on slower-growing enterprise markets
  • Two-GPU Architecture - Cohere's strategic decision to design models that fit within common enterprise hardware constraints

Timestamp: [40:05-47:54]Youtube Icon

🏗️ Why does Cohere's Nick Frosst want to build a generational company?

Building Something That Outlasts You

Nick Frosst explains that building a generational company is fundamentally about creating something bigger than yourself that will endure beyond your personal involvement.

Core Motivation:

  • Inherently Human Drive: The desire to build something lasting is fundamentally human - whether it's art, buildings, philosophies, or ideas
  • Bigger Than Yourself: The reward comes from participating in constructing something that transcends individual contribution
  • Time Scale Perspective: "Generational" refers to time scale generations, not just his own generation

The Reality Check:

  • Inevitable Impermanence: Acknowledges that eventually everything becomes "two feet in the desert" (referencing Ozymandias)
  • Dual Truth: Both the reward of building something lasting AND its ultimate impermanence can be true simultaneously
  • Worth the Effort: Despite knowing nothing lasts forever, the act of building something enduring remains rewarding and exciting

What This Means for Cohere:

  • Long-term Vision: Focus on creating technology and solutions that will serve future generations
  • Beyond Personal Legacy: Not about personal recognition but about contributing to something meaningful
  • Sustainable Impact: Building systems and innovations that continue to provide value long after the founders are gone

Timestamp: [48:02-53:09]Youtube Icon

💼 How does Nick Frosst view different types of work and their value?

Understanding Labor Value and Economic Reality

Nick Frosst provides a nuanced perspective on why different jobs are compensated differently, drawing from his own experience working various roles.

The Compensation Reality:

  • Skill Investment: Jobs requiring longer training periods and specialized skills typically command higher compensation
  • Agency Factor: Roles with more decision-making authority and strategic impact are valued higher
  • Market Dynamics: The economy generally compensates based on the investment required to develop skills and the scarcity of those abilities

Personal Experience Examples:

High-Stress Service Work:

  • Restaurant Experience: Working the grill during chaotic breakfast rushes
  • Physical Challenges: No air conditioning, running across the street for supplies
  • Genuine Difficulty: "Some of the hardest days I ever had at work"
  • Valuable but Different: Challenging and rewarding work, even at minimum wage

Current Impact Work:

  • Scale of Influence: Work affecting thousands of employees at major companies
  • Downstream Effects: Impact on millions of end users through enterprise clients
  • Definitively Higher Impact: More valuable due to broader reach and influence

Balanced Perspective:

  • Avoiding Extremes: Rejects both "invisible hand is perfectly accurate" and "all work is equally valuable" viewpoints
  • Middle Ground: Acknowledges that market compensation isn't perfect but generally reflects skill requirements and impact
  • Respect for All Work: Values the effort and skill in all types of labor while recognizing economic realities

Timestamp: [48:47-51:46]Youtube Icon

🤝 What disagreements have Nick Frosst and Aidan Gomez had at Cohere?

Co-founder Dynamics and Decision-Making

Nick Frosst reveals the nature of disagreements between Cohere's co-founders, emphasizing their collaborative relationship despite tactical differences.

Technical Disagreements:

  • API Design Philosophy: Brief disagreement about creating specific endpoints for different functions
  • Pre-RLHF Era: Debated whether to build separate endpoints for summarization and entity extraction
  • Low-Level Implementation: Focused on technical architecture rather than strategic vision

Day-to-Day Business Decisions:

  • Operational Policies: Regular discussions about business policies and procedures
  • Tactical Choices: Ongoing debates about specific business decisions and approaches
  • Normal Business Friction: Standard disagreements that occur in any leadership team

Overall Relationship Dynamic:

  • Mutual Respect: "I have a huge amount of respect for" both Aidan and Ivan (other co-founders)
  • No Major Conflicts: Haven't experienced any significant strategic disagreements
  • Collaborative Approach: Disagreements focus on execution details rather than fundamental direction
  • Healthy Debate: Regular arguing about "little things" while maintaining alignment on big picture

Leadership Structure:

  • Three Co-founders: Nick works closely with both Aidan Gomez and Ivan (third co-founder)
  • Privilege of Partnership: Views working with his co-founders as a privilege
  • Constructive Disagreement: Sees tactical disagreements as normal and healthy part of building the business

Timestamp: [53:15-53:48]Youtube Icon

📢 Why doesn't Cohere focus on public storytelling like other AI companies?

Enterprise vs Consumer Business Models

Nick Frosst explains why Cohere takes a different approach to public relations compared to consumer-focused AI companies.

Consumer vs Enterprise Distinction:

Consumer-Focused Companies:

  • Direct Consumer Sales: Companies like OpenAI and Anthropic make money through consumer subscriptions
  • Individual Decision Making: Success depends on individual consumers choosing their product
  • Public Story Importance: Consumer interest in the company narrative directly impacts revenue

Cohere's Enterprise Focus:

  • No Consumer Offering: "You can't spend $200 a month on Cohere as a person"
  • B2B Sales Model: Revenue comes from enterprise clients, not individual consumers
  • Different Stakeholders: Decision makers are businesses, not individual consumers influenced by public perception

Strategic Priorities:

  1. Product Development: Building better products is the most important focus
  2. Customer Problem Solving: Addressing enterprise client needs takes precedence
  3. Model Improvement: Creating better AI models for business applications
  4. Public Relations: Important but not the primary driver of business success

Balanced Approach:

  • Selective Engagement: Participates in interviews and public discussions when appropriate
  • Enjoys the Conversation: "I'm excited to come. I'm excited to talk to you. I like talking about Cohere"
  • Realistic Prioritization: Acknowledges storytelling importance while focusing resources on product and customers
  • Could Do Better: Recognizes room for improvement in public communication while maintaining current priorities

Timestamp: [54:51-55:58]Youtube Icon

💎 Summary from [48:02-55:58]

Essential Insights:

  1. Generational Company Vision - Building something that outlasts individual founders is fundamentally human and rewarding, even knowing it won't last forever
  2. Work Value Philosophy - Different jobs have different economic value based on skill requirements, training investment, and impact scale, but all work has inherent worth
  3. Co-founder Harmony - Successful partnerships involve tactical disagreements on implementation while maintaining strategic alignment and mutual respect

Actionable Insights:

  • Long-term Thinking: Focus on building sustainable value that transcends individual involvement
  • Realistic Compensation Understanding: Recognize that market forces generally reflect skill scarcity and impact, while avoiding extremes in valuation
  • Strategic Communication: Prioritize product development and customer needs over public relations when serving enterprise markets
  • Healthy Disagreement: Embrace tactical debates with co-founders while maintaining respect and shared vision
  • Business Model Alignment: Match communication strategy to target market - enterprise clients care more about product capability than founder narratives

Timestamp: [48:02-55:58]Youtube Icon

📚 References from [48:02-55:58]

People Mentioned:

  • Aidan Gomez - Cohere co-founder, frequently mentioned for his vision of building a generational company
  • Ivan - Third co-founder of Cohere, part of the leadership team Nick respects
  • Aravind Srinivas - CEO of Perplexity, discussed in context of AI company leadership and public presence
  • Sam Altman - OpenAI CEO, referenced as example of consumer-focused AI company leader
  • Dario Amodei - Anthropic CEO, mentioned as another consumer-focused AI company leader
  • Adam Smith - Referenced for his "invisible hand" economic theory about market value determination

Companies & Products:

  • Cohere - Nick's company, discussed throughout as enterprise-focused AI platform
  • OpenAI - Referenced as consumer-focused AI company with different business model
  • Anthropic - Mentioned as quasi-consumer company with significant enterprise API usage
  • Perplexity - AI search company led by Aravind, used as example of consumer-focused AI leadership
  • Cursor - AI coding tool mentioned as competing with Anthropic for developer users

Concepts & Frameworks:

  • Generational Company - Building organizations designed to outlast their founders across multiple time-scale generations
  • Invisible Hand Theory - Adam Smith's economic concept about market forces determining value and compensation
  • RLHF (Reinforcement Learning from Human Feedback) - AI training methodology referenced in context of API design decisions
  • Enterprise vs Consumer Business Models - Distinction between B2B and B2C approaches in AI company strategy

Literary References:

  • Ozymandias - Percy Bysshe Shelley poem about the impermanence of human achievements, referenced as "two feet in the desert"

Timestamp: [48:02-55:58]Youtube Icon

🌍 Should countries fund their own AI models for sovereignty?

Model Sovereignty and National Infrastructure

Why Countries Should Build Their Own Models:

  1. Language Infrastructure - Having a language model that speaks your country's language is like building infrastructure for your people
  2. Cultural Context - Models need to understand local dialect, cultural fluency, and context to truly empower citizens
  3. Economic Independence - Using models built by China or America might not set your country and economy up as well as having your own

The Silicon Valley Problem:

  • 20+ Years of Dominance - Technological history has been very defined by Silicon Valley and California
  • Growing Resentment - Many people are rightfully upset with American tech dominance
  • Political Influence - America has shown willingness to turn off tech access based on political reasons

Geopolitical Advantages:

  • Canadian Asset - Being Canadian helps Cohere in global discussions
  • Non-American Appeal - Companies worldwide are interested in working with non-American tech companies
  • Government Connections - The connection between American tech and government becomes less clear over time

Timestamp: [56:14-1:00:06]Youtube Icon

📱 Will phones still be our primary input device in 5 years?

The Future of Human-Computer Interaction

Language as the New Interface:

  • More Important Role - Language will become a more important part of how we interact with computers
  • Not Everything - Language isn't always the best interface; graphic user interfaces are still better for certain tasks
  • Better Integration - The goal is using language models to work with computers more effectively

Failed Attempts and Lessons:

  1. Rabbit R1 and Humane Pin - Recent attempts didn't get it right, but there's something promising about the concept
  2. Google Glass Experience - Initially exciting, but people immediately clocked it on public transport as intrusive
  3. VR Disappointment - Realized he doesn't want to strap a computer to his face or be more disengaged from the world

Core Philosophy:

  • World Engagement - Technology should connect us to the world better, not disconnect us
  • Immediate Connection - Values technology that creates immediate, human connections (like playing music)
  • Community Focus - People want to connect in the moment, not be removed from reality

Timestamp: [1:00:29-1:01:49]Youtube Icon

😟 What happened to Nick Frosst's technological optimism?

From Tech Optimist to Cautious Observer

The Transformation:

  • Past Enthusiasm - Used to be a real technological optimist who loved how technology was built
  • Current State - Wouldn't describe himself as a technological optimist over the past 10 years
  • Specific Concerns - Worries about dissolution of community and technology disconnecting people

Modern Tech Concerns:

  1. Social Issues - Worried about depression, loneliness, eating disorders, and materialism focus among young people
  2. Influencer Culture - The number one job young people want is to be an influencer
  3. Disconnection Problem - Technology removing people from real-world engagement rather than enhancing it

Historical Perspective:

  • Universal Worry Pattern - Acknowledges that every generation thinks "things used to be better" and "kids these days are weird"
  • Greek Philosophers - Even ancient Greeks worried about writing making people not use their memories
  • Newspaper Concerns - People once worried about newspapers making bus riders antisocial
  • Recurring Theme - This pattern of technological concern is "historically ubiquitous"

Personal Solution:

  • Music as Connection - Plays music because it's immediate and connects people in the moment
  • Community Seeking - People come to listen to music to connect, which fulfills a real human need

Timestamp: [57:33-1:03:58]Youtube Icon

💎 Summary from [56:03-1:03:58]

Essential Insights:

  1. Model Sovereignty is Infrastructure - Countries should fund their own AI models like they fund power plants, as language models are national infrastructure that need cultural fluency
  2. Geopolitical Tech Advantages - Being Canadian helps Cohere globally as companies seek alternatives to American tech due to political influence concerns
  3. Technology Should Connect, Not Isolate - The future of human-computer interaction should engage people more with the world, not remove them from it through VR or other disconnecting technologies

Actionable Insights:

  • Language will become more important in computer interfaces, but graphic interfaces still have their place for specific tasks
  • Countries benefit from having AI models that understand local dialect, culture, and context rather than relying on foreign-built systems
  • Historical perspective shows every generation worries about new technology, but current concerns about loneliness and disconnection deserve attention

Timestamp: [56:03-1:03:58]Youtube Icon

📚 References from [56:03-1:03:58]

People Mentioned:

  • Adam Smith - Referenced for his "invisible hand" economic theory in casual conversation
  • Donald Trump - Mentioned as influencing US tech companies and government connections

Companies & Products:

  • Mistral - Discussed as "the Europe play" for sovereign AI models, representing European AI independence
  • Intel - Referenced regarding US government taking a 10% stake, showing government-tech company connections
  • Google Glass - Used as example of initially exciting technology that failed due to social acceptance issues
  • Rabbit R1 - Mentioned as recent attempt at new computer interaction that "didn't get it right"
  • Humane Pin - Another recent failed attempt at reimagining human-computer interaction

Technologies & Tools:

  • VR (Virtual Reality) - Discussed as technology that disconnects people from the world rather than engaging them
  • Language Models - Positioned as infrastructure similar to power plants that countries should develop domestically

Concepts & Frameworks:

  • Model Sovereignty - The concept that countries should have their own AI models for cultural, economic, and political independence
  • Technological Optimism - Personal philosophy shift from embracing all technology to being more cautious about its social impacts
  • Human-Computer Interaction - Discussion of how interfaces should evolve to better connect people rather than isolate them

Timestamp: [56:03-1:03:58]Youtube Icon

🚨 Why has Sam Altman actually done a disservice to AI?

Nick Frosst's Critique of AGI Predictions

Key Issues with Altman's Approach:

  1. False Timeline Predictions - Made several predictions about AGI proximity that were "obviously wrong" at the time
  2. Existential Threat Messaging - Conducted a world tour telling major leaders AI poses existential threats in the near term
  3. Academic Dishonesty - Described Altman's approach as "academically disingenuous" and harmful to the technology

The Funding-Fear Correlation:

  • Balanced Messaging from Well-Funded Companies: Leaders like Demis Hassabis (DeepMind) and Mark Zuckerberg historically provided more measured perspectives when they didn't need external funding
  • Provocative Messaging for Funding: Companies requiring investment capital tend to make more dramatic claims to attract attention and dollars
  • Recent Shift in Tone: Even previously measured leaders like Zuckerberg and Hassabis have become more aggressive about AI's potential impacts

Legitimate vs. Illegitimate Concerns:

  • Real Transformative Potential: AI is genuinely transformative like personal computers, industrial revolution, steam engines, and printing press
  • Valid Discussion Topics: Legitimate concerns about labor patterns and workforce changes deserve attention
  • Misplaced Focus: Too much time spent on illegitimate concerns rather than real, practical implications

Timestamp: [1:05:10-1:07:30]Youtube Icon

🍟 What is Cohere's unique founder ritual after closing funding rounds?

McDonald's Tradition After Every Round

The Ritual Details:

  • Location: McDonald's restaurant
  • Participants: All three co-founders attend together
  • Timing: Immediately after signing each funding round
  • Consistency: Maintained across all 4-5 funding rounds completed

Nick's Personal Order:

  • Standard Meal: Two junior chickens every time
  • Origin Unknown: Can't remember how the tradition started
  • Unwavering Commitment: Despite raising hundreds of millions, they stick to this humble celebration

Why This Matters:

  • Grounding Ritual: Keeps founders connected to their roots despite massive valuations
  • Team Bonding: Shared experience that reinforces co-founder relationships
  • Authentic Celebration: Simple, unpretentious way to mark major milestones

Timestamp: [1:07:30-1:08:09]Youtube Icon

⚖️ What is the worst regulatory mistake governments could make with AI?

The Digital Gods Misunderstanding

Core Regulatory Risk:

  • Fundamental Misunderstanding: Regulators thinking large language models are "digital gods" rather than sophisticated but limited tools
  • Benchmark Fixation: Creating regulations based on arbitrary benchmarks that can be gamed or manipulated
  • AGI Shutdown Policies: Implementing rules to halt development based on misleading benchmark performance

Problems with Benchmark-Based Regulation:

  1. Gaming Potential: Models can be specifically trained to perform better or worse on particular benchmarks
  2. Limited Representation: Single benchmarks don't capture real-world usage patterns and risks
  3. Innovation Stifling: Could shut down beneficial development based on flawed metrics

Better Regulatory Approach:

  • Focus on Actual Use Cases: Understand how technology is actually used and misused
  • Avoid Existential Threat Fixation: Don't build policy around speculative doomsday scenarios
  • Practical Impact Assessment: Evaluate real-world applications rather than theoretical capabilities

Timestamp: [1:08:09-1:09:10]Youtube Icon

🇨🇳 Will China produce leading AI models that beat US models?

Assessment of Chinese AI Capabilities

Current State Analysis:

  • Good But Not Leading: China has produced good models but hasn't yet created models that definitively beat top US models
  • Rapid Development: Recently saw seven new Chinese model providers release models in about a week
  • Quality Recognition: Acknowledges the models were "pretty good" and impressive in their rapid deployment

Competitive Outlook:

  • Continued Development: Expects China will keep building models that will be useful
  • Specialized Strengths: Chinese models will likely excel in areas they specifically train for
  • Not a Major Concern: Doesn't express worry about Chinese AI capabilities as a competitive threat

Strategic Perspective:

  • Natural Competition: Views Chinese AI development as expected market competition rather than existential threat
  • Focused Excellence: Anticipates Chinese models will be particularly strong in their areas of specialization
  • Measured Response: Takes a balanced view without dismissing capabilities or expressing alarm

Timestamp: [1:09:10-1:10:06]Youtube Icon

🔮 What is Nick Frosst's boldest prediction for LLMs in 2026?

Automated Expense Filing Revolution

The Prediction:

  • Simple Command Interface: Users will be able to say "file my expenses" to their computer
  • Complete Automation: The model will automatically understand expense policies, locate photos, and handle all processing
  • Universal Application: This capability will work across different applications and systems

Why This Seems Bold:

  1. Deceptively Simple: Sounds straightforward but requires complex integration and reliability
  2. Not Yet Ubiquitous: Most people don't currently have this seamless experience
  3. Reliability Requirement: Must work consistently enough for business-critical processes

Technical Challenges:

  • Policy Understanding: AI must interpret complex, company-specific expense policies
  • Multi-System Integration: Needs to work across various applications and file systems
  • Error-Free Processing: Business processes require high accuracy and reliability standards

Broader Implications:

  • Workflow Transformation: Represents fundamental shift in how people interact with computers
  • Administrative Revolution: Could eliminate entire categories of routine business tasks
  • User Experience Evolution: Natural language becomes primary interface for complex operations

Timestamp: [1:10:06-1:10:54]Youtube Icon

💻 What AI tools has Nick Frosst added to improve his productivity?

Personal AI Workflow Integration

Core Productivity Tools:

  1. Cohere's North: Uses their own product for various tasks
  2. Cursor: Primary coding application that significantly improves development workflow
  3. Whisper: For transcription and audio processing tasks

Team Approach to AI Tools:

  • No Mandates: Doesn't require the entire team to use specific tools like Cursor
  • Individual Choice: Engineers can choose their preferred AI-enhanced development tools
  • Widespread Adoption: Many team members rely on various AI models for their work

Cost Sensitivity Analysis:

  • Personal Use: Not price-sensitive to tools like Cursor for individual use
  • Business Scale: Would become cost-conscious if prices increased 10x across 100+ engineers
  • Alternative Solutions: Could potentially build internal tools using Cohere's own models if costs became prohibitive

Quality vs. Cost Trade-offs:

  • Current Reality: Acknowledges Cursor has built superior UX and product quality
  • Future Considerations: Would only consider alternatives if costs increased dramatically (1000x)
  • Build vs. Buy Decision: Maintains option to develop internal solutions if market pricing becomes unreasonable

Timestamp: [1:11:00-1:12:25]Youtube Icon

🤔 What trait has both helped and hindered Nick Frosst's success?

The Curious and Contrarian Mindset

The Double-Edged Asset:

  • Core Trait: Being simultaneously curious and contrarian in thinking and approach
  • Success Factor: Enables breakthrough insights when conventional wisdom is wrong
  • Risk Factor: Can lead to being wrong when the majority view is actually correct

When It Works Well:

  1. Early AI Adoption: Recognized language model potential in 2019 when few others did
  2. Market Timing: Joined Aidan Gomez to found Cohere when the view wasn't widespread
  3. Learning Advantage: Deep curiosity drives thorough understanding of complex topics

When It Backfires:

  • Technological Optimism: Early belief that all human progress metrics would continue improving monotonically
  • RLHF Skepticism: Initially doubted the data efficiency of reinforcement learning from human feedback in 2020
  • Contrarian Stubbornness: Sometimes maintains opposing views even when evidence suggests otherwise

Specific Examples of Being Wrong:

  1. Human Progress Assumptions: Believed life expectancy, income inequality, and happiness would all improve continuously
  2. Technical Misjudgments: Underestimated how small datasets of human feedback could improve model performance
  3. Monotonic Progress Fallacy: Failed to account for setbacks and reversals in human advancement

The Learning Process:

  • Balanced Perspective: Now holds both optimistic and pessimistic views simultaneously
  • Humility Development: Recognizes the importance of questioning his own contrarian instincts
  • Continuous Calibration: Uses past mistakes to better evaluate when to trust or doubt conventional wisdom

Timestamp: [1:12:34-1:14:16]Youtube Icon

💎 Summary from [1:04:03-1:14:23]

Essential Insights:

  1. Sam Altman's AI Messaging - Frosst criticizes Altman's "academically disingenuous" predictions about AGI timelines and existential threats, arguing it harms the technology
  2. Regulatory Risks - The worst regulatory mistake would be treating LLMs as "digital gods" and creating shutdown policies based on gameable benchmarks
  3. Founder Authenticity - Cohere maintains humble McDonald's celebrations after every funding round, demonstrating grounded leadership despite billion-dollar valuations

Actionable Insights:

  • AI Development Focus: Concentrate on legitimate technological concerns rather than speculative existential threats
  • Productivity Integration: Tools like Cursor and Whisper can meaningfully improve workflow, but cost-benefit analysis matters at scale
  • Leadership Traits: Curiosity and contrarian thinking drive innovation but require humility to recognize when conventional wisdom is correct

Timestamp: [1:04:03-1:14:23]Youtube Icon

📚 References from [1:04:03-1:14:23]

People Mentioned:

  • Sam Altman - OpenAI CEO criticized for making false AGI timeline predictions and conducting world tour about existential AI threats
  • Demis Hassabis - DeepMind CEO noted for historically balanced AI messaging when not needing external funding
  • Mark Zuckerberg - Meta CEO whose AI rhetoric has become more aggressive recently, causing concern about real changes coming
  • Baz Luhrmann - Director referenced for "Wear Sunscreen" commencement speech about generational perspectives
  • Jordan - Person who revealed Cohere's McDonald's funding celebration ritual
  • Aidan Gomez - Cohere co-founder who invited Frosst to start the company in 2019

Companies & Products:

  • Google DeepMind - AI research company Frosst would bet his career on if not at Cohere
  • Cursor - AI-powered coding application that significantly improves development productivity
  • Whisper - OpenAI's transcription tool used in Frosst's workflow
  • North - Cohere's own product used internally for productivity
  • Windsurf - Alternative AI coding tool mentioned as possibly used by some team members
  • Devon - Another AI development tool referenced in the discussion

Technologies & Tools:

  • VS Code - Microsoft's code editor that could potentially be extended with Cohere's models
  • RLHF (Reinforcement Learning from Human Feedback) - Technique Frosst initially doubted the data efficiency of in 2020

Concepts & Frameworks:

  • Model Sovereignty - Concept of countries funding their own AI models for independence
  • AGI Benchmarks - Metrics used to measure artificial general intelligence progress, which can be gamed
  • Technological Optimism - Frosst's early belief that human progress metrics would improve monotonically

Timestamp: [1:04:03-1:14:23]Youtube Icon