undefined - Synthetic Data and the Future of AI | Cohere CEO Aidan Gomez

Synthetic Data and the Future of AI | Cohere CEO Aidan Gomez

How do companies like Salesforce and Dell scale intelligence across every cloud? Aidan Gomez, co-founder and CEO of Cohere, explains how theyโ€™re building AI that works across all enterprise systems and deploys anywhere, giving companies true flexibility and security. The co-author behind the revolutionary "Attention Is All You Need" paper joins Joubin Mirzadegan for a wide-ranging conversation on why synthetic data went from dismissed to indispensable, and how the race among AI labs is really unfolding. Guest: Aidan Gomez, co-founder and CEO of Cohere

โ€ขNovember 17, 2025โ€ข71:40

Table of Contents

0:00-7:57
8:03-15:59
16:05-23:58
24:04-31:55
32:02-39:55
40:01-47:52
48:00-55:58
56:07-1:03:56
1:04:03-1:11:34

๐Ÿš€ How did Google's Attention Is All You Need paper transform AI?

The Revolutionary Transformer Architecture

The "Attention Is All You Need" paper fundamentally changed AI by introducing the Transformer architecture, which became the foundation for modern large language models like GPT and beyond.

Key Innovation - Efficiency at Scale:

  1. Parallel Processing Power - The Transformer was extremely well-suited to scaling across many GPUs, unlike previous sequential models
  2. Scalability Advantage - As the industry moved toward larger models, Transformers scaled the best and dominated the field
  3. Perfect Timing - The architecture emerged just as everyone started scaling up models, making it the ideal solution

The Collaborative Discovery:

  • Team Effort: Eight co-authors at Google Brain worked together on this breakthrough
  • Intern Contribution: Aidan Gomez was just 19 years old and still an undergraduate when he co-authored this influential paper
  • Building on Existing Ideas: The paper pulled together concepts from BitNet, WaveNet, and sequence-to-sequence models that were "in the ether"

Industry Impact:

  • Open Source Decision: Google's choice to publish rather than keep it proprietary enabled widespread adoption
  • Inevitable Innovation: The ideas were converging naturally - if Google hadn't published it, someone else would have created something similar within 12-18 months
  • Catalytic Effect: Once published, the community momentum took over and figured out the rest of the applications

Timestamp: [0:31-0:58]Youtube Icon

๐ŸŽฏ What is Google's current position in the AI race?

Google's AI Comeback Story

Despite initial perceptions that Google "missed AI," they've made a remarkable comeback and are now potentially leading the field in several key areas.

Google's Competitive Advantages:

  1. Superior Models - Building potentially the best models that may have surpassed OpenAI
  2. Financial Engine - A "money printing machine" (search/ads business) that can fuel massive AI investments
  3. Data Advantage - A "data printing machine" providing continuous training material
  4. Talent Concentration - "Surreal" concentration of top AI talent and expertise

Current Status Assessment:

  • Technology Leadership: They've caught up and potentially exceeded competitors on a technological basis
  • The Product Challenge: The remaining question mark is whether they can compete effectively from a product perspective
  • Market Position: Strong technical foundation but execution in consumer-facing AI products remains to be proven

Strategic Implications:

  • Resource Advantage: Unique combination of talent, data, and capital creates sustainable competitive moats
  • Innovation Capacity: Deep technical capabilities position them well for continued breakthroughs
  • Execution Risk: Success will depend on translating technical excellence into market-winning products

Timestamp: [0:00-0:25]Youtube Icon

๐Ÿ’ผ How does Cohere CEO Aidan Gomez balance travel with personal life?

Strategic Approach to Executive Travel

As CEO of a $6.8 billion AI company, Aidan Gomez has developed a unique approach to managing extensive travel demands while maintaining personal relationships.

The Travel Reality:

  • Sales-Focused Role: CEO position at this stage is fundamentally a sales job requiring constant travel
  • Industry Misconception: People glorify frequent business travel, but the reality is exhausting and isolating
  • Personal Impact: "I wouldn't wish this on my worst enemy" - the toll of weekly travel

Innovative Solution - Bringing Family:

  1. Menu Approach: Present wife with upcoming trip options and let her choose which ones to join
  2. Dinner Integration: Wife attends all business dinners, not just daytime meetings
  3. Client Appreciation: Business partners actually prefer and enjoy meeting his wife
  4. Relationship Maintenance: Transforms isolating business trips into shared experiences

Practical Implementation:

  • Selective Participation: Wife joins dinners but not daytime meetings
  • Client Comfort: Partners are "stoked" to meet his wife and often prefer talking to her
  • Work-Life Integration: Rather than separation, creates meaningful integration of personal and professional spheres

Timestamp: [1:46-4:48]Youtube Icon

๐Ÿง  What was Aidan Gomez's role in creating the Transformer paper at age 19?

The Youngest Co-Author of AI's Most Important Paper

At just 19 years old and still an undergraduate, Aidan Gomez became one of eight co-authors of "Attention Is All You Need," the paper that launched the modern AI revolution.

The Remarkable Context:

  • Age and Experience: Second or third-year undergraduate student working as an intern at Google Brain
  • Equal Contribution: Listed as co-author alongside senior researchers, not just a minor contributor
  • Historical Significance: The paper became "one of the most consequential papers in our industry"

The Development Process:

  1. Collaborative Effort: Eight-person team at Google Brain working together
  2. Building on Foundations: Drew from existing concepts like BitNet, WaveNet, and sequence-to-sequence models
  3. Perfect Timing: Ideas were "in the ether" and ready to be synthesized into something revolutionary

The Bigger Picture:

  • Inevitable Innovation: If this team hadn't created it, someone else would have within 12-18 months
  • Community Momentum: Once published, the broader research community took the ideas and ran with them
  • Catalytic Effect: The paper served as a "seed" that started a "snowball" rolling down the hill of AI advancement

Timestamp: [4:48-7:57]Youtube Icon

๐Ÿ’Ž Summary from [0:00-7:57]

Essential Insights:

  1. Google's AI Resurgence - Despite early perceptions of missing AI, Google has potentially surpassed OpenAI with superior models, backed by their data advantage, financial resources, and talent concentration
  2. Transformer Architecture Impact - The "Attention Is All You Need" paper succeeded because of its efficiency and scalability across GPUs, perfectly timed for the industry's move toward larger models
  3. Executive Travel Innovation - Cohere's CEO has solved the isolation of business travel by bringing his wife to dinner meetings, creating better client relationships and work-life integration

Actionable Insights:

  • Open Source Strategy: Google's decision to publish the Transformer paper rather than keep it proprietary enabled widespread adoption and industry transformation
  • Inevitable Innovation Principle: Revolutionary ideas often emerge from multiple sources simultaneously - if one team doesn't execute, another will within 12-18 months
  • Personal Integration in Business: Including family members in appropriate business settings can enhance rather than hinder professional relationships

Timestamp: [0:00-7:57]Youtube Icon

๐Ÿ“š References from [0:00-7:57]

People Mentioned:

  • Aidan Gomez - Co-founder and CEO of Cohere, co-author of "Attention Is All You Need" paper
  • Joubin Mirzadegan - Partner at Kleiner Perkins, host of the Grit podcast

Companies & Products:

  • Google Brain - Google's AI research division where the Transformer paper was developed
  • Cohere - Enterprise-first LLM infrastructure company valued at $6.8 billion
  • OpenAI - AI company that Google may have surpassed according to the discussion
  • Kleiner Perkins - Venture capital firm where the host is a partner

Research Papers & Publications:

  • "Attention Is All You Need" - The foundational Transformer paper co-authored by Aidan Gomez and seven others at Google Brain
  • BitNet - Earlier research that contributed ideas to the Transformer architecture
  • WaveNet - Google's neural network architecture that influenced the Transformer development
  • Sequence-to-Sequence Models - Earlier models that provided foundational concepts for the Transformer

Concepts & Frameworks:

  • Transformer Architecture - The neural network architecture that became the foundation for modern large language models
  • Auto-regressive Models - Models that were more scalable for training and influenced the Transformer design
  • GPU Scaling - The ability to efficiently distribute computation across multiple graphics processing units

Timestamp: [0:00-7:57]Youtube Icon

๐Ÿ”ฌ What was the core insight in the Transformer paper that changed AI?

The Revolutionary Focus on Efficiency and Scalability

The breakthrough wasn't just about creating another neural network architecture - it was about building something that could scale efficiently across multiple GPUs when "many GPUs" meant just tens of them.

Key Innovation Elements:

  1. Minimal Architecture Design - The Transformer was intentionally simple and streamlined, avoiding unnecessary complexity
  2. GPU Scaling Optimization - Built specifically to work efficiently across multiple processors (32 GPUs was the ambitious target)
  3. Future-Proof Framework - Designed with the assumption that scaling would be essential for machine learning progress

The Scaling Evolution:

  • 2017: "Many GPUs" = tens of GPUs
  • Today: Tens of thousands, maybe hundreds of thousands of GPUs
  • Near Future: Potentially millions of GPUs

Why This Mattered:

The architecture that scaled best would dominate the field. The Transformer's efficiency advantage meant it could leverage increasing computational resources better than competing approaches, making it the foundation for today's AI revolution.

Timestamp: [8:28-9:39]Youtube Icon

๐Ÿ“ How long did it take to create the Transformer paper?

A Compressed 4-Month Sprint That Changed Everything

The entire Transformer project happened in an incredibly short timeframe - just 12 to 16 weeks, roughly 4 months total in 2017. This compressed timeline created intense, memorable moments during the development process.

Memorable Development Moments:

  1. Submission Night Drama - Sleeping at the Google office the night they submitted to NeurIPS conference
  2. Unexpected Wake-Up Call - A cleaner accidentally hit him in the head with a door while he slept in a phone booth
  3. The Prediction Moment - Lying on a couch next to first author Ashish Vaswani after submission

The Realization:

  • Ashish's Prediction: "This is going to be a big deal"
  • Intern's Perspective: "Aren't all papers like this?"
  • Expected Impact: Hundreds of citations
  • Actual Impact: Hundreds of thousands of citations and catalyst for an industrial revolution

The rapid development timeline demonstrates how breakthrough innovations can emerge from focused, intensive collaboration periods.

Timestamp: [10:10-11:54]Youtube Icon

๐Ÿ“ˆ Are AI scaling laws still working or hitting limits?

The Surprising Plateau and Strategic Pivot

Even the experts were caught off guard by how long AI models continued to improve with more compute and data. However, recent developments suggest we're reaching the limits of pure scaling approaches.

The Scaling Surprise:

  • Academic Skepticism: Papers were being written showing models should get worse after initial improvements
  • Contrarian Bet: OpenAI's belief in "bigger is better forever" was highly controversial
  • Extended Success: Scaling worked much longer than most experts predicted

Current Reality Check:

  1. GPT-5 Strategy Shift - Rumored to be smaller than previous models, not larger
  2. Failed Scaling Attempts - The larger Orion series didn't deliver material improvements
  3. New Focus Areas - Better data quality, improved training methods, novel approaches

The Economic Question:

  • Massive Investment: Tens of billions being spent on model training
  • Diminishing Returns: Doubling or 10x-ing spending may not be economically justified
  • Evidence Gap: No strong proof that massive spending increases are still worthwhile

The field is transitioning from pure computational scaling to more sophisticated optimization strategies.

Timestamp: [12:01-15:59]Youtube Icon

๐Ÿ’Ž Summary from [8:03-15:59]

Essential Insights:

  1. Transformer's Core Innovation - The focus on efficiency and GPU scalability, not just performance, made it the dominant architecture as the field scaled up
  2. Rapid Development Timeline - The revolutionary paper was created in just 4 months, with memorable moments of uncertainty about its true impact
  3. Scaling Laws Plateau - The "bigger is better" approach that surprised everyone with its longevity is now hitting economic and performance limits

Actionable Insights:

  • Design for scalability from the beginning - the Transformer's efficiency advantage became crucial as compute resources expanded
  • Even breakthrough innovations can emerge from short, focused development periods with the right team and approach
  • The AI field is pivoting from pure scaling to better data and training methods as traditional approaches reach diminishing returns

Timestamp: [8:03-15:59]Youtube Icon

๐Ÿ“š References from [8:03-15:59]

People Mentioned:

  • Ashish Vaswani - First author of the Transformer paper who predicted its significance during the submission night

Companies & Products:

  • Google - Where the Transformer research was conducted and the paper was developed
  • OpenAI - Made the contrarian bet on scaling that proved successful longer than expected
  • Cohere - Aidan Gomez's company, started during the period when scaling laws were still being debated
  • Anthropic - Mentioned as one of the major AI labs working on next-generation models

Technologies & Tools:

  • Transformer Architecture - The neural network architecture that became the foundation for modern AI systems
  • GPT-5 - OpenAI's rumored next model that may be smaller than previous versions
  • Orion Series - OpenAI's larger experimental models that reportedly didn't deliver expected improvements

Concepts & Frameworks:

  • Scaling Laws - The principle that AI models improve predictably with more compute, data, and parameters
  • NeurIPS Conference - The venue where the original Transformer paper was published
  • GPU Scaling - The ability to efficiently distribute training across multiple graphics processing units

Timestamp: [8:03-15:59]Youtube Icon

๐Ÿ”ฌ Why are AI labs shifting focus from consumer chatbots to scientific breakthroughs?

The Economics of AI Development

The landscape of AI development is experiencing a fundamental shift as companies grapple with diminishing returns on massive investments.

Current Market Saturation:

  • Consumer Impact Plateau: Users can't feel material differences between model generations that justify 10x price increases
  • Slowing Progress: Rate of advancement has considerably decreased despite continued high spending on compute and scale
  • Economic Territory: Consumer chatbot improvements are entering uneconomic territory for pricing

Strategic Pivot to High-Value Applications:

  1. Enterprise Solutions - Focus on business applications with higher willingness to pay
  2. Scientific Breakthroughs - Targeting research applications like cancer cures and millennium problems
  3. Government Partnerships - Pursuing contracts with entities willing to pay premium prices for breakthrough capabilities

The Value Proposition Shift:

  • Consumer Market: Limited pricing power due to marginal perceived improvements
  • Scientific Applications: Potentially unlimited value for breakthrough discoveries
  • Strategic Refocusing: Labs are repositioning around "LLMs for science" to drive innovation forward

Timestamp: [16:05-17:44]Youtube Icon

๐Ÿ‘จโ€๐Ÿ’ผ How has Cohere CEO Aidan Gomez transitioned from researcher to business leader?

From Academic Research to Global Deployment

The transformation from technical researcher to CEO represents a complete career evolution with both challenges and unique opportunities.

The Technical Transition:

  • Self-Assessment: Admits to being "non-technical" and "cooked" compared to his research days
  • Reading Habits: Reduced from consuming research "religiously" to reading papers once or twice per quarter
  • Team Dynamics: Acknowledges being "more annoying than helpful" to the modeling team with constant ideas

New Responsibilities and Privileges:

  1. Policy Influence - Helping set industry and government policy
  2. Global Deployment - Seeing technology implementation at the front lines
  3. Strategic Networking - Meeting influential people and visiting important locations

The Research Distance Problem:

  • Previous Isolation: As a researcher in Oxford, Mountain View, and Toronto, was "so far from anyone actually using the stuff"
  • Current Connection: Now directly involved in deploying technology into the global economy
  • Perspective Shift: Different side of the field with real-world impact visibility

Continued Technical Curiosity:

  • Efficiency Focus: Ideas about multihop prediction, low-rank training, and pre-training optimization
  • Data Curriculum: Concepts about training progression from easy to hard, noisy to clean data
  • Product-Level Modeling: Still contributes to meta and product-level modeling questions

Timestamp: [17:44-20:26]Youtube Icon

๐Ÿง  What breakthrough capability do reasoning models bring to AI systems?

Variable Compute for Variable Complexity

Reasoning models represent a fundamental shift in how AI systems approach problem-solving, introducing human-like adaptability in computational effort.

The Core Innovation:

  • Adaptive Computation: Models can now spend different amounts of compute based on problem complexity
  • Human-Like Approach: Mirrors how humans spend varying time on different problems - seconds for simple questions, decades for complex theorems

Previously Impossible Problem Categories:

  1. Complicated Mathematics - Complex calculations requiring step-by-step reasoning
  2. Code Debugging - Understanding why attempts failed and implementing fixes
  3. Enterprise Tool Use - Navigating failures and rerouting through complex business systems

The Fundamental Problem with Pre-Reasoning Models:

  • Fixed Compute Allocation: Same computational effort for "What's 1+1?" and "How do we cure cancer?"
  • Massive Complexity Variance: Input space ranges from trivial to millennium problems
  • Inefficient Resource Usage: No ability to match effort to problem difficulty

Why Reasoning Was Obvious (In Retrospect):

  • Surprised the Wrong People: Those "one step removed from language models" were caught off-guard
  • Natural Evolution: Logical next step for anyone deeply familiar with LLM limitations
  • Universal Input Space: Language models can receive any question, requiring variable response complexity

Timestamp: [20:26-22:34]Youtube Icon

๐ŸŽฏ What critical intelligence capability are current AI models missing?

Learning from Experience: The Next Frontier

Current AI systems lack a fundamental aspect of human intelligence that could unlock unprecedented capabilities in long-term collaboration and skill development.

The Experience Problem:

  • Reset Limitation: Pressing "new chat" returns the model to its original state, losing all interaction history
  • No Growth: A month of collaboration provides zero learning benefit for future interactions
  • Static Intelligence: Models remain at the same competency level regardless of usage

Human Learning Analogy:

  • Founder's Journey: Building Cohere involved making "every mistake you could make" but learning from each one
  • Iterative Improvement: Humans become "massively more competent" over time through experience
  • Mistake Avoidance: Good experiences teach us not to repeat the same errors

The Intelligence Gap:

  • Obvious Missing Trait: Learning from experience is a fundamental aspect of intelligence
  • Product-Level Impact: This capability would unlock entirely new categories of AI applications
  • Collaborative Potential: Long-term AI partnerships could become genuinely productive

Technical Components:

  1. Memory Systems - Retaining and accessing past interactions
  2. Skill Distillation - Extracting learnable patterns from experiences
  3. Experience Integration - Converting memories into improved capabilities

Strategic Priority:

  • Next Major Breakthrough: Positioned as the logical successor to reasoning models
  • Team Focus: Active area of development and research priority
  • Product Transformation: Could fundamentally change how humans interact with AI systems

Timestamp: [22:34-23:58]Youtube Icon

๐Ÿ’Ž Summary from [16:05-23:58]

Essential Insights:

  1. Market Saturation Reality - AI labs are hitting economic limits with consumer applications as users can't justify paying 10x more for marginal improvements
  2. Strategic Pivot to Science - Companies are refocusing on high-value applications like scientific breakthroughs where governments and large entities will pay premium prices
  3. CEO Transformation Challenge - Technical founders face the difficult transition from hands-on research to business leadership, often feeling disconnected from their technical roots

Actionable Insights:

  • Enterprise Focus: The real value in AI lies in enterprise and scientific applications, not consumer chatbots
  • Reasoning Models Breakthrough: Variable compute allocation based on problem complexity unlocks previously impossible problem categories
  • Next AI Frontier: Learning from experience and memory systems represent the next major capability gap to solve

Timestamp: [16:05-23:58]Youtube Icon

๐Ÿ“š References from [16:05-23:58]

People Mentioned:

  • Aidan Gomez - Co-founder and CEO of Cohere, discussing his transition from researcher to business leader

Companies & Products:

  • Cohere - AI company founded by Aidan Gomez, focusing on enterprise language models
  • Oxford University - Referenced as one of Gomez's previous research locations
  • Google (Mountain View) - Another location where Gomez conducted AI research

Technologies & Tools:

  • Language Models (LLMs) - Core technology discussed throughout the conversation
  • Reasoning Models - New category of AI models that can allocate variable compute based on problem complexity
  • Multihop Prediction - Technical approach mentioned for improving model efficiency
  • Low-rank Training - Method for getting better efficiency from training compute

Concepts & Frameworks:

  • Data Curriculums - Training methodology progressing from easy to hard content, noisy to clean data
  • Variable Compute Allocation - Concept of matching computational effort to problem complexity
  • Learning from Experience - Missing AI capability that allows improvement through interaction history
  • Memory Systems - Technical component needed for AI models to retain and learn from past interactions

Timestamp: [16:05-23:58]Youtube Icon

๐Ÿš€ What is Cohere CEO Aidan Gomez's prediction for GPT-6's breakthrough capability?

AI Models Learning and Evolving Over Time

Aidan Gomez believes the next major advancement in language models will be the ability to continuously learn and improve without requiring new model training. This represents a fundamental shift from static models to dynamic, evolving systems.

The Intern-to-Expert Analogy:

  1. Initial State: AI models arrive like fresh university graduates with general knowledge but no specific business context
  2. Learning Phase: Through continuous interaction, they acquire domain-specific knowledge about your business, products, and customers
  3. Maturation: Over time, they become increasingly valuable assets with higher ROI as they accumulate experience

Key Implications:

  • For Consumers: AI assistants that become smarter and more personalized through use
  • For Enterprise: Systems that develop deep institutional knowledge without manual retraining
  • For ROI: The longer you use the system, the more valuable it becomes to your specific needs

Current Limitations:

  • Models today require complete retraining to incorporate new knowledge
  • Each deployment starts from scratch without accumulated learning
  • Business-specific expertise must be manually programmed or fine-tuned

Timestamp: [24:39-26:09]Youtube Icon

๐Ÿ”„ How did synthetic data go from being dismissed to dominating AI training?

The Complete Reversal on AI Self-Generated Training Data

The AI community has undergone a dramatic shift in perspective on synthetic data, moving from complete skepticism to widespread adoption across all major AI labs.

The Initial Skepticism:

  • "Ouroboros Effect": Critics compared it to a snake eating its own tail
  • "Human Centipede of Data": Concerns about degradation through recursive self-training
  • Fundamental Limitation: Belief that models couldn't make themselves smarter using their own outputs

The Reality Today:

  1. Universal Adoption: Every major AI lab now uses synthetic data as the majority of their training material
  2. Performance Benefits: Models can reformat and structure their own data for more effective learning
  3. Quality Filtering: AI systems excel at identifying and prioritizing higher-quality training examples

Why the Transformation Worked:

  • Data Reformatting: Models optimize data presentation for their own learning processes
  • Quality Enhancement: Synthetic generation allows filtering for better training examples
  • Structural Improvements: AI can organize information in ways that facilitate more effective learning

Current Applications:

  • Training data augmentation and enhancement
  • Domain-specific dataset creation
  • Quality control and filtering of existing datasets

Timestamp: [27:24-28:18]Youtube Icon

๐ŸŽญ Why does Cohere CEO Aidan Gomez call AI self-awareness claims "BS"?

Skepticism Toward AI Consciousness and Self-Awareness

Aidan Gomez expresses strong skepticism about recent claims of self-awareness in AI models, viewing them as anthropomorphization rather than genuine technological breakthroughs.

His Direct Assessment:

  • Clear Position: Calls self-awareness claims "BS" without hesitation
  • Source Attribution: Links these claims to the "effective altruist crowd"
  • Personification Problem: Believes people are incorrectly attributing human-like consciousness to AI systems

The Pattern He Observes:

  1. Anthropomorphization: Tendency to assign human characteristics to AI systems
  2. Misinterpretation: Confusing sophisticated pattern matching with genuine self-awareness
  3. Hype Cycle: Part of broader pattern of exaggerated claims in AI development

Technical Reality vs. Claims:

  • Current Capabilities: Models demonstrate sophisticated responses but not genuine consciousness
  • Pattern Recognition: Advanced text generation can mimic self-reflective behavior
  • No Evidence: Lack of concrete proof for actual self-awareness or consciousness

Broader Implications:

  • Skepticism extends to other grandiose AI claims
  • Focus should remain on practical capabilities rather than consciousness speculation
  • Important to distinguish between impressive performance and actual sentience

Timestamp: [26:52-27:18]Youtube Icon

๐ŸŽฏ What does Aidan Gomez say about the "ladder pulling" strategy in AI?

Competitive Tactics and Market Manipulation in AI Development

Aidan Gomez provides a scathing critique of how leading AI companies have used fear-based messaging and regulatory capture to maintain competitive advantages and discourage new entrants.

The Strategy Components:

  1. Existential Threat Messaging: Claims that AI models pose world-ending risks
  2. Resource Intimidation: Exaggerating the money and energy requirements for competitive AI development
  3. Regulatory Capture: Lobbying for restrictions that only apply to competitors

Historical Claims vs. Reality:

  • Past Predictions: Claims that GPT-3 "might literally kill the world"
  • Exponential Takeoff Theory: Predictions that the first to achieve AGI would "take the whole cake"
  • Current Reality: 5-7 models have converged to similar capabilities and are largely interchangeable

The Effectiveness:

  • Investor Deterrence: Successfully scared off potential investors from funding competitors
  • Regulatory Impact: Influenced policymakers to consider restrictive regulations
  • Market Psychology: Created artificial barriers to entry through fear rather than technical limitations

Intellectual Dishonesty:

  • False Extrapolations: Making unreasonable predictions about technology development
  • Goalpost Moving: Continuously shifting definitions and timelines when predictions fail
  • Conviction vs. Evidence: Presenting speculation with unwarranted certainty

Real-World Outcomes:

  • Multiple companies successfully developed competitive models
  • No evidence of the predicted exponential advantages
  • Market remains competitive despite fear-based predictions

Timestamp: [29:00-31:55]Youtube Icon

๐Ÿ’Ž Summary from [24:04-31:55]

Essential Insights:

  1. Next-Gen AI Capability - GPT-6 and similar models will likely feature continuous learning, allowing them to improve over time without retraining, similar to how human employees gain experience
  2. Synthetic Data Revolution - The AI industry completely reversed its position on synthetic data, moving from dismissal to universal adoption across all major labs
  3. Competitive Strategy Critique - Leading AI companies used fear-based messaging and regulatory capture as "ladder pulling" tactics to discourage competition, despite technical reality showing multiple companies achieving similar capabilities

Actionable Insights:

  • Businesses should prepare for AI systems that become more valuable with extended use rather than requiring constant updates
  • The success of synthetic data suggests AI self-improvement is already happening in practical applications
  • Market competition in AI remains viable despite attempts to create artificial barriers through fear-based messaging

Timestamp: [24:04-31:55]Youtube Icon

๐Ÿ“š References from [24:04-31:55]

People Mentioned:

  • OpenAI Team - Referenced for their AI researcher goals and GPT model development timeline
  • Effective Altruist Community - Criticized for anthropomorphizing AI systems and promoting self-awareness claims

Companies & Products:

  • OpenAI - Mentioned for GPT-3, GPT-5, GPT-6 development and AI researcher ambitions
  • Anthropic - Referenced through Claude model's alleged self-awareness capabilities
  • Cohere - Aidan Gomez's company, discussed in context of enterprise AI learning

Technologies & Tools:

  • GPT-3 - Historical example of fear-based messaging around AI safety
  • GPT-5 - Current generation model referenced as sufficient for memory improvements
  • GPT-6 - Predicted to feature continuous learning capabilities
  • Claude - Anthropic's model mentioned regarding self-awareness claims
  • Synthetic Data - Training methodology that evolved from dismissed to universally adopted

Concepts & Frameworks:

  • Continuous Learning - AI systems that improve over time without retraining
  • Synthetic Data Generation - Models creating their own training data for self-improvement
  • Ouroboros Effect - Metaphor for concerns about AI training on its own outputs
  • Exponential Takeoff Theory - Prediction that first AGI achiever would dominate market
  • Regulatory Capture - Strategy of using government regulation to limit competition

Timestamp: [24:04-31:55]Youtube Icon

๐Ÿš€ What is Aidan Gomez's perspective on AI fear-mongering?

Shifting from Fear to Opportunity

Aidan Gomez expresses strong opposition to the fear-based narrative surrounding AI technology that dominated discussions 1-2 years ago. He believes this "world should be afraid" mentality has been counterproductive and represents an "ugly strategy" despite its effectiveness.

Key Philosophical Shifts:

  1. From Destruction to Salvation - Views AI as potentially world-saving rather than world-destroying technology
  2. High Leverage Potential - Emphasizes AI's transformative power across multiple fields and industries
  3. Acceleration Over Caution - Advocates for rapid deployment rather than fear-based hesitation

The Fear Migration Pattern:

  • Past Focus: General AI apocalypse scenarios
  • Current Shift: Fear now targeting AI integration with robotics
  • Strategic Concern: Fear as a motivational tool that shifts pressure points rather than disappearing

Gomez's Position:

"We shouldn't be scared of it. We should be running towards it. We should be deploying it as quickly as we can."

The CEO notes that while the world has broadly accepted AI's potential, fear continues to migrate to new areas like robotics integration, suggesting a pattern of shifting anxiety rather than genuine risk assessment.

Timestamp: [32:02-32:56]Youtube Icon

๐Ÿข What is Cohere's business model and target market?

Enterprise-First AI Platform

Cohere positions itself as a modeling company that builds both AI models and the complete deployment framework for enterprise integration, focusing specifically on critical industries with stringent security requirements.

Core Business Components:

  1. Model Development - Custom AI models designed for enterprise constraints
  2. Platform Integration - Framework enabling seamless business system connectivity
  3. Deployment Flexibility - Multi-environment deployment capabilities

Enterprise Integration Philosophy:

Human-Level Access Requirements:

  • Sales software integration
  • HR system connectivity
  • Marketing platform access
  • Email and calendar management
  • Supply chain software integration
  • "It needs the same level of access that you grant trust and access that you grant to your humans"

Target Industries:

  • Public Sector - Government agencies and departments
  • Energy - Critical infrastructure companies
  • Financial Services - Banks and financial institutions
  • Healthcare - Medical and health organizations
  • Telecommunications - Telecom providers and infrastructure

Unique Deployment Advantage:

Multi-Environment Capability:

  • All major cloud platforms
  • On-premises deployment
  • Air-gapped environments for maximum security
  • Designed for "extremely secure settings" from inception

Timestamp: [33:02-34:33]Youtube Icon

๐Ÿ’ก What is Cohere's two-GPU constraint strategy?

Right-Sizing AI for Enterprise Reality

Cohere has implemented a deliberate constraint limiting their models to operate within two GPUs maximum, prioritizing enterprise accessibility over raw model size and capability.

Strategic Constraint Rationale:

  1. Enterprise Market Reality - Large enterprises cannot consume massive models due to cost constraints
  2. Infrastructure Limitations - Companies lack hundreds of millions of consumers to justify massive compute costs
  3. Practical Deployment - Models must fit existing enterprise infrastructure budgets

The Two-GPU Rule:

Hard Constraint Philosophy:

  • Maximum footprint: Two GPUs only
  • Optimization goal: "Squeeze the most intelligence out of those two GPUs that we can"
  • Rejection criteria: If a model requires 3+ GPUs, Cohere won't build it
  • Six-Year Focus: This constraint has guided development since company inception

Competitive Landscape Context:

Industry Model Sizes:

  • DeepSeek: 700 billion parameters, requiring 8-16 GPUs
  • Meta's Behemoth: ~1 trillion parameters, dramatically higher GPU requirements
  • GPT-4 Rumors: Potentially 32+ GPUs required
  • Cohere's Approach: Intentionally smaller, more efficient models

Market Positioning:

This constraint-based approach allows Cohere to serve enterprise customers who need AI capabilities but cannot justify the infrastructure costs of frontier models, creating a distinct competitive advantage in the enterprise market.

Timestamp: [34:40-36:34]Youtube Icon

๐ŸŒ How has Cohere scaled globally over six years?

International Expansion and Customer Base

Cohere has achieved rapid global scaling with a strategic office network and diverse international customer portfolio, reflecting the universal demand for enterprise AI solutions.

Global Office Network:

Headquarters and Regional Presence:

  • Toronto - Primary headquarters (founders' home base)
  • San Francisco - US West Coast operations
  • New York - US East Coast presence
  • Montreal - Additional Canadian operations
  • Paris - European hub
  • London - UK and European operations
  • Seoul - Asian Pacific presence
  • Riyadh - Upcoming Middle East expansion

International Customer Portfolio:

Geographic Diversity:

  • Saudi Telecom - Middle East telecommunications
  • Fujitsu - Japanese technology conglomerate
  • LG - South Korean electronics giant
  • British Government - UK public sector
  • Canadian Government - Home country public sector
  • Canadian Telcos - Domestic telecommunications
  • American Banks - US financial services

Scaling Timeline:

The company has experienced particularly rapid growth "over the past couple years," suggesting accelerated expansion following initial product-market fit establishment during their six-year development period.

This global footprint demonstrates Cohere's ability to serve critical infrastructure needs across different regulatory environments and cultural contexts.

Timestamp: [35:16-35:59]Youtube Icon

๐Ÿ“Š What does Aidan Gomez observe about enterprise AI adoption patterns?

From Experimentation to Production Deployment

Gomez identifies a fundamental shift in enterprise AI adoption, moving from scattered proof-of-concepts to focused, large-scale production deployments across entire organizations.

The Current Enterprise Reality:

Board-Driven AI Mandate:

  • Boards pressuring CEOs for AI implementation
  • Executive teams receiving "we need AI" directives
  • Decision-making driven by personal ChatGPT experiences
  • "Just like whatever that means, we need it"

Two-Phase Evolution Prediction:

  1. Aperture Narrowing - Companies will reduce from "thousand flowers bloom" approach to focused investments
  2. Model Scaling Slowdown - AI model improvements have already begun plateauing

The Transformation Gomez Observes:

From Low to High Appetite:

Previous Phase (1-2 years ago):

  • 30+ different use case experiments
  • Small teams (5 people per project)
  • Proof-of-concept mentality
  • Board pressure driving unfocused initiatives
  • "My board is going to fire me if I don't figure out some sort of strategy around AI"

Current Phase:

  • Narrowed focus on proven use cases
  • Deployment across tens of thousands of employees
  • Production-scale implementation
  • ROI-driven conviction
  • Massive scaling in actual deployment

Cohere's Advantage:

Gomez notes that while many POCs failed industry-wide, Cohere helped customers avoid unsuccessful experiments from the outset, contributing to their success in this transition period.

Timestamp: [36:46-39:55]Youtube Icon

๐Ÿ’Ž Summary from [32:02-39:55]

Essential Insights:

  1. AI Philosophy Shift - Gomez advocates moving from fear-based AI narratives to embracing rapid deployment for world-saving potential
  2. Enterprise-First Strategy - Cohere's two-GPU constraint prioritizes enterprise accessibility over frontier model capabilities
  3. Global Market Validation - Rapid international expansion across critical industries demonstrates universal enterprise AI demand

Actionable Insights:

  • Enterprise AI adoption is transitioning from scattered experiments to production-scale deployment across entire organizations
  • Security-conscious industries require on-premises and air-gapped deployment capabilities that cloud-only providers cannot serve
  • Strategic constraints can create competitive advantages by aligning product capabilities with market realities rather than pursuing maximum technical performance

Timestamp: [32:02-39:55]Youtube Icon

๐Ÿ“š References from [32:02-39:55]

People Mentioned:

  • Nick - Cohere co-founder from Toronto
  • Ivan - Cohere co-founder from Toronto

Companies & Products:

  • Cohere - AI modeling company focused on enterprise deployment
  • DeepSeek - AI company with 700 billion parameter models
  • Meta - Developer of the "Behemoth" trillion-parameter model
  • Saudi Telecom - Cohere customer in telecommunications
  • Fujitsu - Japanese technology conglomerate and Cohere customer
  • LG - South Korean electronics company and Cohere customer
  • ChatGPT - Referenced as driving enterprise AI awareness

Technologies & Tools:

  • GPT-4 - Advanced language model requiring significant GPU resources
  • GPU Computing - Hardware constraint driving Cohere's two-GPU strategy
  • On-premises Deployment - Secure AI deployment model for sensitive industries
  • Air-gapped Systems - Highest security deployment environment for AI models

Concepts & Frameworks:

  • Two-GPU Constraint - Cohere's strategic limitation for enterprise accessibility
  • Enterprise AI Integration - Human-level access requirements for business systems
  • Proof-of-Concept (POC) Phase - Initial enterprise AI experimentation period
  • Production Deployment - Large-scale AI implementation across organizations

Timestamp: [32:02-39:55]Youtube Icon

๐Ÿš€ How does Cohere scale AI across enterprise businesses?

Enterprise AI Deployment Strategy

Cohere has moved beyond the experimental phase into large-scale enterprise deployment, where companies are finally seeing real ROI from AI investments.

Current Market Phase:

  • Past: Companies ran small pilot programs with 30-1,000 users
  • Present: Enterprises are scaling AI solutions across entire organizations
  • Impact: True ROI only emerges with broad deployment across businesses

Two-Pronged Go-to-Market Approach:

1. Self-Service API Platform:

  • Direct access to Cohere's model suite via API
  • Core Models: Rerank, Embed, Command, and additional specialized models
  • Completely self-served for technical teams

2. Strategic Enterprise Engagements:

  • Full-Service Partnership: Cohere provides models plus complete tech stack
  • Collaborative Execution: Joint development of customer's AI agent roadmap
  • Knowledge Transfer: Deploy FTEs to train client teams on implementation
  • Gradual Independence: Eventually transition clients to run independently on Cohere's platform

Enterprise Transformation Process:

Companies now understand where AI works and where it doesn't, making informed bets and scaling successful implementations across their entire business operations.

Timestamp: [40:01-41:48]Youtube Icon

๐Ÿ’ฐ How much funding has Cohere raised and where does it go?

$1.7 Billion Investment Breakdown

Cohere has raised approximately $1.6-1.7 billion over six years to build their enterprise AI platform.

Primary Investment Categories:

1. Compute Infrastructure (Major Portion):

  • Massive computational resources for training and running AI models
  • Scaling to support global enterprise deployments

2. Talent Acquisition (Significant Investment):

  • Premium Talent: Very expensive, specialized AI researchers and engineers
  • Global Team: Six years of building international presence
  • Expertise Premium: Reflects the scarcity and value of AI talent

3. Go-to-Market Expansion:

  • Building global presence and sales infrastructure
  • Enterprise engagement and support capabilities

Resource Allocation:

80% of funding goes toward the combination of compute, data, and people - the three critical pillars for scaling enterprise AI solutions.

Investment Rationale:

The substantial funding reflects the capital-intensive nature of building enterprise-grade AI infrastructure that can serve major corporations globally while maintaining performance and reliability standards.

Timestamp: [41:48-42:41]Youtube Icon

๐ŸŽฏ How many S-tier AI researchers exist globally?

The Elite AI Talent Pool

According to Cohere CEO Aidan Gomez, there are approximately 150-200 S-tier AI researchers that all major AI labs are competing to recruit.

The Talent Landscape:

S-Tier Researcher Characteristics:

  • Universal Recognition: Everyone in the industry knows who they are
  • Master Lists: All major labs track these individuals
  • Competitive Intelligence: Companies monitor their career satisfaction and management relationships

Market Competition Reality:

  • Extreme Compensation: Some offers reach $100 million annually (particularly from Meta)
  • Cohere's Strategy: Deliberately avoids competing on pure compensation
  • Talent Philosophy: Focus on mission-driven individuals rather than "mercenary talent"

Growing Talent Pool:

The number may be expanding to 4x the original estimate (potentially 600-800) as more people gain hands-on experience in AI research and development over recent years.

Cohere's Recruitment Philosophy:

  • Mission-Driven Focus: Seek researchers committed to building generational organizations
  • Long-term Commitment: Prefer talent motivated by purpose over pure financial incentives
  • Equity Upside: Offer potential for 10x returns versus established companies with limited growth potential

Timestamp: [42:41-45:28]Youtube Icon

๐Ÿง  Was Aidan Gomez considered S-tier talent when leaving Google?

Elite AI Research Background

Aidan Gomez confirms he possessed extremely rare expertise when founding Cohere, representing one of the most exclusive groups in AI research.

Unique Experience Profile:

  • Planetary Scarcity: Only 15-30 people globally had trained language models when Cohere started
  • Specialized Knowledge: Tens of people worldwide had worked on similar transformer architecture projects
  • Founding Advantage: This exclusive experience base provided crucial competitive positioning

Current Status of Early Pioneers:

  • Leadership Positions: Most early language model researchers now hold senior roles across the industry
  • Entrepreneurial Path: Very few became heads of startups or founded companies
  • Distributed Impact: These pioneers are "all over the place" in various leadership capacities

The Researcher-to-CEO Transformation:

  • Natural Introversion: Researchers typically more shy and focused on technical work
  • Learned Extraversion: Public-facing CEO role requires "faking extraversion"
  • Energy Drain: Speaking engagements and public appearances feel like "drudgery"
  • Skill Development: Five years of practice has made public communication less tiring
  • Ongoing Challenge: Still requires significant effort to perform extraverted behaviors

This transformation from Oxford researcher to Google AI researcher to global CEO represents an extreme metamorphosis that goes against natural tendencies.

Timestamp: [45:28-47:52]Youtube Icon

๐Ÿ’Ž Summary from [40:01-47:52]

Essential Insights:

  1. Enterprise AI Maturity - Companies have moved from small pilots to organization-wide AI deployments, finally achieving meaningful ROI
  2. Talent Scarcity Reality - Only 150-200 S-tier AI researchers exist globally, creating intense competition among major labs
  3. Strategic Funding Allocation - Cohere's $1.7B investment focuses 80% on compute, data, and premium talent acquisition

Actionable Insights:

  • Enterprise Approach: Successful AI deployment requires scaling beyond test groups to achieve real business impact
  • Talent Strategy: Mission-driven recruitment outperforms pure compensation competition for building lasting organizations
  • Market Timing: The AI talent pool is expanding 4x as more professionals gain hands-on experience, creating new opportunities

Timestamp: [40:01-47:52]Youtube Icon

๐Ÿ“š References from [40:01-47:52]

People Mentioned:

  • Aidan Gomez - Co-founder and CEO of Cohere, former Google AI researcher with expertise in transformer architecture

Companies & Products:

  • Cohere - Enterprise AI platform providing language models and deployment infrastructure
  • Google - Former employer where Gomez developed foundational AI research experience
  • Meta - Competitor offering extreme compensation packages ($100M annually) for top AI talent
  • Palantir - Referenced as comparison for enterprise platform sales approach

Technologies & Tools:

  • Rerank Model - Cohere's specialized model for improving search and retrieval results
  • Embed Model - Cohere's embedding model for semantic understanding and similarity
  • Command Model - Cohere's flagship language model for enterprise applications
  • API Platform - Self-service interface for accessing Cohere's model suite

Concepts & Frameworks:

  • S-tier Researchers - Elite category of AI researchers (150-200 globally) that all major labs compete to recruit
  • Enterprise AI Deployment - Strategic approach to scaling AI solutions across entire organizations rather than small pilot programs
  • Mission-driven Talent - Recruitment philosophy focusing on purpose-driven individuals over purely compensation-motivated candidates

Timestamp: [40:01-47:52]Youtube Icon

๐ŸŽค How does Cohere CEO Aidan Gomez feel about public speaking?

Personal Growth and Leadership Development

Speaking Challenges:

  • Natural introversion: Prefers small groups and working on interesting problems over being on stage
  • Not a natural public speaker: Admits it's not something he naturally enjoys
  • Persistent shyness: The shyness never completely goes away, but becomes manageable

Learning Process:

  1. Gradual improvement - Has gotten "good enough at it to be productive"
  2. Finding enjoyment - Can learn to see the nice parts in public speaking
  3. Overcoming barriers - Shyness gets easier to manage over time

Benefits of Public Speaking:

  • Meeting interesting people - Opportunity to connect with cool individuals
  • Discussing fascinating topics - Platform to talk about interesting subjects
  • Sharing perspectives - Chance to give opinions and views to the world
  • Personal growth - Learning and growing through the process

Practical Approach:

  • Seeing past discomfort - Can appreciate the benefits even when the process feels uncomfortable
  • Enjoying conversations - Finds genuine enjoyment in meaningful discussions like interviews
  • Focusing on value - Concentrates on the positive aspects rather than the anxiety

Timestamp: [48:00-49:26]Youtube Icon

๐Ÿข What does Aidan Gomez think about Google's AI comeback?

Google's AI Renaissance and Market Position

Leadership Impact:

  • Demis Hassabis saved Google - CEO of DeepMind (now absorbed into Google) led the turnaround
  • Gemini leadership - Demis technically runs the Gemini project
  • Strategic transformation - Moved Google from "missing the AI wave" to potential market leader

Current Competitive Position:

  1. Potentially the best model - May have surpassed OpenAI with current capabilities
  2. Strong technological foundation - Building excellent models that seem great
  3. Uncertain future - Will need to see performance with Gemini 3

Google's Advantages:

  • Money printing machine - Search revenue can fuel AI development
  • Data printing machine - Continuous data absorption capabilities
  • Talent concentration - Google Brain and DeepMind like "Bell Labs" with surreal expertise
  • Complete resources - Had everything needed for technological catch-up

Market Challenge:

Consumer Focus Imperative:

  • Core identity: Google is "a consumer company through and through"
  • Enterprise limitations: Cloud and enterprise efforts are secondary
  • Critical success factor: Must win consumer market or face serious consequences
  • Low market share: Gemini currently has very low but growing market share

Personal Connection:

  • Mentorship relationship: "Google sort of raised me" - worked there as a student
  • Continued support: Hopes Google succeeds and is proud of their comeback
  • Professional respect: Impressed by what Gemini team is accomplishing

Timestamp: [49:45-52:34]Youtube Icon

๐Ÿก What was Aidan Gomez's childhood like in the Canadian wilderness?

Unique Canadian Upbringing and Early Technology Passion

Wilderness Childhood:

  • Born in Toronto but moved before age one
  • Log home construction - Father built a log house in 150 acres of Canadian wilderness
  • Traditional activities - Tapped maple trees in March when sap runs, operated a sugar shack
  • Immigrant parents - Father from Spain, mother British, giving him "the world's most Canadian upbringing"

Early Technology Obsession:

Gaming Console Modifications:

  1. Wii hacking - Would unscrew the console and install chips from China
  2. Free games access - Chips provided access to numerous games without cost
  3. PlayStation jailbreaking - Modified gaming systems for enhanced functionality
  4. Hands-on learning - Physical hardware manipulation and modification

Technology as Passion:

  • Obsessive interest - "Just obsessed" with computers and technology from early age
  • Self-directed learning - Independently researched and implemented modifications
  • Problem-solving focus - Enjoyed figuring out how to hack and modify devices
  • Foundation building - Early experiences shaped later academic and career choices

Timestamp: [52:34-53:57]Youtube Icon

๐ŸŽ“ Why did Aidan Gomez choose AI at University of Toronto?

Academic Journey and AI's Fundamental Appeal

University Selection:

  • Local choice - Chose University of Toronto (UofT) because it was his local university
  • Timing - Started in 2013-2014
  • Serendipitous outcome - Accidentally ended up at one of the world's top AI centers

Toronto's AI Ecosystem:

World-Class Talent Concentration:

  • Geoffrey Hinton - Legendary AI researcher based there
  • Ilya Sutskever - Notable alumnus who came from UofT
  • Incredible faculty - "Tons of incredible people who have defined the entire field"
  • Deep expertise - Access to professors, PhD students, and postdocs with unmatched AI knowledge

Unique Learning Environment:

  • Immersive experience - "You sort of get raised into AI"
  • Knowledge concentration - Toronto had exceptional concentration of AI expertise
  • Historical significance - "Toronto was really the place that got built" for AI

AI's Philosophical Appeal:

The Great Mystery:

  1. Most interesting unsolved question - AI represented the ultimate intellectual challenge
  2. Intelligence as uniqueness - What separates humans qualitatively from all other animals
  3. Fundamental mystery - Understanding how and why humans became so intelligent

Scientific Comparison:

  • Physics limitations - While we know much about physics and can predict with precision, intelligence remains mysterious
  • Human explanations - Many different theories about how, why, and what intelligence means
  • Exploration opportunity - Thought exploring intelligence was "the most exciting" pursuit

Timestamp: [53:57-55:58]Youtube Icon

๐Ÿ’Ž Summary from [48:00-55:58]

Essential Insights:

  1. Personal growth in leadership - Aidan overcame natural introversion to become an effective public speaker, learning to appreciate the benefits despite persistent shyness
  2. Google's AI renaissance - Under Demis Hassabis's leadership, Google transformed from "missing the AI wave" to potentially having the best AI model, though consumer market success remains critical
  3. Formative experiences shape careers - Growing up in Canadian wilderness with early technology obsession, combined with accidentally choosing UofT, led to immersion in world-class AI research environment

Actionable Insights:

  • Leadership development: Introverted leaders can learn public speaking by focusing on the value and connections rather than the discomfort
  • Competitive analysis: Companies with strong technical foundations and resources can make dramatic comebacks with proper leadership
  • Career serendipity: Sometimes local choices and following genuine interests lead to extraordinary opportunities in emerging fields

Timestamp: [48:00-55:58]Youtube Icon

๐Ÿ“š References from [48:00-55:58]

People Mentioned:

  • Demis Hassabis - CEO of DeepMind, credited with saving Google's AI efforts and leading Gemini
  • Geoffrey Hinton - Legendary AI researcher at University of Toronto who helped define the field
  • Ilya Sutskever - Notable University of Toronto alumnus who became a key figure in AI

Companies & Products:

  • Google - Aidan's former employer that "raised him" and is now competing strongly in AI
  • DeepMind - AI research company absorbed into Google, led by Demis Hassabis
  • Gemini - Google's AI model that may have surpassed OpenAI's capabilities
  • OpenAI - AI company that Google's Gemini may have surpassed
  • University of Toronto - Where Aidan studied and got immersed in AI research

Technologies & Tools:

  • Nintendo Wii - Gaming console Aidan modified by installing chips for free games
  • PlayStation - Gaming system he jailbroke as part of his early technology exploration

Concepts & Frameworks:

  • Google Brain - Google's AI research division compared to Bell Labs for talent concentration
  • Bell Labs - Historical research institution used as comparison for Google's AI talent concentration
  • Consumer vs Enterprise Markets - Strategic distinction crucial for Google's AI success

Timestamp: [48:00-55:58]Youtube Icon

๐ŸŽ“ What was Aidan Gomez's decision between joining OpenAI and pursuing a PhD?

Career Crossroads at the Dawn of AI

The OpenAI Opportunity:

  • Early conversations with Ilya Sutskever - Discussions about joining OpenAI around 8 years ago, coinciding with the Transformer paper development
  • Strategic timing - This was before GPT-3 existed, when the AI landscape looked completely different
  • Personal connection - Knew Ilya from his student days at University of Toronto

The PhD Choice:

  1. Desire for continued learning - Wanted to deepen his academic foundation
  2. Oxford opportunity - Chose to pursue doctoral studies instead of immediate industry work
  3. No regrets - Would make the same decision 100% if faced with it again today

Industry vs Academia Dynamics:

  • Frontier research shift - Most cutting-edge AI work now happens in industry rather than academia
  • Knowledge sharing concerns - Industry research often doesn't get published or shared publicly like academic work
  • Strategic advantage - PhD background provided a more interesting foundation for his current position

Timestamp: [56:51-58:07]Youtube Icon

๐Ÿค– What was the first transformer language model output that amazed Aidan Gomez?

The Moment Everything Changed

The Breakthrough Email:

  • Source: Manager Lukasz sent an email with the subject "Aidan, check this out"
  • The prompt: Simply "Title: The Transformer"
  • The output: Complete Wikipedia page about a fictional Japanese punk rock band
  • The revelation: Manager wrote only the title - the machine wrote everything else

Historical Context:

  1. Before this moment - Computers could barely string sentences together properly
  2. The transformation - Went from "dumb things that bumbled" to human-level writing quality
  3. Instant realization - This represented a fundamental leap in AI capabilities

Personal Impact:

  • Still preserved - Gomez keeps this email on his phone as a historical artifact
  • Paradigm shift - Demonstrated that fluent computer writing had arrived seemingly overnight
  • Wake-up call - Recognized this as a pivotal moment in AI development

The Irony:

Today, generating a Wikipedia page about "The Transformer" seems obvious and unremarkable, but at that moment it represented an impossible leap forward in machine capabilities.

Timestamp: [58:42-59:59]Youtube Icon

๐Ÿข Why does Aidan Gomez see massive enterprise AI opportunities with current technology?

The Untapped Potential of Today's AI

Current Enterprise Reality:

  • Basic applications dominate - Most companies still using AI for simple tasks like email summaries and meeting notes
  • Test phase mentality - Models are essentially still in experimental stages doing lowest-value work
  • Foundational level - Enterprise adoption remains at very elementary stages

The White Collar Opportunity:

  1. Supply-constrained market - White collar workers are called that because you have to pay them well due to scarcity
  2. High demand, limited supply - Tons of demand for skilled workers but not enough people to meet global needs
  3. Perfect AI match - Models excel at exactly the types of tasks these workers perform

Proven Success Areas:

  • Coding - Already a "home run" for AI augmentation
  • Legal work - Demonstrating strong AI integration success
  • Finance - Next major field expected to see significant AI adoption

Transformative Potential:

  • Beyond seat licenses - Moving from traditional software licensing to actually augmenting human capabilities
  • Job augmentation - AI doing portions of people's actual work rather than just supporting tools
  • Economic integration - Models becoming integral parts of the economy rather than peripheral tools

Timestamp: [1:00:05-1:02:26]Youtube Icon

๐Ÿ”ฎ What does Aidan Gomez predict will be obvious about AI's impact in 5-10 years?

The Coming Economic Transformation

Productivity Revolution:

  1. Economic percolation - AI technology will deeply integrate into the broader economy
  2. Company productivity gains - Businesses will see significant efficiency improvements
  3. Impossible efficiency levels - Teams of 1,000 people doing work that currently requires hundreds of thousands

Labor Market Evolution:

  • Fundamental capability shifts - Products and companies will be built that were previously impossible
  • Workforce augmentation - AI will handle larger portions of what people do today, especially white-collar work
  • Market transformation - Moving beyond traditional software models to actual job function replacement

Global Economic Recovery:

  • Stagnation reversal - Hope for resumed growth across developed economies
  • Productivity renaissance - Potential end to the productivity stagnation of recent decades
  • Regional challenges - Many countries have seen flat or declining GDP per capita over 10-15 years

Geographic Concerns:

  • Canada and UK struggles - Both countries facing economic stagnation despite fewer regulatory barriers than EU
  • European challenges - EU-style regulations creating additional obstacles
  • Asian market impacts - Various parts of Asia also experiencing economic flatness

Timestamp: [1:02:26-1:03:56]Youtube Icon

๐Ÿ’Ž Summary from [56:07-1:03:56]

Essential Insights:

  1. Strategic career decisions - Gomez chose PhD over early OpenAI opportunity, providing stronger foundation for current success
  2. Historical AI moment - First transformer output (fictional Wikipedia page) marked the instant transition from basic to human-level computer writing
  3. Enterprise opportunity - Current AI applications remain basic despite massive potential for white-collar work augmentation

Actionable Insights:

  • Companies should look beyond basic AI applications like email summaries to more substantial workflow integration
  • White-collar professions represent the highest-impact areas for AI implementation due to supply constraints
  • The next 5-10 years will likely bring fundamental economic transformation through AI productivity gains

Timestamp: [56:07-1:03:56]Youtube Icon

๐Ÿ“š References from [56:07-1:03:56]

People Mentioned:

  • Ilya Sutskever - Former OpenAI co-founder who had early conversations with Gomez about joining the company
  • Geoffrey Hinton - AI pioneer mentioned as part of the University of Toronto AI community
  • Yann LeCun - AI researcher noted as another University of Toronto alumnus
  • Nick - Gomez's co-founder who plays chess with Geoffrey Hinton every Monday
  • Lukasz - Gomez's former manager who sent the first transformer output email

Companies & Products:

  • OpenAI - Company that tried to recruit Gomez around 8 years ago, before GPT-3 existed
  • University of Toronto - Academic institution that produced many leading AI researchers
  • Oxford University - Where Gomez pursued his PhD instead of joining OpenAI

Technologies & Tools:

  • Transformer - The foundational AI architecture that Gomez co-authored in "Attention Is All You Need"
  • GPT-3 - Referenced as coming after Gomez started Cohere, marking a timeline in AI development

Concepts & Frameworks:

  • White collar work augmentation - The concept that AI will primarily impact knowledge workers due to supply constraints
  • Enterprise AI adoption phases - Current basic applications vs. future transformative integration

Timestamp: [56:07-1:03:56]Youtube Icon

๐ŸŒ Why does Aidan Gomez believe economic growth prevents global conflicts?

Economic Growth and Global Stability

The Growth-Peace Connection:

  1. Historical prosperity pattern - Over the past century, everyone has gotten richer with access to healthcare, education, and improved living standards
  2. Expanding pie theory - When the economic pie grows, people don't need to fight because everyone benefits from increasing wealth
  3. Zero-sum mentality emerges - When growth stops, people must take from others to improve their position, leading to conflict

Dangerous Consequences of Economic Stagnation:

  • Rise of xenophobia - People blame immigrants for taking their share of a static economic pie
  • Territorial conflicts - Nations seek to expand access to resources by taking from others
  • Political regression - Movement away from liberal democracies toward authoritarianism and autocratic rule

AI's Role in Resuming Growth:

  • Technology diffusion - Spreading AI technology throughout the economy can restart growth cycles
  • Extended prosperity - Potential for another half century to century of economic expansion
  • Democratic preservation - Continued march away from authoritarianism and toward liberal democratic systems

Timestamp: [1:04:03-1:05:38]Youtube Icon

๐Ÿ‡ช๐Ÿ‡บ What does Cohere CEO think is holding Europe back in technology?

Europe's Technology Challenge

The Regulatory-First Approach:

  1. Police mentality - The EU has positioned itself as the regulator of other countries' tech companies
  2. Single tool syndrome - Their primary strategy is "regulate, regulate, regulate" rather than building
  3. Misplaced focus - Celebrating minor regulatory wins like USB-C standardization while missing bigger opportunities

What Europe Should Do Instead:

  • Build, don't police - Focus on creating competitive European tech companies rather than controlling foreign ones
  • Leverage existing strengths - Europe has incredible universities and pools of capital available
  • Organizational imperative - Europeans need to organize themselves to build the next generation of great companies

Personal Investment in Europe:

  • Dual citizenship - Aidan holds both British and Spanish passports
  • Family connections - His wife studied in Barcelona and he lives in London
  • Deep care - Genuinely invested in Europe's success and technological future

The Path Forward:

Building their own companies will be the only way to strengthen Europe - not protecting themselves from other companies, but creating competitive alternatives.

Timestamp: [1:06:00-1:08:47]Youtube Icon

๐Ÿข What positions is Cohere actively hiring for right now?

Cohere's Hiring Expansion

Current Hiring Status:

  • Everything - Hiring across all functions and departments
  • Extremely resource constrained - Way too small for current needs and growth trajectory
  • Rapid scaling required - Need to double or triple team sizes in multiple areas

Specific Roles in High Demand:

Research and Development:

  • ML researchers - Direct outreach encouraged via Twitter/social media
  • Technical positions - Part of the core 200-800 person target range

Sales and Customer Success:

  • Sales team expansion - Need to double or triple current sales force
  • Delivery and implementation - Field Technical Engineering Employees (FTEEs) for customer deployment

Growth Across All Functions:

  • Universal expansion - Every department needs significant headcount increases
  • Immediate need - Company is currently way too small for market demands

How to Apply:

  • ML researchers - Message Aidan directly on Twitter
  • Other positions - Standard application processes for various functions

Timestamp: [1:09:53-1:10:42]Youtube Icon

๐Ÿ’ช How does Aidan Gomez define grit?

Personal Definition of Grit

Core Elements:

  1. Dirt and toughness - Raw, fundamental resilience
  2. Pain tolerance - The ability to withstand and endure difficult situations
  3. Endurance mindset - Capacity to persist through challenging circumstances

Simple but Powerful:

The definition comes down to the ability to withstand pain - a straightforward but profound understanding of what it takes to succeed in challenging endeavors.

Timestamp: [1:10:47-1:11:09]Youtube Icon

๐Ÿ’Ž Summary from [1:04:03-1:11:34]

Essential Insights:

  1. Economic growth prevents conflicts - When economies stagnate, zero-sum thinking leads to xenophobia, territorial disputes, and political regression away from democracy
  2. Europe's regulatory trap - The EU focuses on policing other countries' tech companies instead of building their own competitive alternatives
  3. AI as growth catalyst - Diffusing AI technology throughout the economy could resume growth for another half century to century

Actionable Insights:

  • Europe needs to leverage its universities and capital to build great tech companies rather than just regulate foreign ones
  • Cohere is aggressively hiring across all functions, particularly ML researchers and sales teams
  • True grit means developing the ability to withstand pain and persist through difficult circumstances

Timestamp: [1:04:03-1:11:34]Youtube Icon

๐Ÿ“š References from [1:04:03-1:11:34]

People Mentioned:

  • CIO of Deutsche Bank - Dinner conversation partner who discussed European technology challenges and cultural preservation

Companies & Products:

  • Deutsche Bank - German multinational investment bank whose CIO provided perspective on European tech challenges
  • European Union (EU) - Regulatory body discussed in context of technology policy and approach to foreign tech companies
  • USB-C standardization - EU regulatory achievement cited as example of focus on control rather than innovation

Locations:

  • London, Soho - Aidan's office location and residence area
  • Barcelona - Where Aidan's wife studied
  • Toronto - Mentioned as having the best Indian food in Aidan's opinion

Concepts & Frameworks:

  • Zero-sum economic thinking - When economic growth stops, people must take from others to improve their position
  • Liberal democracy progression - Historical march away from authoritarianism and kings toward democratic systems
  • Protectionism vs. innovation - Europe's regulatory approach versus building competitive companies

Timestamp: [1:04:03-1:11:34]Youtube Icon