Skip to content

Interview Preparation (DeepMind, CoreAI)-Phil’s MacBook Air

:::note] :::

  • The Copilot product is in a good place
  • The cross-company branding is a mess
  • Need to focus on:
    • Distribution
    • Memory
    • Use emerging open protocols (MCP) to take advantage of others’ capabilities (M365 data, etc.)
  • Drive the cost of experimentation to zero
  • definitely have the end-to-end
  • fine balance for rigor commerciality and science

[!Who I am]

  • Intersection of: new technology x business strategy
  • Enjoy working on different ambiguous, interesting projects and bringing clarity
  • Experience across software engineering, program management, strategy and partnerships.
  • Worked as an engineer, business development & strategy, program management
    • OpenAI, AMD, ASML / Disney / Lenfest Institute
  • Designing and nurturing partnerships, bringing technology to market.
    • NOW LOOKING FOR AN EXCITING ROLE WITH A MISSION-DRIVEN ORGANIZATION WHERE I CAN WORK AT THE FOREFRONT OF TECHNOLOGY, BUT HELP BRING IT TO MARKET IN A REAL WAY: INVENTION → INNOVATION
  • Experience working at the intersection of new technology and business strategy.
  • Demonstrated success leading multi-disciplinary, cross-company teams and programs.
  • Understanding of how large companies (Microsoft) work
  • A mission driven effort that helps improve the world in which we live.
  • Honesty, team work, solving for the global maximum
  • ‘Inhuman’ intelligence
  • Focus on utility and identify human tasks that technology can help with.
  • 5 second tasks → 5 minutes → 5 hours / days / months
  • Memory is key → people will forgo 10 IQ points if only a system remembered you better.
  • Concern about subscription business.
  • Concern about impact on core business.

What Do I Think is Most Interesting about GDM’s Current Work?

Section titled “What Do I Think is Most Interesting about GDM’s Current Work?”
  • David Silver’s “Era of Experience” paper
  • World Models

Why GDM? What’s the Unique Draw for PJW?

Section titled “Why GDM? What’s the Unique Draw for PJW?”
  • End-to-end focus on solving intelligence, and delivering is safely and responsibly.
  • Demis: understand the world around us and help advance human knowledge.
  • Act in the world!

Open: image.png

4 partnership principles:

  1. Build trust – we will always have disagreements. Trust is the medicine.
  2. Maintain alignment – we will define and redefine shared goals and desired outcomes often
  3. Show persistence – we tackle complex problems together
  4. Have fearlessness – we will be optimists, and won’t be afraid to do hard things or take risks

Setting goals and running projects:

  • People often assume there’s a ‘what’ problem, when it’s really a ‘how’ problem
  • Process goals > performance goals > outcome goals
  • Creativity is keeping going after the first “good enough” answer
  • Building a team is recognizing individuals while managing for team goals
    • hire the best → solve for ability to learn and improve
    • lead so you can push effective decision making to the edge
    • walk the walk
  • Drive ambition → develop the platform → differentiate

Run good experiments:

  • what should I do → what should I learn?
  • help learn new, specific, valuable information
  • are worthwhile
  • predicated on a reasonable hypothesis about what will happen
  • is as small as possible to deliver
  • can scale those learnings
  • takes advantage of any unique capabilities you have

Run good projects:

  • focus
  • keep a detailed plan for victory
  • run a fast observe / orient / decide / act loop
  • overcommunicate
  • break off sub-projects


Why AI Product Management Is Different

Playbook Tip:

  • Lead with output quality.
  • In AI, the model is both the compiler and the product.
  • Ensuring it meets user needs and quality benchmarks is the foundation upon which great UX can then be layered.

AI product management differs from traditional software in three critical ways:

  1. Probabilistic Outputs: AI generates variable results influenced by training data, prompts, and real-world usage.
  2. Quality First: If model outputs are poor, no UI/UX polish can compensate.
  3. Rapid Evolution: AI frameworks, best practices, and models change quickly, requiring constant adaptation.

Traditional PM vs. AI PM

Traditional Product ManagementAI Product Management
Definition of “Good”Features are defined by a set of functional requirements and deterministic logic. If the feature meets specs, it’s “good.”Quality is probabilistic; “good” is defined by metrics like accuracy, relevance, clarity, or user satisfaction. Continuous measurement and clear criteria (golden sets, test sets) are essential.
Spec & RequirementsSpecifications center on predefined features, acceptance criteria, and deterministic logic. Requirements are mostly about how the system should behave under various conditions.Specs must explicitly define what good looks like through sample prompts, golden sets, and evaluation metrics. AI PMs must provide annotated examples, success benchmarks, and clear criteria for acceptable vs. unacceptable outputs.
Empirical MindsetValidation relies on predefined use cases, acceptance criteria, and manual QA.Demands a data-driven, experimental approach. Product teams must continuously test, measure output quality, and refine based on real-world feedback and metrics.
Core FocusThe UI/UX and workflow design often take precedence. If the feature’s logic is correct, a polished experience is enough.AI output quality is paramount, overshadowing UI design. A subpar model output can negate even the best-designed interface.
Feature Crew DisciplinesPrimary collaboration: Product Managers, Engineers, UX Designers, and Copywriters.Deep collaboration is needed with applied research (for AI model development, prompt engineering, data pipelines) and technical writers (to craft prompts, refine model responses), in addition to classic disciplines (UX, copy, eng).
Data RequirementsMostly static requirements and configurations; data typically is for analytics or minimal business logic.Robust, high-quality datasets drive output evaluation and improvement.
IterationIteration is usually tied to feature roadmaps and version releases; updates are less frequent once feature logic stabilizes.An ongoing cycle of prompt tuning, model retraining, and evaluation. AI features often see continuous updates as the model and data evolve.
Evaluation & TestingTest cases and QA checklists ensure deterministic outcomes match the specification.Golden sets, automated metrics, LLM-as-judge pipelines, and human reviews. Success is assessed against empirical benchmarks and user feedback loops.
Stakeholder CollaborationProduct, marketing, and user research typically align on messaging once core feature functionality is locked.Tight cross-functional alignment is critical. Marketing must understand AI’s capabilities and limits; user research must inform ongoing prompt/model refinements.
Risk of FailureBugs or mismatched features can lead to user frustration, but issues are often binary and more predictable.AI outputs can fail in subtle ways—incorrect facts, biased or confusing responses. Failures may be less predictable and require robust risk mitigation (e.g., human-in-the-loop evaluations).
User ExpectationsConsistent functionality once a feature “works.”Variable output quality; must manage expectations and clarify limitations.
Safety & RAIPrivacy & Security requirements focus on data protection, regulatory compliance, and standard code-of-conduct.Goes beyond privacy/security to include algorithmic bias detection, content moderation, ethical usage guidelines, and frameworks for responsible AI (e.g., fairness, transparency, governance).