Key Moments
How To Hire AI Engineers (ft. James Brady and Adam Wiggins of Elicit)
Key Moments
Hiring AI engineers requires a blend of software skills, ML curiosity, and a fault-tolerant mindset, with interviews simulating real work.
Key Insights
AI engineering is approximately 90% software engineering with a critical 10% specialized AI/ML knowledge.
Key attributes for AI engineers include strong conventional software engineering skills, curiosity for ML/LLMs, and a "fault-first" mindset for building resilient systems.
LLMs introduce unpredictability in latency and response content, necessitating robust error handling, retries, fallbacks, and strong typing.
Interview processes should simulate real-world tasks, focusing on defensive programming and system design that accounts for potential failures.
Curiosity for LLM capabilities is crucial, but it should be tied to product goals and user needs, not just technological advancements.
Sourcing AI engineers involves a mix of clear employer branding, outbound outreach, and engagement in relevant communities like hackathons and specialized job boards.
DEFINING THE AI ENGINEER ROLE
The role of an AI engineer blends conventional software engineering with specialized AI/ML knowledge. James Brady, Head of Engineering at Elicit, notes that while AI engineering is often described as 90% software engineering, the remaining 10% is highly differentiated and critical. This blend stems from the need to build robust applications on top of inherently unpredictable technologies like large language models (LLMs).
CORE SKILLS: SOFTWARE ENGINEERING AND FAULT TOLERANCE
Effective AI engineers possess strong conventional software engineering skills as a baseline. Crucially, they must also have a "fault-first" mindset, meaning they proactively build systems that can handle failures. This is essential due to the high variability in LLM latency and response content, requiring techniques like retries, fallbacks, and careful error handling.
NAVIGATING LLM UNPREDICTABILITY
Working with LLMs introduces significant challenges compared to traditional APIs. Latency can vary by a factor of ten, and response formats and semantics are naturally unpredictable. AI engineers need to build resilient applications that provide a stable user experience despite this underlying chaos. This often involves applying principles from distributed systems engineering at the application level, such as strong typing and checked exceptions, to manage data shapes and potential errors.
INTERVIEWING FOR THE RIGHT MINDSET
Interviewing AI engineers requires moving beyond traditional happy-path coding challenges. Technical exercises should incorporate adding features and fixing bugs within a codebase that simulates the unpredictable nature of LLMs. System design interviews can effectively probe for a fault-first mindset by posing hypothetical failure scenarios, such as node failures or network slowness.
CULTIVATING CURIOSITY AND PRODUCT FOCUS
A genuine curiosity and enthusiasm for machine learning and LLM capabilities are vital. However, this must be coupled with a product mindset. AI engineers should be excited by new models and features but always frame their exploration around how these advancements can solve user problems or improve the product's strategic goals, rather than just pursuing them for their own sake.
SOURCING TALENT AND EMPLOYER BRANDING
Effective sourcing for AI engineers involves a multifaceted approach. This includes maintaining an active online presence through blogs and social media, engaging in relevant communities (hackathons, conferences), and conducting targeted outbound outreach. For smaller organizations, demonstrating a clear mission, great teammates, and a compelling product through employer branding is crucial for attracting top talent.
THE ML-FIRST MINDSET ADJUSTMENT
Adopting an "ML-first" approach requires a significant mindset shift, moving away from the need for complete control over every component. It involves relinquishing some control to opaque black-box models and developing comfort with unexpected outputs. While this can lead to powerful emergent capabilities, it necessitates careful integration, often using regular expressions or other post-processing for validation, balancing innovation with necessary robustness.
BALANCING INNOVATION AND STANDARDIZATION
In the rapidly evolving AI landscape, there's a natural tension between the need for rapid experimentation and the desire for standardization. While large organizations might benefit from AI gateways for control and security, smaller, agile teams often need the flexibility to quickly switch between models and prompts. Finding the right time to introduce abstractions and standards is a key judgment call, especially in this "wild west" phase of AI development.
THE EVOLVING ROLE OF PROMPT ENGINEERING
While prompt engineering is currently a key skill, it's unlikely to remain a durable differentiator. Instead, the ability to structure ML problems and ask the right questions will likely become more crucial. The operational challenges of LLMs, such as managing latency and handling unpredictable inputs, will continue to demand defensive engineering Socratic methods of inquiry into their capabilities.
ASSESSING CANDIDATE MATURITY AND FIT
Modern young professionals often exhibit a higher degree of maturity, capability, and drive than previous generations. Identifying and nurturing this talent is key. The interview process should simulate collaborative work, allowing candidates to assess the company as much as the company assesses them, particularly important for an emerging field like AI engineering where established playbooks are still being written.
Mentioned in This Episode
●Software & Apps
●Companies
●Organizations
●Concepts
●People Referenced
Common Questions
An effective AI engineer needs strong conventional software engineering skills, a genuine curiosity and enthusiasm for machine learning and language models, and a fault-first mindset to build resilient systems.
Topics
Mentioned in this video
Mentioned in relation to the impact of his downfall on the Effective Altruism (EA) community, prompting reflection and caution.
Head of engineering at Elicit, discussed his transition from traditional VP of Technology roles to AI due to the generational shift in AI/ML capabilities.
Co-founder of Heroku, previously worked on Muse, and currently acts as an internal journalist for Elicit, focusing on supporting James's articles and learning about AI applications.
A company focused on applying language model capabilities to science and literature search.
A programming language used at Elicit for front-end development, sharing types with backend Python code via OpenAPI specs.
Used as an example of a conventional API with predictable latency, contrasted with language model APIs.
A new model released by Anthropic, discussed regarding its capabilities and how to assess them.
A JavaScript library for building user interfaces, mentioned as an example of a technology that eventually led to standardized abstractions after a period of rapid evolution.
A cloud platform as a service founded by Adam Wiggins, which standardized the early development experience.
Mentioned in the context of discussions around standardization and abstractions in the rapidly evolving AI space.
A specification used at Elicit to generate TypeScript types dynamically from Python definitions, facilitating type sharing between backend and frontend.
A philosophy and social movement focused on using evidence and reason to do the most good. Its community is undergoing reflection after the issues related to Sam Bankman-Fried.
More from Latent Space
View all 186 summaries
86 minNVIDIA's AI Engineers: Brev, Dynamo and Agent Inference at Planetary Scale and "Speed of Light"
72 minCursor's Third Era: Cloud Agents — ft. Sam Whitmore, Jonas Nelle, Cursor
77 minWhy Every Agent Needs a Box — Aaron Levie, Box
42 min⚡️ Polsia: Solo Founder Tiny Team from 0 to 1m ARR in 1 month & the future of Self-Running Companies
Found this useful? Build your knowledge library
Get AI-powered summaries of any YouTube video, podcast, or article in seconds. Save them to your personal pods and access them anytime.
Try Summify free