Careers

Lead Validation Engineer

About us

AIM builds autonomy for the real world - robots that move mountains. Our systems integrate software, electronics, mechanical systems, perception, and mission-critical infrastructure into rugged, safety-critical machines operating on construction and mining sites globally.

Our machines operate in dynamic, unpredictable environments where performance, reliability, and safety must hold under real-world conditions - not just in simulation. Delivering autonomy at scale requires not only building systems, but proving they work consistently, safely, and reliably in the field.

Validation is how we make that real.

About You

You are a systems-oriented engineer who cares deeply about whether things actually work - not just whether they were designed to work.

You:

  • Think in terms of system behavior, not isolated components

  • Build mechanisms that scale, not manual processes

  • Use data and real-world evidence to drive decisions

  • Have strong judgment in ambiguous situations with incomplete information

  • Hold a high bar for reliability, performance, and safety

You are equally comfortable:

  • Designing validation strategies and frameworks

  • Building infrastructure and tooling

  • Debugging failures across hardware, software, and AI systems

  • Working in the field to understand real-world conditions

You take ownership of outcomes - ensuring validation directly improves system performance, safety, and reliability.

About Us Together

We’re building autonomous systems that must perform under real-world conditions - and we validate them by:

  • Ensuring end-to-end system behavior across electrical, mechanical, software, and AI subsystems

  • Validating performance under harsh environments: dust, vibration, temperature extremes, and dynamic terrain

  • Ensuring reliability across long-duration operation in real jobsites

  • Testing complex interactions across autonomy, controls, perception, and hardware systems

  • Capturing real-world edge cases and converting them into repeatable validation scenarios

We move fast, test rigorously, and build validation systems that scale with the fleet.

Why this role exists

Autonomous systems fail at system boundaries - in the interactions between perception, planning, controls, hardware, and real-world environments.

While engineering teams validate their own components, the Lead Validation Engineer ensures that the integrated system actually works in reality.

This role exists to:

  • Close the gap between engineering intent and real-world behavior

  • Build the validation system that scales with the company

  • Provide an independent signal of system readiness

This is not a QA role. This role builds the system that makes quality measurable, enforceable, and scalable across AIM.

What You Will Own

As the Lead Validation Engineer, you will own the system that ensures AIM’s autonomous machines work reliably, safely, and predictably in the real world. You are accountable not just for validation activities, but for making validation scalable, measurable, and deeply integrated into how we build and deploy systems.

Validation System Ownership

Own how validation is performed across AIM.

  • Define the end-to-end validation strategy across:

    • simulation

    • software-in-the-loop (SIL)

    • hardware-in-the-loop (HIL)

    • proving ground (PG)

    • field deployments

  • Establish the test pyramid, shifting validation earlier in development

  • Define validation coverage across:

    • system behaviors

    • environmental conditions

    • Operational Design Domain (ODD)

  • Ensure validation scales through automation, replay systems, and data-driven testing

Validation Infrastructure & Tooling

Build systems that enable engineers to validate their own work.

  • Develop:

    • automated test frameworks

    • simulation and replay systems

    • data capture and labeling pipelines

    • validation dashboards and reporting systems

  • Integrate validation into CI/CD and development workflows

  • Enable self-service validation across AI, software, and hardware teams

System-Level Validation

Ensure the integrated system behaves correctly across domains.

  • Validate interactions across:

    • perception

    • planning

    • controls

    • hardware systems

    • operator interfaces

  • Identify emergent failure modes not visible at the component level

  • Design validation scenarios that reflect real-world complexity

Real-World Validation & Field Feedback

Ensure validation reflects actual operating conditions.

  • Define validation strategies for PG and field deployments

  • Build systems to:

    • capture real-world data

    • reproduce field failures

    • convert issues into repeatable tests

  • Ensure field learnings feed back into validation systems

Release Gating & Readiness

Define and enforce system readiness.

  • Define validation gates for:

    • feature releases

    • system releases

    • customer deployments

  • Provide validation input into ORRs and release decisions

  • Ensure systems meet performance, reliability, and safety thresholds

Acts as an independent signal of readiness, not a bottleneck.

Metrics, Observability & Coverage

Make validation measurable.

  • Define metrics including:

    • coverage across ODD conditions

    • failure and escape rates

    • MTBI (Mean Time Between Interventions)

    • validation effectiveness

  • Build dashboards and reporting systems

  • Identify and close validation gaps

Incident Analysis & Continuous Improvement

Ensure failures improve the system.

  • Partner on root cause analysis (CoE)

  • Identify validation gaps from incidents and near misses

  • Feed learnings back into validation frameworks and test coverage

Cross-Functional Leadership

Drive validation across engineering teams.

  • Define validation expectations across AI, software, and hardware

  • Ensure teams own validation of their components

  • Drive alignment on validation practices and standards

Validation Culture & Mechanisms

Establish validation as a core engineering discipline.

  • Define validation standards and best practices

  • Establish validation reviews and processes

  • Train teams on validation methodologies

  • Ensure validation is treated as an engineering output

Decision Authority

Serve as the technical authority on validation.

  • Provide independent input on system readiness

  • Gate releases when validation is insufficient

  • Escalate validation risks to leadership

  • Challenge assumptions where system behavior is not proven

Qualifications

Basic Qualifications

  • Bachelor’s degree in Mechanical, Electrical, Software, Systems, Robotics, or related field (or equivalent experience)

  • Experience in system-level validation, test engineering, or reliability engineering

  • Experience validating complex systems integrating hardware, software, and autonomy

  • Strong experience designing validation strategies and frameworks

  • Hands-on experience debugging system-level failures

Preferred Qualifications

  • Experience in robotics, autonomy, or safety-critical systems

  • Experience building validation infrastructure, simulation, or replay systems

  • Experience with functional safety validation

  • Experience scaling validation from prototype to production

  • Experience with real-world field deployments

How You’ll Stand Out

  • You build validation systems that scale, not manual processes

  • You identify failure modes before they reach the field

  • You translate real-world behavior into testable systems

  • You balance rigor with speed in ambiguous environments

  • You turn validation into a competitive advantage

  • You drive teams to confront reality and improve system performance

Details

Department

Engineering

Location

Seattle

Type

Full Time

Apply Now