[< BACK]
// POSTED: May 3, 2026

Peach Pilot — Principal QA Engineer (AI Systems & Platform) Remote — Latin America

APPLY NOW

Peach Pilot — Principal QA Engineer (AI Systems & Platform) Remote — Latin America | Full-Time Contract | US Eastern Timezone Overlap Required (5+ hours daily)

The Mission: Trust Has to Be Earned — Every Release

95% of enterprise AI pilots fail — not because the technology is broken, but because users don't trust it. At Peach Pilot we are building an enterprise AI operating system where trust is the product. That means every feature we ship must work exactly as the user expects, every time. One broken interaction at the wrong moment can undo months of adoption. You are the last line of defense before our platform reaches a CFO's desk.

Peach Pilot is a funded US-based AI startup building an enterprise AI operating system for business leaders. We are closing the AI trust gap — making powerful AI feel effortless and reliable for the people who run companies, not just the engineers who build software.

We are an early-stage founding team moving fast and hiring remotely across Latin America.

The Role

This is a hands-on, high-ownership role. You will build and own the QA function at Peach Pilot — writing test code, designing eval pipelines, and setting the quality bar as we move from early-stage development into full production and enterprise deployment. We are not looking for someone who manages spreadsheets and delegates everything. We are looking for someone who can do the work, knows what good looks like, and raises the bar across the entire engineering team.

This is a fully remote contract role based in Latin America. As the company scales, there is a path to a larger leadership role.  For now the focus is getting the product right.

You will work directly with the US-based founding engineering team and must be available during US Eastern business hours with a minimum of 5 hours of daily overlap.

The Challenge: QA for AI is a Different Problem

Traditional QA assumes deterministic outputs. LLMs don't give you that. You will be building a quality function from scratch in an environment where:

What You Will Own & Build

Who You Are

Even Better If

The Stack You'll Test Against

AI/LLM: Anthropic Claude, OpenAI GPT-4o, xAI Grok, Gemini Frontend: React/Next.js, TypeScript, Tailwind CSS Backend: Python, Node.js/TypeScript (FastAPI/Express) Data & Graph: Neo4j, Snowflake, Azure Cosmos DB, Azure AI Search Infrastructure: Azure (Functions, Key Vault), CI/CD pipelines Visualization: Plotly, D3, Recharts, Mermaid

Compensation

Competitive contractor rate commensurate with experience. Paid monthly via Deel in USD.

The Clincher

Tell us about a quality failure — one you caught before it shipped, or one that got through. What did you build or change after it, and how did you make sure your team could catch the next one without you?

 

Interested in this role?Apply on iHire