AI product eval/ applied LLM eval/Human-in-the-Loop Evaluation / Annotation Advisor(1-3hr pre week/potential co founder)

Posted 2026-05-06
Remote, USA Full-time Immediate Start

Looking for a Part-Time Evaluation Advisor (human annotation/llm eval)

The direction method is already validated by multiple ai leaders experts and ux researcher, there’s early buyer interest /pre pilot from big tech ai team , and the immediate goal is to turn the current build into a small, stable, pilot-ready product.

We’re building an early evaluation and workflow product for AI teams. The current focus is helping teams structure and operationalize real-world failure cases, especially in scenarios where an assistant recommends too early, becomes overconfident, or fails to verify what matters before responding. The short-term wedge is a lightweight regression and review workflow. The longer-term opportunity is much bigger: infrastructure for how AI systems are tested, reviewed, and controlled in production decision flows.

    Looking for someone who can help with:
  • human annotation design
  • label guideline writing
  • evaluation schema
  • translating
    Best fit: someone with experience in human eval, annotation design, ranking/review quality, AI eval, or related areas. Bonus if you’ve worked on shopping, marketplace, trust, search, ranking, or agent systems.
  • Part-time / advisor to start.
    Best fit: someone with experience in human eval, annotation design, ranking/review quality, AI eval, or related areas. Bonus if you’ve worked on shopping, marketplace, trust, search, ranking, or agent systems.
  • Part-time / advisor to start.

offer equity or cash

Similar Jobs

Back to Job Board