Back to all jobs
IT

Senior Go Software Engineer (AI Evaluation)

United States Contract

About the Role

A high-impact AI evaluation initiative focused on improving the reliability and performance of conversational systems in software engineering contexts. The work centers on analyzing how models generate, reason about, and explain code across varying levels of complexity and real-world use cases.

This opportunity is ideal for experienced engineers with deep expertise in Go and strong problem-solving ability, particularly those comfortable evaluating complex logic, debugging edge cases, and assessing algorithmic correctness. Candidates should be detail-oriented and capable of independently validating technical outputs.

The work involves reviewing and testing AI-generated code, validating correctness through execution, and providing structured feedback on reasoning quality and clarity, where precision and consistency are critical to improving system performance.

What You'll Do

  • Evaluate AI-generated responses to coding and software engineering problems
  • Execute and validate code to verify correctness and performance
  • Identify logical errors, inefficiencies, and edge case failures
  • Annotate outputs with detailed feedback on strengths and weaknesses
  • Assess code readability, maintainability, and algorithmic soundness
  • Perform fact-checking using reliable technical references
  • Apply standardized evaluation frameworks and scoring guidelines
  • Ensure outputs align with expected engineering and conversational standards

Requirements

  • 5+ years of professional experience in software engineering or related fields
  • Strong expertise in Go programming language
  • Ability to solve medium to hard algorithmic problems independently
  • Experience executing, testing, and debugging production-level code
  • Strong understanding of data structures, algorithms, and system design principles
  • High attention to detail in reviewing technical reasoning and outputs
  • Fluent English communication skills
  • Experience using LLMs in coding workflows and understanding their limitations
  • Ability to follow structured evaluation frameworks and guidelines
  • Bachelor’s degree or higher in Computer Science or related discipline
  • Experience contributing to open-source projects with accepted contributions
  • Familiarity with multiple programming languages or paradigms (preferred)
  • Experience in model evaluation, RLHF, or data annotation (preferred)
  • Background in competitive programming or technical assessments (preferred)
  • Experience reviewing code in production environments (preferred)
  • Ability to explain complex technical concepts clearly to varied audiences (preferred)
Application Note: By submitting your profile for this partnered position, our team can quickly review your background and reach out to present you with this specific opportunity or match you with similar AI Training projects.