Remote Otter LogoRemoteOtter

English LLM Evaluator - AI Output Review - Remote

Posted 1 week ago
All others
Contract
USA

Overview

Join Project Hermes 2.0, a global initiative focused on enhancing LLM output across a range of domains!

In Short

  • Review input-output pairs, assessing AI-generated responses for quality and accuracy
  • Evaluate content based on criteria such as clarity, tone, relevance, and alignment with intent
  • Provide structured responses to evaluation questions to support model improvement
  • Maintain high attention to detail and quality across all assigned tasks
  • Native or near-native fluency in English required
  • Strong reading comprehension and writing skills needed
  • Minimum commitment of 20 hours per week
  • Bachelor’s degree or relevant work experience in content evaluation, writing, linguistics, or a related field
  • Comfortable reviewing content that may include sensitive or potentially harmful material
  • Project-based opportunity with CrowdGen

Requirements

  • Native or near-native fluency in English
  • Strong reading comprehension and writing skills
  • Minimum commitment of 20 hours per week
  • Bachelor’s degree or relevant work experience in content evaluation, writing, linguistics, or a related field
  • High attention to detail and ability to follow detailed guidelines
  • Comfortable reviewing content that may include sensitive or potentially harmful material

Benefits

  • Make an impact on the future of AI
  • Start contributing from the comfort of your home
  • Flexible working hours
  • Opportunity to join the CrowdGen Community
  • Chance to secure a spot quickly as slots are filling up fast
Appen logo

Appen

Appen is a global leader in AI enablement, specializing in critical tasks such as model improvement, supervision, and evaluation. With over 25 years of experience, Appen leverages a diverse crowd of more than one million skilled contractors from 130 countries, speaking over 180 languages and dialects. The company utilizes an advanced AI-assisted data annotation platform to collect and label various data types, including images, text, speech, audio, and video. Trusted by the world's largest technology companies, Appen plays a crucial role in building and enhancing innovative AI systems across multiple sectors, including automotive, finance, retail, healthcare, and government. Committed to creating an inclusive and diverse workplace, Appen fosters a learn-it-all culture that values growth, innovation, and customer obsession.

Share This Job!

Save This Job!

Similar Jobs:

CrowdGen by Appen logo

Music Evaluator – English - Remote

CrowdGen by Appen

11 weeks ago

Join the CrowdGen LLM Pre-Qualification Community to gain access to high-paying LLM projects and enhance your skills.

USA
Freelance
All others
CrowdGen by Appen logo

Danish LLM Evaluator - Remote

CrowdGen by Appen

8 weeks ago

Join Project Spearmint as a Danish LLM Evaluator to assess AI model outputs focused on Tone and Fluency.

Denmark
Contract
All others
Welocalize logo

AI Response Evaluator - Remote

Welocalize

41 weeks ago

Join us as an AI Response Evaluator to enhance AI technologies and gain valuable experience in a flexible freelance role.

Worldwide
Freelance
All others
15USD/hour
Welocalize logo

AI Response Evaluator - Remote

Welocalize

42 weeks ago

Join us as an AI Response Evaluator to enhance AI technologies and gain valuable experience in a flexible freelance role.

Worldwide
Freelance
All others
Welocalize logo

AI Response Evaluator - Remote

Welocalize

42 weeks ago

Join us as an AI Response Evaluator to enhance AI technologies and gain valuable experience in a flexible freelance role.

Worldwide
Freelance
All others