Remote Otter LogoRemoteOtter

Chinese Speakers — Contribute to AI LLM Evaluation - Remote

Posted 21 weeks ago
All others
Contract
Worldwide

Overview

Join Project Spearmint — a multilingual AI response evaluation project focused on reviewing large language model (LLM) outputs in different languages!

In Short

  • Shape the future of AI in your native language.
  • Work from home with flexible hours.
  • Contribute to advancing cutting-edge language model development.
  • Review short, pre-segmented datasets.
  • Evaluate model-generated replies based on Tone or Fluency.
  • Read a user prompt and two model replies, then rate each using a five-point scale.
  • Provide short rationales for extreme ratings.
  • Determine if replies are helpful, engaging, fair, and appropriately formal.
  • Assess grammatical accuracy, clarity, coherence, and natural flow.
  • This is a project-based opportunity with CrowdGen.

Requirements

  • Native-level fluency in Persian language.
  • Strong comprehension of English.
  • Residing in United States.
  • Can commit to working 25 hours per week.

Benefits

  • Work from home.
  • Flexible hours.
  • Make an impact on the future of AI.
  • Project-based opportunity with payment details provided upon selection.
Appen logo

Appen

Appen is a global leader in AI enablement, specializing in critical tasks such as model improvement, supervision, and evaluation. With over 25 years of experience, Appen leverages a diverse crowd of more than one million skilled contractors from 130 countries, speaking over 180 languages and dialects. The company utilizes an advanced AI-assisted data annotation platform to collect and label various data types, including images, text, speech, audio, and video. Trusted by the world's largest technology companies, Appen plays a crucial role in building and enhancing innovative AI systems across multiple sectors, including automotive, finance, retail, healthcare, and government. Committed to creating an inclusive and diverse workplace, Appen fosters a learn-it-all culture that values growth, innovation, and customer obsession.

Share This Job!

Save This Job!

Similar Jobs:

Appen logo

Czech Speakers — AI LLM Evaluation Contributor - Remote

Appen

21 weeks ago

Join Project Spearmint to evaluate AI language model outputs in Czech from the comfort of your home.

Czech Republic
Contract
All others
CrowdGen by Appen logo

Chinese–Tamil Translation Evaluator - Remote

CrowdGen by Appen

4 weeks ago

Join as a Chinese–Tamil Translation Evaluator to ensure high-quality translations.

India
Contract
All others
Welocalize logo

Digital Content Evaluation Specialist - Traditional Chinese (Taiwan) - Remote

Welocalize

31 weeks ago

Join Welo Data as a Digital Content Evaluation Specialist, reviewing and enhancing digital content quality for AI applications.

TW
Freelance
All others
$10.00/hour
RWS logo

AI Data Services Evaluator - Japanese Speakers - Remote

RWS

16 weeks ago

Join our voice AI evaluation project as a native Korean speaker with Japanese language skills.

South Korea
Freelance
All others

Binance is seeking an AI Evaluation Specialist to design and manage evaluation frameworks for AI agents across various domains.

HK
Full-time
Software Development