Join Project Hermes 2.0, a global initiative focused on enhancing LLM output across a range of domains!
We’re looking for contributors to take part in a fast-paced evaluation task. In this role, you'll review pairs of inputs and AI-generated outputs, then assess the quality of those outputs based on accuracy, relevance, and clarity—following the project’s established guidelines. Your evaluations will directly support improvements to model performance and help shape the future of responsible AI development.
Key Responsibilities:
Review input-output pairs, assessing AI-generated responses for quality and accuracy- Evaluate content based on criteria such as clarity, tone, relevance, and alignment with intent
- Provide structured responses to evaluation questions to support model improvement
- Maintain high attention to detail and quality across all assigned tasks
Requirements:
Native or near-native fluency in English- Strong reading comprehension and writing skills
- Minimum commitment of 20 hours per week
- Bachelor’s degree or relevant work experience in content evaluation, writing, linguistics, or a related field
- High attention to detail and ability to follow detailed guidelines
- Comfortable reviewing content that may include sensitive or potentially harmful material
Slots are filling up fast, and early starters will have the best chance to secure their spot and begin earning sooner!
This role is a project-based opportunity with CrowdGen, where you will join the CrowdGen Community as an Independent Contractor. If selected, you will receive an email from CrowdGen regarding the creation of an account using your application email address. You will need to log in to this account and reset the password, complete the setup requirements, and proceed with your application for this project-based role.
Make an impact on the future of AI – apply today and start contributing from the comfort of your home.
This is a pay-per-unit task, with the pay rate based in USD.
Report job