Job Search and Career Advice Platform

Enable job alerts via email!

Arabic Trust & Safety Data Trainer

SME Careers

Remote

SAR 48,000 - 120,000

Part time

Yesterday
Be an early applicant

Generate a tailored resume in minutes

Land an interview and earn more. Learn more

Job summary

A fast-growing AI Data Services company is seeking a fully remote contractor to review AI-generated responses and ensure safety and accuracy. Responsibilities include curating training examples in both Arabic and English, evaluating model responses against safety policies, and continuously auditing model behavior. Candidates must have near-native Arabic proficiency and at least C1 English proficiency. A background in Trust & Safety, emotional resilience, and familiarity with AI training tools is highly valued.

Qualifications

  • Bachelor’s degree or higher in Communications, Linguistics, Psychology, Law/Policy, or equivalent.
  • Near-native or native Arabic proficiency for high-precision labeling.
  • Experience in Trust & Safety or content moderation.
  • LLM red teaming experience is essential.
  • Emotional resilience to handle explicit content.

Responsibilities

  • Curate and label training examples in English and Arabic.
  • Review and document model responses against safety policies.
  • Audit model behavior and propose clearer decision rules.

Skills

Fluent Arabic
C1 English proficiency
Trust & Safety experience
LLM red teaming experience
Excellent judgment

Education

Bachelor’s degree in a relevant field

Tools

Perplexity
Gemini
ChatGPT
Job description

In this hourly-paid, fully remote contractor role, you will review AI-generated responses and/or generate safety-focused evaluation content, assess reasoning quality and step-by-step problem-solving, and provide expert feedback so outputs are accurate, safe, and clearly explained. You will annotate and evaluate content in English and Arabic; you must be fluent/proficient (near-native or native) in Arabic and able to apply Minimum C1 English to interpret prompts and policies precisely. Your annotations on these explicit tasks will be used to prevent the Large Language Model from generating unintentional or adversarial, toxic, or unsafe outputs. The types of explicit content you may be exposed to may include, but are not limited to, those of a sexual, violent, or psychologically disturbing nature. This role is with SME Careers, a fast-growing AI Data Services company and a subsidiary of SuperAnnotate that provides AI training data for many of the world’s largest AI companies and foundation-model labs, helping improve the world’s premier AI models.

Key Responsibilities
  • Curate and label safety-focused training examples (including adversarial/red-team cases) in English and Arabic that probe model behavior across hate/harassment, sexual content, self-harm, violence, bias, illegal services, malicious activity, malicious code, and misinformation—capturing nuance and intent with Minimum C1 English and near-native French proficiency.
  • Review, score, and compare multiple model responses against safety policy and quality rubrics, documenting why an output is safe/unsafe and identifying failure modes such as evasion, normalization, escalation, or procedural enablement.
  • Continuously stress-test and audit model behavior for policy gaps and edge cases; flag ambiguous scenarios, propose clearer decision rules, and help maintain consistent annotation standards across reviewers.
Your Profile
  • Bachelor’s degree or higher in a relevant field (e.g., Communications, Linguistics, Psychology, Law/Policy, Security Studies) or equivalent professional experience.
  • Near-native or native Arabic proficiency (reading/writing) for high-precision safety labeling and cultural-linguistic nuance.
  • Minimum C1 English proficiency (reading/writing) for policy interpretation, prompt understanding, and consistent documentation.
  • Experience in Trust & Safety, content moderation, policy enforcement, risk operations, investigations, or safety evaluation work.
  • LLM red teaming experience is a must (proven ability to probe safety boundaries and document adversarial patterns).
  • Strong knowledge of safety domains: Hate & Harassment, Sexual content, Suicide & Self-Harm, Violence, Bias, Illegal goods/services, malicious activities, malicious code, and deliberate misinformation.
  • Emotional resilience: an understanding that this role requires annotating texts that contain unsafe, explicit, and/or toxic content, including content of a sexual, violent, or psychologically disturbing nature.
  • Excellent judgment under ambiguity, with the ability to apply written policies consistently and explain decisions succinctly.
  • Comfort working as an hourly contractor: dependable throughput, clear documentation, and responsiveness across time zones.
  • Previous experience with AI data training / annotation / evaluation is preferred.
  • Strong hands‑on experience using tools like Perplexity, Gemini, ChatGPT and others.
Get your free, confidential resume review.
or drag and drop a PDF, DOC, DOCX, ODT, or PAGES file up to 5MB.