
Enable job alerts via email!
Generate a tailored resume in minutes
Land an interview and earn more. Learn more
A leading technology firm in Singapore is searching for an LLM Engineer to design and implement algorithms that steer LLM output. Responsibilities include optimizing inference architecture, collaborating with interdisciplinary teams, and investigating advanced techniques in LLM alignment. The ideal candidate holds an MS or PhD in relevant fields and possesses strong Python programming skills and experience with ML frameworks. Strong analytical and software engineering skills are essential for success in this role.
Design and implement algorithms that steer or direct LLM output such as: controlling tone/voice/persona, enforcing constraints (e.g., style, accuracy, safe behavior), prompting/control flow, tool‑use by the model, and multi‑step reasoning workflows.
Work on post‑training processes such as supervised fine‑tuning (SFT), reinforcement learning from human feedback (RLHF), reward modelling and alignment techniques to shape model behavior. For example, one job listing describes “advanced post‑training of large language models (e.g. SFT, RLHF/RLAIF…)”.
Build and maintain evaluation & benchmark pipelines for behavior: e.g., factuality, bias/ethics, latency, throughput, tool‑calls, JSON‑schema compliance, multi‑round interaction, hallucination detection. For instance: “Build offline and live eval pipelines for alignment, factuality, grounding, and hallucinations.”
Optimize inference architecture and serve models at scale: e.g., context window management, checkpoint routing, efficient serving, adapter/LoRA training, distributed training/inference frameworks. As one role states: “Design, deploy, and operate Model Context Protocol (MCP) servers … manage context windows.”
Collaborate with research, product, safety, and infrastructure teams to integrate directed‑LLM capabilities into product experiences—ensuring the model’s behavior meets user intent, product goals, and safety/ethics standards.
Investigate and adopt state‑of‑the‑art techniques in LLM alignment, prompting/steering, retrieval‑augmented generation (RAG), multi‑modal integration, agentic behavior, and emergent capabilities.