Ativa os alertas de emprego por e-mail!
A technology services firm is seeking an AI Engineer (Search Engine) for a fully remote, 10-month contract. This role focuses on building and optimizing RAG models for search across large-scale medical documents. Candidates must have 1+ years of experience in search technology and 2+ years of Python development. Strong communication skills are essential, along with the ability to work independently. The position offers a flexible working schedule from 10AM to 6PM BRT.
Job Title : AI Engineer (Search Engine)
About the Company : Insight Global's Client
Type : 10 month extending contract
Compensation : Shared upon initial WhatsApp screening call
Location : Fully Remote
Working Hours : 10AM - 6PM BRT (9AM-5PM EST)
Interview Process : immediate interviews available - 2 rounds can close by end of next week.
Project Overview: We are hiring one AI / LLM Engineers (1 Mid-Level) to join a small, focused team building backend features for an existing Retrieval-Augmented Generation (RAG) service. Your primary focus will be building and optimizing RAG models for search across large-scale medical and scientific documents, including pre-processing and embedding over half a billion documents to ensure they are searchable and contextually accurate. You’ll work on semantic chunking strategies, improving the automated evaluation pipeline, and fine-tuning LLMs for textual RAG use cases. The role involves hands-on experimentation, model development, and backend engineering, with deployments to non-prod environments and collaboration with DevOps for production rollout. While this is a heads-down, individual contributor role, strong communication and collaboration with the team are essential.
We are hiring one AI / LLM Engineers (1 Mid-Level) to join a small, focused team building backend features for an existing Retrieval-Augmented Generation (RAG) service. Your primary focus will be building and optimizing RAG models for search across large-scale medical and scientific documents, including pre-processing and embedding over half a billion documents to ensure they are searchable and contextually accurate. You’ll work on semantic chunking strategies, improving the automated evaluation pipeline, and fine-tuning LLMs for textual RAG use cases. The role involves hands-on experimentation, model development, and backend engineering, with deployments to non-prod environments and collaboration with DevOps for production rollout. While this is a heads-down, individual contributor role, strong communication and collaboration with the team are essential.