Enable job alerts via email!

Generative AI Data Engineer - Remote

The Hartford

Frisco (TX)

Remote

USD 125,000 - 189,000

Full time

30+ days ago

Boost your interview chances

Create a job specific, tailored resume for higher success rate.

Job summary

An established industry player is seeking a Staff Data Engineer to join their innovative Generative AI team. In this pivotal role, you will design and build fault-tolerant infrastructure that supports cutting-edge Generative AI applications. You'll collaborate closely with cross-functional teams, ensuring the transition from research to scalable solutions. This hybrid position offers the chance to work in a dynamic environment, where your contributions will directly influence the future of technology. If you are passionate about data engineering and eager to drive innovation, this opportunity is perfect for you.

Qualifications

  • 4+ years of AWS experience and 8+ years in data-intensive solutions.
  • Extensive programming experience in Python and Java.

Responsibilities

  • Design and build fault-tolerant infrastructure for Generative AI.
  • Collaborate with cross-functional teams to execute essential functionalities.

Skills

Problem-solving skills
Excellent communication skills
Experience with CI/CD pipelines
Distributed computing
AWS cloud
Data engineering
Programming in Python
Programming in Java

Education

Bachelor's degree in Computer Science

Tools

AWS OpenSearch
Terraform
CloudFormation
Jenkins
GitHub Actions
SageMaker
ECS/Docker

Job description

Staff Data Engineer - GE07CE

We’re determined to make a difference and are proud to be an insurance company that goes well beyond coverages and policies. Working here means having every opportunity to achieve your goals – and to help others accomplish theirs, too. Join our team as we help shape the future.

At the Hartford, we are seeking a GEN AI Data Engineer who is responsible for building fault-tolerant infrastructure to support the Generative AI applications, and also designing, developing, and deploying data pipelines to solve complex problems and drive innovation at scale.

We are founding a dedicated Generative AI platform engineering team to build our internal developer platform and are looking for an experienced Staff Data Platform Engineer - Generative AI, to help us build the foundation of our Generative AI capability. You will work on a wide range of initiatives, whether that’s building ETL pipeline, training a retrieval re-ranker, working with the DevSecOps team to build the CICD pipeline, or designing a Generative AI Infrastructure that conforms to our strict security standards/guardrails, or working with the data science team in their pursuit of improving the accuracy of the LLM models.

The Generative AI team is comprised of a multiple cross-functional group that works in unison and ensures a sound move from our research activities to scalable solutions. You will collaborate closely with our cloud, security, infrastructure, enterprise architecture, and data science teams to conceive and execute essential functionalities.

This role has a Hybrid work schedule. Candidates who live near one of our office locations (Hartford, CT, Charlotte, NC, Chicago, IL, Columbus, OH) will have the expectation of working in an office 3 days a week (Tuesday through Thursday).

Candidates must be eligible to work in the US without sponsorship now or in the future

Responsibilities:

  • Design and build fault-tolerant infrastructure to support the Generative AI Ref architecture (RAG, Summarization, Agent etc).
  • Ensure code is delivered without vulnerabilities by enforcing engineering practices, code scanning, etc.
  • Build and maintain IAC (terraform/Cloud Formation), CICD (Jenkins) scripts, CodePipeline, uDeploy, & GitHub Actions.
  • Partner with our shared service teams like Architecture, Cloud, Security, etc to design and implement platform solutions.
  • Collaborate with the DS team to develop a self-service internal developer Generative AI platform.
  • Design and build the Data ingestion pipeline for Finetuning LLM Models.
  • Create templates (Architecture As Code) implementing Ref architecture application’s topology.
  • Build a feedback system using HITL for Supervised finetuning.

Qualifications:

  • Bachelor's degree in Computer Science, Computer Engineering, or a technical field.
  • 4+ years of experience with AWS cloud.
  • At least 8 years of experience designing and building data-intensive solutions using distributed computing.
  • 8+ years building and shipping software and/or platform infrastructure solutions for enterprises.
  • Experience with CI/CD pipelines, Automated Testing, Automated Deployments, Agile methodologies, Unit Testing, and Integration Testing tools.
  • Experience with building scalable serverless application (real-time / batch) on AWS stack (Lambda + step function).
  • Knowledge of distributed NoSQL database systems.
  • Experience with data engineering, ETL technology, and conversation UX is a plus.
  • Experience with HPCs, vector embedding, and Hybrid/Semantic search technologies.
  • Experience with AWS OpenSearch, Step/Lambda Functions, SageMaker, API Gateways, ECS/Docker is a plus.
  • Proficiency in customization techniques across various stages of the RAG pipeline, including model fine-tuning, retrieval re-ranking, and hierarchical navigable small-world graph (HNSW) is a plus.
  • Strong proficiency in embeddings, ANN/KNN, vector stores, database optimization, & performance tuning.
  • Extensive programming experience with Python, Java.
  • Experience with LLM orchestration frameworks like Langchain, LlamaIndex etc.
  • Foundational understanding of Natural Language Processing, and Deep Learning.
  • Excellent problem-solving skills and the ability to work in a collaborative team environment.
  • Excellent communication skills.
  • Candidate must be authorized to work in the US without company sponsorship. The company will not support the STEM OPT I-983 Training Plan endorsement for this position.

Compensation

The listed annualized base pay range is primarily based on analysis of similar positions in the external market. Actual base pay could vary and may be above or below the listed range based on factors including but not limited to performance, proficiency and demonstration of competencies required for the role. The base pay is just one component of The Hartford’s total compensation package for employees. Other rewards may include short-term or annual bonuses, long-term incentives, and on-the-spot recognition. The annualized base pay range for this role is:

$125,760 - $188,640

Equal Opportunity Employer/Females/Minorities/Veterans/Disability/Sexual Orientation/Gender Identity or Expression/Religion/Age

About Us | Culture & Employee Insights | Diversity, Equity and Inclusion | Benefits

Get your free, confidential resume review.
or drag and drop a PDF, DOC, DOCX, ODT, or PAGES file up to 5MB.

Similar jobs

Sr Machine Learning Engineer - REMOTE

RealPage, Inc.

Richardson

Remote

USD 120,000 - 160,000

Today
Be an early applicant

Principal Data Scientist - Generative AI, Machine Learning, Python, R - Remote

Lensa

Fort Worth

Remote

USD 120,000 - 180,000

Today
Be an early applicant

Chief Data Scientist

Blue Yonder

Town of Texas

Remote

USD 186,000 - 285,000

7 days ago
Be an early applicant

Sr. Machine Learning Engineer

Mr. Cooper Group Inc.

Lewisville

Remote

USD 90,000 - 150,000

2 days ago
Be an early applicant

Data Scientist (AI)

Wiraa

Virginia

Remote

USD 146,000 - 184,000

2 days ago
Be an early applicant

Sr. Data Scientist

Apkudo

Dallas

Remote

USD 90,000 - 150,000

8 days ago

GenAI Systems/ML Data Engineer

Motion Recruitment

New Jersey

Remote

USD 72,000 - 205,000

Today
Be an early applicant

LLM Data Engineer | United States | Fully Remote

Halo Media

Tallahassee

Remote

USD 90,000 - 150,000

Today
Be an early applicant

Artificial Intelligence (AI) Data Scientist

General Dynamics Information Technology

Fairfax

Remote

USD 146,000 - 184,000

7 days ago
Be an early applicant