AI Engineer (Generative AI / MLOps / AI Agents)
Diamondpick
Job Title:
AI Engineer (Generative AI / MLOps / AI Agents)
Location : Warren NJ- Hybrid role
Key Responsibilities
Generative AI & LLM Engineering
Design, fine-tune, and deploy
Large Language Models (LLMs)
for insurance-specific use cases including document intelligence, claims summarization, policy interpretation, and underwriting Q&A.
Build
Retrieval-Augmented Generation (RAG)
pipelines using vector databases (e.g., Azure AI Search, Pinecone, ChromaDB) to ground LLM outputs in enterprise knowledge bases.
Develop
prompt engineering frameworks
and systematic evaluation pipelines to ensure LLM output quality, consistency, and safety in regulated insurance contexts.
Integrate LLM capabilities with internal data platforms via
LangChain, LlamaIndex, or Semantic Kernel .
Evaluate and benchmark foundational models (OpenAI GPT-4o, Azure OpenAI, Claude, Mistral, Llama) against insurance-specific tasks to guide platform selection.
AI Agents & Automation
Architect and implement
autonomous AI agents
capable of multi-step reasoning, tool use, and decision-making for workflows such as FNOL triage, claims routing, policy lookup, and compliance checks.
Build agentic frameworks using patterns such as
ReAct, Chain-of-Thought, and Tool-Augmented Agents
to handle complex, multi-turn insurance workflows.
Design
human-in-the-loop (HITL)
checkpoints and escalation logic to ensure AI agents operate within defined risk and compliance boundaries.
Integrate agents with internal APIs, data platforms, and enterprise systems using orchestration tools such as
Azure Logic Apps, Apache Airflow, or Databricks Workflows .
Develop guardrails, monitoring, and audit logging for all deployed agents to meet regulatory and governance standards.
MLOps & Model Deployment
Build and maintain
end-to-end MLOps pipelines
covering model training, versioning, validation, deployment, and monitoring using
MLflow, Azure ML, and Databricks .
Implement
CI/CD pipelines for ML models
using Azure DevOps or GitHub Actions, enabling reliable, repeatable model releases.
Deploy models as
REST APIs or batch inference services
on Azure Kubernetes Service (AKS) or Azure Container Apps, ensuring scalability and low-latency response.
Establish
model monitoring frameworks
to detect data drift, model degradation, and prediction anomalies in production.
Manage the
model registry and lineage tracking
to maintain governance and auditability of all AI assets.
Collaborate with data engineering teams to ensure feature pipelines are production-grade, versioned, and integrated with the
Feature Store
on Databricks or Azure ML.
Collaboration & Delivery
Work closely with business analysts, actuaries, underwriters, and claims professionals to translate domain requirements into AI solution designs.
Participate in
Agile/Scrum ceremonies
including sprint planning, standups, and retrospectives as an active delivery contributor.
Produce clear, well-structured
technical documentation
including solution designs, API specs, model cards, and deployment runbooks.
Mentor junior engineers and contribute to internal AI engineering best practices and standards
Experience
3-5 years
of professional experience in AI/ML engineering, with demonstrated delivery of
production-grade AI systems .
Hands-on experience building and deploying
LLM-powered applications
using frameworks such as LangChain, LlamaIndex, or Semantic Kernel.
Proven experience implementing
MLOps pipelines
in cloud environments (Azure preferred).
Experience developing
AI agents or automation workflows
using agentic frameworks.
Prior experience in
financial services, insurance, or regulated industries
is strongly preferred.