D

Generative AI Engineer (Data/ML/GenAI)

DATAECONOMY
Full-time
On-site
Overview

About Us

About DATAECONOMY: We are a fast-growing data & analytics company headquartered in Dublin with offices in Dublin, OH, Providence, RI, and an advanced technology center in Hyderabad, India. We are clearly differentiated in the data & analytics space via our suite of solutions, accelerators, frameworks, and thought leadership. Job Description

Job Title

Generative AI Engineer (Data/ML/GenAI)

Jerser City, NJ

Full-time

Job Summary

We’re hiring a Generative AI Engineer with 6 years across Data/ML/GenAI who can design, build, and productionize LLM-powered systems end-to-end. You’ll select and fine-tune models (OpenAI, Anthropic, Google, Meta, open-source), craft robust RAG/agentic workflows (AutoGen, LangGraph, CrewAI, LangChain/LlamaIndex), and ship secure, observable services with FastAPI, Docker, and Kubernetes. You pair strong software engineering with MLOps/LLMOps rigor—evaluation, monitoring, safety/guardrails, and cost/latency optimization.

Responsibilities

Solution architecture: Own E2E design for chat/agents, structured generation, summarization/classification, and workflow automation. Choose the right model vs. non-LLM alternatives and justify trade-offs. Prompting & tuning: Build prompt stacks (system/task/tool), synthetic data pipelines, and fine-tune or LoRA adapters; apply instruction tuning/RLHF where warranted. Agentic systems: Implement multi-agent/tool-calling workflows using AutoGen, LangGraph, CrewAI (state management, retries, tool safety, fallbacks, grounding). RAG at scale: Stand up retrieval stacks with vector DBs (Pinecone/Faiss/Weaviate/pgvector), chunking and citation strategies, reranking, and caching; enforce traceability. APIs & deployment: Ship FastAPI services, containerize (Docker), orchestrate (Kubernetes/Cloud Run), wire CI/CD and IaC; design SLAs/SLOs for reliability and cost. LLMOps & observability: Instrument evals (unit/regression/AB), add tracing and metrics (Langfuse, LangSmith, OpenTelemetry), and manage model/version registries (MLflow/W&B). Safety & governance: Implement guardrails (prompt injection/PII/toxicity), policy filters (Bedrock Guardrails/Azure AI Content Safety/OpenAI Moderation), access controls, and compliance logging. Data & pipelines: Build/maintain data ingestion, cleansing, and labeling workflows for model/retrieval corpora; ensure schema/version governance. Performance & cost: Optimize with batching, streaming, JSON-schema/function calling, tool-use, speculative decoding/KV caching, and token budgets. Collaboration & mentoring: Partner with product/engineering/DS; review designs/PRs, mentor juniors, and drive best practices/playbooks. Qualifications

Preferred Qualifications Agent ecosystems: Deeper experience with multi-agent planning/execution, tool catalogs, and failure-mode design. Search & data stores: Experience with pgvector/Elasticsearch/OpenSearch; comfort with relational/NoSQL/graph stores. Advanced evals: Human-in-the-loop pipelines, golden sets, regression suites, and cost/quality dashboards. Open-source & thought leadership: OSS contributions, publications, talks, or a strong portfolio demonstrating GenAI craftsmanship. Nice to Have Eventing & rate limiting: Redis/Celery, task queues, and concurrency controls for bursty LLM traffic. Enterprise integrations: Experience with API gateways (e.g., MuleSoft), authN/Z, and vendor compliance reviews. Domain experience:

Prior work in data-heavy or regulated domains (finance/health/gov) with auditable GenAI outputs. Requirements Experience: 6 years across Data/ML/GenAI, with 1–2 years designing and shipping LLM or GenAI apps to production. Languages & APIs: Strong Python and FastAPI; proven experience building secure, reliable REST services and integrations. Models & frameworks: Hands-on with OpenAI/Anthropic/Gemini/Llama families and at least two of: AutoGen, LangGraph, CrewAI, LangChain, LlamaIndex, Transformers. RAG & retrieval: Practical experience implementing vector search and reranking, plus offline/online evals (e.g., RAGAS, promptfoo, custom harnesses). Cloud & DevOps: Docker, Kubernetes (or managed equivalents), and one major cloud (AWS/Azure/GCP); CI/CD and secrets management. Observability: Familiarity with tracing/metrics tools (e.g., Langfuse, LangSmith, OpenTelemetry) and setting SLIs/SLOs. Security & governance: Working knowledge of data privacy, PII handling, content safety, and policy/controls for enterprise deployments. Communication: Clear technical writing and cross-functional collaboration; ability to translate business goals into architecture and milestones.

#J-18808-Ljbffr