Bestkaam Logo
The Hartford India Logo

Applied AI Scientist

Actively Reviewing the Applications

The Hartford India

India, Telangana, Hyderabad Full-Time On-site
Posted 9 hours ago Apply by June 14, 2026

Job Description

The Applied AI Scientist is responsible for designing, building, and deploying advanced AI solutions—spanning traditional machine learning, generative AI, and agentic workflows—that address complex business and regulatory needs. The role partners closely with stakeholders across Product, Operations, Underwriting, Claims, Legal, Compliance, and Risk, while working in tight alignment with the broader Technology organization, including AI engineering, data, platform, architecture, security, and IT teams, to deliver scalable, secure, and productionready solutions. Key responsibilities include developing RAG pipelines, assistants, forecasting and classification models, regulatory intelligence and filing automation, and domainspecific knowledge bases, with full ownership of the AI lifecycle and strong emphasis on responsible AI, governance, and compliancebydesign. Successful candidates combine deep applied ML and GenAI expertise, strong Python and cloud skills, rigorous evaluation and MLOps practices, and the ability to translate technical outcomes into clear business value‑ready solutions. Key responsibilities include developing RAG pipelines, assistants, forecasting and classification models, regulatory intelligence and filing automation, and domain‑specific knowledge bases, with full ownership of the AI lifecycle and strong emphasis on responsible AI, governance, and compliance‑by‑design. Successful candidates combine deep applied ML and GenAI expertise, strong Python and cloud skills, rigorous evaluation and

Key Responsibilities

  • Design & Deliver AI Solutions: Build statistical, ML, and generative/agentic AI solutions spanning RAG pipelines, chat/assistants, classification, forecasting, and recommendation systems using a fit‑for‑purpose toolkit from traditional predictive modeling to agentic workflows.
  • Regulatory Intelligence & Filing Automation: Design and deploy GenAI capabilities to automate regulatory filing support, including DOI objection response generation and ingestion of legacy filings into searchable knowledge bases. Partner with Legal and Compliance to ensure outputs meet evolving standards and enable direct API integrations with regulatory bodies.
  • Knowledge Base Engineering for Strategic Domains: Engineer and maintain domain‑specific knowledge bases (regulatory intelligence, competitive insights, customer sentiment) to power generative applications across underwriting, pricing, and service.
  • Domain & Compliance Integration: Develop deep understanding of The Hartford’s business structures, processes, and data sources. Embed domain taxonomies, regulatory constraints, access controls, and security into the solution design. Ensure adherence to responsible AI practices—fairness, bias mitigation, transparency, and observability with compliance‑by‑design.
  • Stakeholder Collaboration: Partner with leaders and SMEs across Product, Operations, Claims, Underwriting, and Risk to align initiatives to business goals. Define success criteria that balance accuracy, reusability, cost, and performance. Translate model behavior into actionable strategies with clear ROI.
  • End‑to‑End Solution Development: Own the AI lifecycle from problem framing through deployment: data prep, modeling, evaluation, CI/CD, orchestration, observability, safety filters/guardrails, and rollback plans. Collaborate closely with AI engineers, data engineers, platform, security, and IT for seamless integration.
  • Unstructured Data & Retrieval Design: Prepare multi‑format content (PDF, Office, HTML, images, audio) with normalization, metadata/lineage management, and PII detection/redaction. Design retrieval strategies (chunking, embeddings, hybrid search) and tune for cost, latency, and domain fit; leverage rerankers where appropriate.
  • Prompt & Agent Design: Author robust system prompts, few‑shot patterns, and structured outputs (e.g., JSON schemas). Define safe tool‑use policies and function/structured calling for reliable agent behavior.
  • Evaluation & Monitoring: Define metrics across use cases for classification, information retrieval, RAG/chat, forecasting, plus customer/ops KPIs. Build gold/synthetic test sets, support A/B testing, and monitor drift. Provide economic, qualitative, and statistical analysis supporting thresholds and decisions.
  • Synthetic Data Generation & Augmentation: Develop and validate synthetic data pipelines to alleviate sparsity and accelerate convergence, especially for low‑frequency perils and emerging segments, while preserving privacy and distributional fidelity.
  • Customer Experience Optimization: Apply GenAI to elevate self‑service, virtual assistants, and inspection automation, driving personalization, speed, and operational efficiency.
  • Architectural Collaboration & MLOps Integration: Partner with enterprise architects and platform teams to ensure scalable, secure deployments via unified systems. Standardize experiment tracking, registries, evaluation gates, and CI/CD patterns across clouds and services.
  • Innovation & Continuous Learning: Identify and pilot emerging methods (OCR, rerankers, PEFT/LoRA, distillation). Build reusable accelerators (chunking templates, prompt registries, evaluation harnesses). Stay current on AI/ML, LLMOps, NLP, RAG, and responsible AI.

Required Skills & Experience:

  • 5–7 years of professional experience with a Bachelor’s degree, or 3–5 years of experience with a Master’s or Ph.D.; Master’s degree or Ph.D. in Machine Learning, Applied Mathematics, Data Science, Computer Science, or a closely related analytical field preferred, or demonstrated progress toward a relevant professional designation.
  • 5+ years of experience in statistical modeling and machine learning using Python, including extensive use of pandas, NumPy, scikit-learn, and strong SQL for data exploration, feature development, and knowledge preparation; familiarity with PyTorch and/or TensorFlow preferred.
  • 5+ years of experience across the end-to-end modeling lifecycle, including problem framing and requirements gathering, experiment design, offline evaluation, and ongoing production validation and monitoring.
  • 5+ years of solid understanding and practical application of core machine learning methods, with 3+ years of experience applying deep learning architectures in real-world use cases.
  • 4+ years of experience designing and operationalizing model evaluation and monitoring approaches, including test set creation (gold and/or synthetic), metric definition and tracking (e.g., classification, forecasting, ranking/IR, and business KPIs), and supporting A/B testing, drift detection, and performance regression monitoring.
  • 3+ years of experience working with unstructured data, including document parsing and OCR fundamentals, text normalization, metadata and lineage awareness, and PII detection or redaction considerations.
  • 5+ years of experience using Git and Unix-based development environments, with experience building reproducible notebooks or pipelines and ensuring repeatable analytical workflows; 3+ years of exposure to basic container and cloud fundamentals supporting deployment workflows
  • 4+ years of experience communicating modeling decisions, design tradeoffs, evaluation results, and risks to both technical and non-technical audiences, and translating analytical outcomes into measurable business impact.
  • 3+ years of experience working with cloud-based AI platforms such as Google Vertex AI, AWS SageMaker or Bedrock, or Azure AI Services, supporting experimentation, model training, and deployment.
  • 3+ years of experience deploying models and integrating scoring logic into production systems, including operation within complex enterprise or packaged application environments (e.g., Duck Creek, Ratabase).
  • 2+ years of experience with NLP and Generative AI capabilities, including embeddings, retrieval strategies (dense and hybrid), chunking approaches, prompt engineering, structured outputs, and contributing to Retrieval-Augmented Generation (RAG) solutions and evaluations.
  • 1+ year of experience or exposure to advanced GenAI applications and extensions, such as agent or tool-use concepts, domain-specific knowledge graph integration, synthetic data generation, sentiment modeling, and GenAI use cases in filing or compliance contexts.
  • 2+ years of experience working within enterprise AI governance expectations, including aligning model development with compliance, privacy, documentation, and ethical standards.


Nice to Have

  • RAG Expertise: Handson with vector databases and search (e.g., Vertex AI RAG Engine, OpenSearch, pgvector/Postgres), ANN indexing (HNSW), rerankers (crossencoders), and evaluation frameworks (RAGAS, TruLens, DeepEval).
  • Document AI Tooling: PyMuPDF/pdfplumber, Apache Tika; OCR (Tesseract); layoutaware models (LayoutLM); table extraction (Camelot/Tabula).
  • Embedding Model Selection: Experience comparing OpenAI/Cohere/Voyage vs. opensource (bge/e5/gte) for domain corpora; understanding dimension/quality/cost/latency tradeoffs and multilingual needs.
  • Orchestration Frameworks: Familiarity with LangChain, LangGraph, or LlamaIndex; structured tool/function calling and guardrails for AI agents.
  • CloudNative ML: Handson with Vertex AI, SageMaker, or Azure ML; experiment tracking (MLflow/W&B), registries, and CI evaluation gates.
  • Responsible AI & Safety: Bias/fairness testing, hallucination mitigation, grounding checks, safety filters; basic model risk documentation.
  • Broader Modalities (Nice to Have): Timeseries forecasting, recommenders, anomaly/fraud detection, speech/vision/multimodal.
  • Fine tuning LLMs and Diffusion models using PEFT/LoRA, experience with distillation

What We Offer

  • Collaborative work environment with global teams.
  • Competitive salary and comprehensive benefits.
  • Continuous learning and professional development opportunities.

Required Skills

Machine Learning Engineering Forecasting Documentation Git Automation Safety Compliance Monitoring Python SQL Training AWS Statistical Analysis HTML Pandas NumPy Deep Learning TensorFlow PyTorch Scikit-learn MLOps Recommendation Systems Azure LangChain MLflow JSON Drift KPIs A/B Testing Data Science CI/CD Grounding Framing Continuous Learning Fraud Detection Testing Postgres Apache NLP RAG Validation Unix Pricing Mathematics Metadata Governance Regression Applied Mathematics Risk Statistical modeling Orchestration Solution design Underwriting Toolkit Requirements gathering Bias mitigation Model development Predictive OCR Graph Modeling Information retrieval Vector Qualitative Generative Images BedRock Indexing API integrations Guardrails Model Evaluation Feature development Data sources Schemas Machine learning methods Ingestion Predictive Modeling Notebooks Experiment design Statistical Data Exploration Data pipelines Normalization Embedding Data generation Synthetic Sagemaker OpenAI Solution development Synthetic Data Privacy Speech Format Tika LLMOps OpenSearch Vertex Orchestration frameworks Chunking Experimentation Detection Extraction Experiment Retrieval-Augmented Generation Exploration Knowledge Graph Tesseract GenAI Prompt engineering Unstructured data Retrieval Distillation LLMs Legal Model Training Observability Virtual Audio Rollback Generative AI Fraud Container Classification GenAI capabilities Table Production Validation AWS Sagemaker Azure ML Professional Development AI/ML SAFe Computer Science SciKit Agentic AI LoRA PEFT Augmented Generation Stakeholder Collaboration Retrieval-augmented Vector Databases Vertex AI Advanced AI Dimension Langgraph Embeddings AI Agents Filters
Check Qualification

Quick Tip

Customize your resume and cover letter to highlight relevant skills for this position to increase your chances of getting hired.