Senior AI Pipeline Engineer (Python, AWS, LLMs) – Document Intelligence / CLM
Actively Reviewing the ApplicationsInhubber
On-site
Posted 17 hours ago
•
Apply by June 10, 2026
Job Description
About Inhubber
Inhubber is a security-first, AI-powered Contract Lifecycle Management (CLM) platform built for organizations with high compliance and data protection requirements. We combine end-to-end encryption and modern cloud architecture with advanced AI to extract, analyze, and generate contract intelligence.
Our platform processes sensitive legal documents for companies worldwide. We are now scaling our AI capabilities and looking for a senior engineer to take ownership of our production AI pipelines and help build the next generation of document intelligence.
The Role
We are looking for a hands-on Senior AI Pipeline Engineer (Python) who will:
What Youll Do
1. Own & Extend Production Pipelines
Weeks 1–2
Inhubber is a security-first, AI-powered Contract Lifecycle Management (CLM) platform built for organizations with high compliance and data protection requirements. We combine end-to-end encryption and modern cloud architecture with advanced AI to extract, analyze, and generate contract intelligence.
Our platform processes sensitive legal documents for companies worldwide. We are now scaling our AI capabilities and looking for a senior engineer to take ownership of our production AI pipelines and help build the next generation of document intelligence.
The Role
We are looking for a hands-on Senior AI Pipeline Engineer (Python) who will:
- Own and optimize our production extraction pipelines.
- Deliver new document-analysis pipelines end-to-end.
- Build the foundation for next-generation GenAI features (agentic contract drafting & interpretation).
What Youll Do
1. Own & Extend Production Pipelines
- Maintain and optimize our Python-based extraction pipelines (AWS Lambda + S3 + Docker components).
- Ensure stable document processing and downstream triggering.
- Improve observability: logging, metrics, alerting, traceability, cost monitoring.
- Debug and stabilize real-world failure modes in production.
- Design and implement end-to-end pipelines for new document families.
- Build evaluation datasets and regression tests.
- Prevent silent quality degradation through measurable metrics.
- Improve Q&A and structured extraction using LLMs.
- Implement structured outputs, retrieval (RAG where useful), and deterministic validation.
- Add robust failure handling (timeouts, retries, fallbacks, safe defaults).
- Build agentic building blocks in Python behind stable APIs.
- Contribute to a contract-generation/editing kernel (planner, drafter, risk checks).
- Collaborate with backend/frontend teams for clean integration.
- Ensure scalability, cost-efficiency, and security.
- Contribute to deployment/versioning/rollback strategies.
- Help define operational runbooks.
- Strong production-grade Python (clean architecture, testing, packaging, APIs).
- Experience owning code in production.
- AWS serverless (Lambda + S3 required; Step Functions/SQS/CloudWatch a plus).
- Docker and containerized services.
- Proven experience maintaining/debugging automated pipelines.
- Hands-on experience with LLMs (OpenAI/Azure OpenAI/Anthropic or similar):
- Structured outputs
- Prompt iteration
- Retrieval (RAG)
- Evaluation approaches
- Document AI experience (OCR, layout extraction, noisy PDFs).
- Evaluation-driven development (test sets, regression checks, quality metrics).
- Experience with cost/latency budgeting.
- Familiarity with TypeScript/Node.
- Experience integrating REST services.
Weeks 1–2
- Fully understand current pipeline architecture.
- Stabilize staging/local environments.
- Define quality, cost, and latency baselines.
- Improve logging and monitoring.
- Deliver a new production-ready pipeline for a new document family.
- Implement evaluation datasets and regression checks.
- Deploy with monitoring and rollback strategy.
- Improve Q&A/extraction accuracy measurably.
- Deliver a first version of a GenAI contract drafting kernel.
- Harden operations (cost controls, retries, fallbacks, documentation).
- Frontend: React (TypeScript)
- Backend: Java (JEE)
- AI Pipelines: Python (AWS Lambda), Dockerized OCR/NLP
- Storage: AWS S3
- LLMs: Azure-hosted LLM APIs
- Infrastructure: AWS + Azure (hybrid)
- Senior-level role
- Long-term collaboration preferred
- Strong ownership mindset required
- Experience in regulated/sensitive data environments is a strong plus
Required Skills
Documentation
Monitoring
Python
AWS
Editing
TypeScript
Docker
Interpretation
Azure
React
Testing
Lambda
NLP
Anthropic
Debugging
RAG
Validation
Regression
Risk
SQS
AWS Lambda
Packaging
OCR
Node
Clean architecture
Azure OpenAI
Traceability
PDFs
Rest services
Kernel
Sets
Pipeline architecture
Logging
OpenAI
JEE
Extraction
LLM APIs
Regression tests
AWS S3
GenAI
Retrieval
Java
LLMs
Budgeting
Storage
Observability
Rollback
Drafting
Planner
SAFe
LLM
Serverless
Step Functions
Quick Tip
Customize your resume and cover letter to highlight relevant skills for this position to increase your chances of getting hired.
Related Similar Jobs
View All
Executive Assistant
ARV Companies
India
Full-Time
₹3–4 LPA
Communication
Time Management
Multitasking
+10
Manager - MIS & Reporting
SBI Card
India
Full-Time
Communication
Data Collection
Data Analysis
+55
DevOps Engineer
Methods
Git
Python
Terraform
+7
Software Engineer
Plaid
India
Full-Time
Testing
Debugging
Software engineering
Full-stack Developer - Golang & React (Remote)
Orca Pacific: A Monks Company
Mumbai
Full-Time
JavaScript
PostgreSQL
Docker
+2
Share
Quick Apply
Upload your resume to apply for this position