Principal DevOps / Edge AI Engineer
Actively Reviewing the ApplicationsMulya Technologies
Bengaluru
Full-Time
Remote
Posted 4 months ago
•
Apply by May 4, 2026
Job Description
Principal DevOps / Edge AI Engineer
Bangalore
Founded in 2023,by Industry veterans HQ in California,US
We are revolutionizing sustainable AI compute through intuitive software with composable silicon
Principal DevOps / Edge AI Engineer
Overview:
You will be responsible for building, deploying, and maintaining the local infrastructure that powers high-performance multimodal AI models (text, image, audio, video) on a compact AI appliance. You?ll bridge the gap between hardware, ML inference, and user-facing applications - ensuring reliability, scalability, and efficiency of on-device AI workloads.
Key Responsibilities:
System Deployment & Orchestration
Containerize AI inference services and web applications using Docker or Podman.
Design lightweight orchestration layers for local systems (Kubernetes, Nomad, or custom orchestration).
Automate build, test, and deployment pipelines (CI/CD) for local-first AI workloads.
Performance Optimization & Resource Management
Optimize compute utilization for concurrent multimodal workloads.
Develop monitoring tools for system health, thermal management, and memory/bandwidth usage.
Tune OS, drivers, and I/O subsystems for maximum throughput and low latency.
Edge Infrastructure & Networking
Configure low-latency local networking for browser-based access to the AI appliance.
Set up secure local APIs and data isolation layers ? ensuring zero external data leakage.
Integrate hardware accelerators and manage firmware updates across different SKUs.
Reliability, Testing, and Scaling
Build test harnesses to validate multimodal model performance (e.g., LLM + diffusion + ASR pipelines).
Implement over-the-air (OTA) update mechanisms for edge devices without exposing user data.
Develop monitoring dashboards and alerting for real-time performance metrics.
Required Qualifications:
Strong background in Linux systems engineering and containerization (Docker, Podman, LXC).
Experience deploying AI inference services locally or at the edge (llama.cpp, ollama, vLLM, ONNX).
Proficiency in CI/CD tools (GitHub Actions, Jenkins, ArgoCD) and infrastructure-as-code (Terraform, Ansible).
Expertise in GPU/accelerator optimization, CUDA stack management, or similar.
Solid understanding of networking, security, and firewall configurations for local appliances.
Scripting and automation skills (Python, Bash, Go, or Rust).
Preferred Qualifications:
Experience with embedded systems or edge AI devices (e.g., Jetson, Coral, FPGA-based accelerators).
experience minimum 10 years
Familiarity with low-bit quantization, model partitioning, or distributed inference.
Background in hardware/software co-design or systems integration.
Knowledge of browser-based local apps (WebSocket, WebRTC, RESTful APIs) and AI service backends.
Prior work in privacy-preserving AI systems or local-first architectures.
Contact:
Uday
Mulya Technologies
[email protected]
"Mining The Knowledge Community"
Quick Tip
Customize your resume and cover letter to highlight relevant skills for this position to increase your chances of getting hired.
Related Similar Jobs
View All
Job Posting Title Engineer, Site Reliability Engineering
LSEG
Bengaluru
Full-Time
Linux
SQL
Murex Architecture
+1
Python Full-Stack Developer
Cognisol
India
Full-Time
₹3–6 LPA
Django
PostgreSQL
Python
+12
Node Fullstack developer
Grid Dynamics
Hyderabad
Full-Time
JavaScript
TypeScript
React
+1
Senior Engineer, Firmware Testing (Protocol)
Landis+Gyr
Noida
Full-Time
Linux
Python
Freelance Senior Python Developer - Functional Test Frameworks
Mindrift
Git
Python
Rust
+5
Share
Quick Apply
Upload your resume to apply for this position