Backend / ML-Ops Engineer — Speech Model Deployment & Inference Optimization - Bengaluru
1 week ago

Job description
, consectetur adipiscing elit. Nullam tempor vestibulum ex, eget consequat quam pellentesque vel. Etiam congue sed elit nec elementum. Morbi diam metus, rutrum id eleifend ac, porta in lectus. Sed scelerisque a augue et ornare.
Donec lacinia nisi nec odio ultricies imperdiet.
Morbi a dolor dignissim, tristique enim et, semper lacus. Morbi laoreet sollicitudin justo eget eleifend. Donec felis augue, accumsan in dapibus a, mattis sed ligula.
Vestibulum at aliquet erat. Curabitur rhoncus urna vitae quam suscipit
, at pulvinar turpis lacinia. Mauris magna sem, dignissim finibus fermentum ac, placerat at ex. Pellentesque aliquet, lorem pulvinar mollis ornare, orci turpis fermentum urna, non ullamcorper ligula enim a ante. Duis dolor est, consectetur ut sapien lacinia, tempor condimentum purus.
Access all high-level positions and get the job of your dreams.
Similar jobs
We are seeking a highly skilled and innovative Inference Optimization (LLM and Runtime) to design, develop, and optimize cutting-edge AI systems that power intelligent, scalable, and agent-driven workflows. · Ph.D. in Computer Science or a related field with specialization in Dee ...
1 month ago
We are seeking an experienced LLM Inference Engineer to optimize and scale production-grade large language model (LLM) inference systems. · Key ResponsibilitiesOwn end-to-end production inference, including release engineering, capacity planning, cost optimization, and incident r ...
3 weeks ago
We're Hiring | Software Engineer – ML Engineer · Location: Bangalore (Hybrid – 3 Days WFO per Week) · Experience: 6–9 Years · Notice Period: Immediate to 60 Days · Join our AI Platform team and help remove compute bottlenecks for production-grade Large Language Models (LLMs). Thi ...
2 weeks ago
This position involves designing efficient AI inference pipelines for production environments. · ...
1 month ago
Role: AI/ML Ops Engineer (GPU Acceleration & AI Inference) · Location: Offshore – Bangalore (BCIT) · Experience: 5+ Years / 7+ Years · We are looking for passionate · AI/ML Ops Engineers · to build and scale enterprise-grade AI platforms with a strong focus on · GPU acceleration, ...
2 weeks ago
We are seeking an experienced LLM Inference Engineer to optimize and scale production-grade large language model (LLM) inference systems. · This is a high-impact role for engineers passionate about scaling AI systems and delivering performant,reliable LLM infrastructure in produc ...
3 weeks ago
We are building next-generation AI inference processors optimized for ultra-low latency, high-throughput workloads. As a Senior / Principal Engineer, · you will play a critical role in designing and optimizing low-level software and compute kernels that extract maximum performanc ...
1 month ago
We are seeking an experienced C++ AI Inference Engineer to design, optimize, and deploy high-performance AI inference engines using modern C++ and processor-specific optimizations. You will collaborate with research teams to productionize cutting-edge AI model architectures for C ...
1 week ago
We're Hiring | Software Engineer – ML Engineer · Location: Bangalore (Hybrid – 3 Days WFO per Week) · Experience: 6–9 Years · Notice Period: Immediate to 60 Days · Join our AI Platform team and help remove compute bottlenecks for production-grade Large Language Models (LLMs). Thi ...
2 weeks ago
Build the world's fastest CPU-based AI inference engines. · Architect C libraries that power production-grade LLMs and vision models. · Squeeze every cycle from modern processors using AVX-512, operator fusion, and NUMA-aware memory systems. · ...
3 weeks ago
We're changing the way the world shops and sells. Our platform empowers millions of buyers and sellers in more than 190 markets around the world. · We're committed to pushing boundaries and leaving our mark as we reinvent the future of ecommerce for enthusiasts.Own production inf ...
2 weeks ago
At eBay, we're more than a global ecommerce leader — we're changing the way the world shops and sells. Our platform empowers millions of buyers and sellers in more than 190 markets around the world. We're committed to pushing boundaries and leaving our mark as we reinvent the fut ...
4 days ago
About Gruve · Gruve is an innovative software services startup dedicated to transforming enterprises to AI powerhouses. We specialize in cybersecurity, customer experience, cloud infrastructure, · We are seeking a highly experienced and visionary Senior Engineering Manager – Infe ...
1 month ago
End-to-End ML Pipeline Development: Build and optimize model training, evaluation, and deployment pipelines for large-scale production environments. · High-Performance Inference Engineering: Architect and scale distributed inference systems capable of processing large request vol ...
1 month ago
Role: AI/ML Ops Engineer (GPU Acceleration & AI Inference) · Location: Offshore – Bangalore (BCIT) · Experience: 5+ Years / 7+ Years · We are looking for passionate AI/ML Ops Engineers to build and scale enterprise-grade AI platforms with a strong focus on GPU acceleration, infer ...
2 weeks ago
Role & responsibilities · ROLE OVERVIEW · We are looking for an experienced AI Infrastructure Engineer to build and operate the GPU-powered infrastructure that underpins our LLM training and inference platform. This is a DevOps-meets-MLOps role requiring expertise in GPU cluster ...
1 week ago
Experience: 1215 Years · Role Overview · We are hiring a Principal AI Scientist to lead applied research and optimization of LLM systems, with deep expertise in tokenization, inference efficiency, and cost modeling. · Key Responsibilities · Optimize token usage and context window ...
2 weeks ago
About AION · AION is building an interoperable AI cloud platform by transforming the future of high-performance computing (HPC) through its decentralized AI cloud. Purpose-built for bare-metal performance, AION democratizes access to compute and provides managed services, aiming ...
4 days ago
Job Title: · AI/ML Engineer Quantization & Optimization for CNN Models · Job Summary: · We are seeking a highly skilled AI/ML Engineer with deep expertise in quantization, optimization, and deployment of CNN-based models for efficient edge and embedded inference. The ideal candid ...
1 week ago
Fine-tune open-source LLMs using techniques such as SFT Supervised Fine-Tuning GRPO Group Relative Policy Optimization DPO and RLHF Leverage Unsloth and similar frameworks for memory-efficient high-speed fine-tuning on consumer and cloud GPUs Design and curate high-quality traini ...
2 weeks ago