The future of AI is inference
With the rise of agentic workflows and reasoning models, enterprises now need 100x more compute and 10x more throughput to run state-of-the-art AI models. Building robust, scalable inference systems has become a top priority—but it's also a major bottleneck, requiring deep expertise in low-level systems, snapshotters, Kubernetes, and more.
Tensorfuse removes this complexity by helping teams run serverless GPUs in their own AWS account. Just bring:
We handle the rest—deploying, managing, and autoscaling your GPU containers on production-grade infrastructure. Teams use Tensorfuse for:
We’re building the runtime layer for AI-native companies. Join us.
This is an in person role at our office in Bangalore. We’re an early stage company which means that the role requires working hard and moving quickly. Please only apply if that excites you.
fulltimeBengaluru, KA, INDevops₹2.5M - ₹4M INR0.50% - 0.80%1+ years