Maitai manages the LLM stack for enterprise companies, enabling the fastest and most reliable inference. The future of enterprise AI revolves around mosaics of small, domain-specific models powering powerful, responsive agents, and Maitai is well positioned to capture the market. If you're looking at getting in early with a company redefining how large companies build with AI, then let's talk.
Join Maitai to reshape how enterprise companies build with open-source LLMs. You’ll be at the forefront, driving cutting-edge innovations in model fine-tuning, distillation, and automation to continuously enhance LLM performance. You’ll collaborate directly with founders, engineers, and enterprise customers, building the core management layer that defines enterprise AI infrastructure. We're scaling rapidly and looking for engineers who deeply understand open-source LLM ecosystems and can confidently automate and optimize model improvements at scale.
You will lead the fine-tuning, distillation, and deployment of open-source LLMs tailored for enterprise customers. Your role involves:
Maitai ensures LLMs never fail by optimizing reliability, speed, and resilience. Acting as an intelligent proxy, we apply real-time autocorrections, route requests intelligently, and fine-tune models for maximum performance. We're experiencing explosive growth, are well-capitalized, and seizing a massive opportunity to redefine how enterprises build with AI. Our platform delivers AI models that significantly outperform closed-source alternatives in speed and accuracy, supported by robust online guardrails. Leading YC startups and public enterprises trust Maitai to manage their LLM infrastructure.
As LLMs are core to our customers' products, resiliency and uptime are our top priorities. Since we act as a proxy, our uptime must exceed that of the providers themselves. We’re multi-cloud, multi-region, and built for seamless failover. Our infrastructure runs on Kubernetes, managed with Terraform, and deployed across AWS and GCP. We use GitHub Actions for CI/CD, with Datadog for monitoring, tracing, and performance insights.
Infra stack: Kubernetes, Terraform, AWS, GCP, GitHub Actions, PostgreSQL, Redis, Datadog.
Our backend is a set of microservices running Python with Quart for web services and Python-based fine-tuning jobs optimized for speed, cost, and accuracy. We use PostgreSQL for conventional data persistence and vector storage. Go is being introduced where performance gains are critical.
Tech stack: Python (Quart), Go (in transition), PostgreSQL.
Tech stack: React (Typescript)
Quick Chat (15-minute Video Call)
Let’s discuss your experience, interests, and ambitions.
Tech Discussion
Get on a call and talk tech. What's going on in the industry, what have you worked with recently, latest model you've fine-tuned, last meetup you were at, etc.
Hands-On Technical
Join us at our office to work through a problem with our team.
In-person Meetup
Coffee or lunch with our team. Assess fit from both sides and move quickly to a decision.
fulltimeRedwood City, CA, USFull stack$100K - $225K0.10% - 0.75%Any (new grads ok)
fulltimeRedwood City, CA, USMachine learning$100K - $225K0.10% - 0.75%Any (new grads ok)
fulltimeRedwood City, CA, USFull stack$150K - $225K1.00% - 4.00%3+ years
fulltimeRedwood City, CA, US$60K - $100K0.10% - 0.50%Any (new grads ok)