Responsibilities About the Team The Inference Infrastructure team is the creator and open-source maintainer of AIBrix, a Kubernetes-native control plane for large-scale LLM inference. We are part of ByteDance’s Core Compute Infrastructure organization, responsible for designing and operating the pla
About Abridge Abridge was founded in 2018 with the mission of powering deeper understanding in healthcare. Our AI-powered platform was purpose-built for medical conversations, improving clinical documentation efficiencies while enabling clinicians to focus on what matters most—their patients. Our en
About Armadin Armadin is an AI-native cybersecurity company building the ultimate attacker: an autonomous, agentic AI system that continuously finds and eliminates exploitable risk before adversaries do. We are redefining cybersecurity for the era of AI-driven Hyperattacks, deploying a fleet of spec
About Cartesia Our mission is to build the next generation of AI: ubiquitous, interactive intelligence that runs wherever you are. Today, not even the best models can continuously process and reason over a year-long stream of audio, video and text—1B text tokens, 10B audio tokens and 1T video tokens
Responsibilitie About the Team The Inference Infrastructure team is the creator and open-source maintainer of AIBrix, a Kubernetes-native control plane for large-scale LLM inference. We are part of ByteDance's Core Compute Infrastructure organization, responsible for designing and operating the plat
*About Rivian** Rivian is on a mission to keep the world adventurous forever. This goes for the emissions-free Electric Adventure Vehicles we build, and the curious, courageous souls we seek to attract. As a company, we constantly challenge what’s possible, never simply accepting what has always bee
Tech Lead Software Engineer - AI Inference Infrastructure Location: San Jose Team: Infrastructure Employment Type: Regular Job Code: A201019 Share this listing: Responsibilities Design and build large-scale, container-based cluster management and orchestration systems with extreme performance, scala
A cutting-edge AI company in San Francisco seeks an Inference Engineering Manager to lead a team dedicated to building scalable AI infrastructure. This role involves developing APIs, architecting inference systems, and enhancing reliability. Candidates should have over 5 years of engineering experie
About the Company Models are what they eat. But a large portion of training compute is wasted training on data that are already learned, irrelevant, or even harmful, leading to worse models that cost more to train and deploy. At DatologyAI, we've built a state of the art data curation suite to autom
A leading AI platform provider in New York seeks a Member of Technical Staff to develop and deploy large language models. You will ensure high availability and low latency for AI applications, working in a collaborative environment. Required skills include extensive engineering experience, knowledge
About Abridge Abridge was founded in 2018 with the mission of powering deeper understanding in healthcare. Our AI-powered platform was purpose-built for medical conversations, improving clinical documentation efficiencies while enabling clinicians to focus on what matters most—their patients. Our en
Senior ML Infrastructure Engineer, Inference Platform page is loaded## Senior ML Infrastructure Engineer, Inference Platformremote type: Hybridlocations: Sunnyvale, California, United States of America: Austin, Texas, United States of America: Mountain View, California, United States of America: War
Overview Rivian is on a mission to keep the world adventurous forever. This goes for the emissions-free Electric Adventure Vehicles we build, and the curious, courageous souls we seek to attract. As a company, we constantly challenge what's possible, never simply accepting what has always been done.
Responsibilitie About the Team The Inference Infrastructure team is the creator and open-source maintainer of AIBrix, a Kubernetes-native control plane for large-scale LLM inference. We are part of ByteDance's Core Compute Infrastructure organization, responsible for designing and operating the plat
About the Company Models are what they eat. But a large portion of training compute is wasted training on data that are already learned, irrelevant, or even harmful, leading to worse models that cost more to train and deploy. At DatologyAI, we've built a state of the art data curation suite to autom
Who are we? Our mission is to scale intelligence to serve humanity. We’re training and deploying frontier models for developers and enterprises who are building AI systems to power magical experiences like content generation, semantic search, RAG, and agents. We believe that our work is instrumental
Machine Learning Infrastructure Engineer Join to apply for the Machine Learning Infrastructure Engineer role at Abridge . Base pay range $179,000.00/yr - $248,000.00/yr About Abridge Abridge was founded in 2018 with the mission of powering deeper understanding in healthcare. Our AI‑powered platform
Responsibilitie About the Team The Inference Infrastructure team is the creator and open-source maintainer of AIBrix, a Kubernetes-native control plane for large-scale LLM inference. We are part of ByteDance's Core Compute Infrastructure organization, responsible for designing and operating the plat
Responsibilitie About the Team The Inference Infrastructure team is the creator and open-source maintainer of AIBrix, a Kubernetes-native control plane for large-scale LLM inference. We are part of ByteDance's Core Compute Infrastructure organization, responsible for designing and operating the plat
A dynamic AI company is seeking an Infrastructure Software Engineer in San Francisco to build and maintain components of an ML inference platform. The successful candidate will develop infrastructure components using Python and Go, manage Kubernetes deployments, and enhance monitoring systems for mo
Frequently asked questions about Inference Infrastructure
How much does a Inference Infrastructure earn in United States?
The estimated salary for Inference Infrastructure in United States ranges from $59,000 to $94,000 USD per year, depending on experience and location.
How many Inference Infrastructure jobs are available?
There are currently 122 job offers for Inference Infrastructure in United States listed on BeBee.
Which cities have the most Inference Infrastructure jobs?
The cities with the most Inference Infrastructure jobs in United States are: San Francisco, Seattle, San Jose, Palo Alto, New York.
How can I apply for a Inference Infrastructure job?
Sign up for free on BeBee, complete your professional profile, and apply directly to the Inference Infrastructure positions that interest you with one click.