Job Description
As a member of the LLM inference team, you will help build state-of-the-art software with the goal of enabling LLM inference to become more efficient, scalable, and accessible. Are you interested in architecting and implementing the best inference stacks in the LLM world? Work and collaborate with a diverse set of teams involving resource orchestration, distributed systems, inference engine optimization, and writing high-performance GPU kernels. Come join our team and contribute towards democratizing Machine Learning for the world!
CentML is a machine learning startup that offers optimization solutions for ML inference and training. Its software allows data scientists and engineers to find and apply value optimizations to their training and inference loads and helps clients have a platform that lowers compute costs.