We are seeking an exceptional AI Research Scientist to join Naptha AI at the ground floor, focusing on advancing the state of the art in test time compute optimization for large language models. In this role, you will be responsible for researching and developing novel approaches to improve inference efficiency, reduce computational requirements, and enhance model performance at deployment.
Working directly with our technical team, you will help shape the fundamental architecture of our inference optimization platform and solve core technical challenges around model compression, efficient inference strategies, and deployment optimization. You will work at the intersection of machine learning, systems optimization, and hardware acceleration to develop practical solutions for real‑world model deployment and scaling.
Naptha AI is committed to building a diverse and inclusive workplace. We are an equal opportunity employer and welcome applications from all qualified candidates regardless of background.