Sorry, looks like this job is no longer open 😔
Check out other job openings on our job board!
Our team builds ML-inference applications and services on Apple Silicon in the datacenter, specifically focusing in recent years on generative AI as part of the Private Cloud Compute component of Apple Intelligence.
Description
As part of the team you will help engineer continuous improvements in stability and performance for private cloud compute, as well as help implement entirely new functionality as it emerges from the research community, in collaboration with product teams throughout Apple. We write performant and scalable frameworks (in Swift and C++) to distribute and coordinate ML inference tasks to different hardware acceleration IP blocks on different SoCs. We’re a collection of highly skilled and friendly engineers who value each other’s opinions and experience. We strive for excellence and believe strongly in the quality of our output. We have formed a team of domain experts who specialize in specific core subject areas, and also have broad experience of cloud software services and platforms. You will integrate inference code into a full service stack to ensure that user traffic is served reliably and performantly, and will have a strong focus on developing code that is easy and safe to develop, update and monitor.
Minimum Qualifications
Preferred Qualifications
Apple is an equal opportunity employer that is committed to inclusion and diversity. We take affirmative action to ensure equal opportunity for all applicants without regard to race, color, religion, sex, sexual orientation, gender identity, national origin, disability, Veteran status, or other legally protected characteristics.
#J-18808-Ljbffr