Join Apple's Video Computer Vision organization as an Applied Research Engineer focusing on multimodal LLMs. This role sits within a centralized applied research and engineering team responsible for developing real-time on-device Computer Vision and Machine Perception technologies across Apple products. You'll work on cutting-edge research projects involving foundation models and multimodal large language models that integrate various types of data such as text, image, video, and audio. The team balances research and product development to deliver Apple-quality, state-of-the-art experiences, innovating through the full stack, and partnering with Hardware, Software, and ML teams. The position offers competitive compensation including base salary, stock options, comprehensive benefits, and opportunities for professional growth. The ideal candidate will have strong expertise in AI/ML, with a focus on multimodal systems and a track record of translating research into practical applications.