Apple has yet another research paper. Their model ReALM can understand the context on the screen and perform tasks based on that information. Even with fewer parameters than GPT-4, ReALM’s heaviest models already outperform it in personal assistance tasks.