More on News
Apple has unveiled a groundbreaking AI image editing tool that allows users to write desired changes to a photograph using simple language, reducing the need for traditional photo editing software. MGIE (MLLM-Guided Image Editing) model is the outcome of a collaboration between Apple and the University of California, Santa Barbara.
MGIE is a significant development in image editing technology. It can perform several editing tasks, such as cropping, resizing, flipping, and applying filters. Its ability to understand and implement editing commands conveyed through natural language prompts sets it apart.
MGIE interprets user instructions using multimodal language models and generates corresponding edits accurately. For example, a request to “make the sky bluer” would result in adjustments to the brightness of the sky portion of an image. This approach ensures precise execution of editing commands, whether straightforward or complex.
In a statement accompanying the release, the researchers emphasized MGIE’s capability to understand explicit visual-aware intentions, leading to meaningful image enhancements. Extensive studies have validated MGIE’s effectiveness across various editing scenarios, showcasing its improved performance while maintaining efficiency.
Apple has made MGIE available for download via GitHub, with a web demo also accessible on Hugging Face Spaces. However, the company has not disclosed its plans for the model beyond research purposes.
While similar capabilities exist in other image generation platforms, such as OpenAI’s DALL-E 3 and Adobe’s Firefly AI in Photoshop, Apple’s entry into the generative AI space underscores its commitment to integrating advanced AI features into its products. CEO Tim Cook has previously expressed the company’s intention to expand AI functionalities across its devices, with recent initiatives including the release of the open-source machine learning framework MLX aimed at facilitating AI model training on Apple Silicon chips.