MGIE—Guiding Instruction-based Image Editing—demonstrates how a multimodal LLM can parse natural-language editing instructions and then drive image transformations accordingly. The project focuses on making edits explainable and controllable: the model interprets text guidance, reasons over image content, and outputs edits aligned with user intent. It’s positioned as an ICLR 2024 Spotlight work, with code and references that show how to connect language planning to concrete image operations. This bridges a gap between free-form prompts and precise edits by letting users describe “what” and “where” in everyday language. The repo includes instructions, examples, and links that situate MGIE within Apple’s broader line of multimodal research. For practitioners, MGIE provides a blueprint for text-to-edit systems that are more semantically grounded than naive prompt-only pipelines.

Features

  • Natural-language instruction parsing for image editing
  • Multimodal reasoning that ties text plans to visual changes
  • Examples and demos aligned with the research paper
  • Fine-grained, region-aware editing behavior
  • Open code for reproducibility and adaptation
  • Basis for controllable, explainable image-editing agents

Project Samples

Project Activity

See All Activity >

License

MIT License

Follow MGIE

MGIE Web Site

Other Useful Business Software
The Most Powerful Software Platform for EHSQ and ESG Management Icon
The Most Powerful Software Platform for EHSQ and ESG Management

Addresses the needs of small businesses and large global organizations with thousands of users in multiple locations.

Choose from a complete set of software solutions across EHSQ that address all aspects of top performing Environmental, Health and Safety, and Quality management programs.
Learn More
Rate This Project
Login To Rate This Project

User Reviews

Be the first to post a review of MGIE!

Additional Project Details

Programming Language

Python

Related Categories

Python Large Language Models (LLM)

Registered

2025-10-08