MaGGIe excels in hair rendering and instance separation on natural images, outperforming MGM and InstMatt in complex, multi-instance scenarios.MaGGIe excels in hair rendering and instance separation on natural images, outperforming MGM and InstMatt in complex, multi-instance scenarios.

Robust Mask-Guided Matting: Managing Noisy Inputs and Object Versatility

Abstract and 1. Introduction

  1. Related Works

  2. MaGGIe

    3.1. Efficient Masked Guided Instance Matting

    3.2. Feature-Matte Temporal Consistency

  3. Instance Matting Datasets

    4.1. Image Instance Matting and 4.2. Video Instance Matting

  4. Experiments

    5.1. Pre-training on image data

    5.2. Training on video data

  5. Discussion and References

\ Supplementary Material

  1. Architecture details

  2. Image matting

    8.1. Dataset generation and preparation

    8.2. Training details

    8.3. Quantitative details

    8.4. More qualitative results on natural images

  3. Video matting

    9.1. Dataset generation

    9.2. Training details

    9.3. Quantitative details

    9.4. More qualitative results

8.4. More qualitative results on natural images

Fig. 13 showcases our model’s performance in challenging scenarios, particularly in accurately rendering hair regions. Our framework consistently outperforms MGM⋆ in detail preservation, especially in complex instance interactions. In comparison with InstMatt, our model exhibits superior instance separation and detail accuracy in ambiguous regions.

\ Fig. 14 and Fig. 15 illustrate the performance of our model and previous works in extreme cases involving multiple instances. While MGM⋆ struggles with noise and accuracy in dense instance scenarios, our model maintains high precision. InstMatt, without additional training data, shows limitations in these complex settings.

\ The robustness of our mask-guided approach is further demonstrated in Fig. 16. Here, we highlight the challenges faced by MGM variants and SparseMat in predicting missing parts in mask inputs, which our model addresses. However, it is important to note that our model is not designed as a human instance segmentation network. As shown in Fig. 17, our framework adheres to the input guidance, ensuring precise alpha matte prediction even with multiple instances in the same mask.

\ Lastly, Fig. 12 and Fig. 11 emphasize our model’s generalization capabilities. The model accurately extracts both human subjects and other objects from backgrounds, showcasing its versatility across various scenarios and object types.

\ All examples are Internet images without ground-truth and the mask from r101fpn400e are used as the guidance.

\ Figure 13. Our model produces highly detailed alpha matte on natural images. Our results show that it is accurate and comparable with previous instance-agnostic and instance-awareness methods without expensive computational costs. Red squares zoom in the detail regions for each instance. (Best viewed in color and digital zoom).

\ Figure 14. Our frameworks precisely separate instances in an extreme case with many instances. While MGM often causes the overlapping between instances and MGM⋆ contains noises, ours produces on-par results with InstMatt trained on the external dataset. Red arrow indicates the errors. (Best viewed in color and digital zoom).

\ Figure 15. Our frameworks precisely separate instances in a single pass. The proposed solution shows comparable results with InstMatt and MGM without running the prediction/refinement five times. Red arrow indicates the errors. (Best viewed in color and digital zoom).

\ Figure 16. Unlike MGM and SparseMat, our model is robust to the input guidance mask. With the attention head, our model produces more stable results to mask inputs without complex refinement between instances like InstMatt. Red arrow indicates the errors. (Best viewed in color and digital zoom).

\ Figure 17. Our solution works correctly with multi-instance mask guidances. When multiple instances exist in one guidance mask, we still produce the correct union alpha matte for those instances. Red arrow indicates the errors or the zoom-in region in red box. (Best viewed in color and digital zoom).

\ Table 12. Details of quantitative results on HIM2K+M-HIM2K (Extension of Table 5). Gray indicates the public weight without retraining.

\ Table 12. Details of quantitative results on HIM2K+M-HIM2K (Extension of Table 5). Gray indicates the public weight without retraining. (Continued)

\ Table 12. Details of quantitative results on HIM2K+M-HIM2K (Extension of Table 5). Gray indicates the public weight without retraining. (Continued)

\ Table 12. Details of quantitative results on HIM2K+M-HIM2K (Extension of Table 5). Gray indicates the public weight without retraining. (Continued)

\ Table 13. The effectiveness of proposed temporal consistency modules on V-HIM60 (Extension of Table 6). The combination of bi-directional Conv-GRU and forward-backward fusion achieves the best overall performance on three test sets. Bold highlights the best for each level.

\

:::info Authors:

(1) Chuong Huynh, University of Maryland, College Park (chuonghm@cs.umd.edu);

(2) Seoung Wug Oh, Adobe Research (seoh,jolee@adobe.com);

(3) Abhinav Shrivastava, University of Maryland, College Park (abhinav@cs.umd.edu);

(4) Joon-Young Lee, Adobe Research (jolee@adobe.com).

:::


:::info This paper is available on arxiv under CC by 4.0 Deed (Attribution 4.0 International) license.

:::

\

Market Opportunity
Mask Network Logo
Mask Network Price(MASK)
$0.58
$0.58$0.58
+1.15%
USD
Mask Network (MASK) Live Price Chart
Disclaimer: The articles reposted on this site are sourced from public platforms and are provided for informational purposes only. They do not necessarily reflect the views of MEXC. All rights remain with the original authors. If you believe any content infringes on third-party rights, please contact service@support.mexc.com for removal. MEXC makes no guarantees regarding the accuracy, completeness, or timeliness of the content and is not responsible for any actions taken based on the information provided. The content does not constitute financial, legal, or other professional advice, nor should it be considered a recommendation or endorsement by MEXC.

You May Also Like

Fundstrat’s Internal Report Contradicts CIO Tom Lee’s Bold Crypto Forecasts

Fundstrat’s Internal Report Contradicts CIO Tom Lee’s Bold Crypto Forecasts

The post Fundstrat’s Internal Report Contradicts CIO Tom Lee’s Bold Crypto Forecasts appeared on BitcoinEthereumNews.com. Key Points: Fundstrat internal report
Share
BitcoinEthereumNews2025/12/21 13:19
SEC Backs Nasdaq, CBOE, NYSE Push to Simplify Crypto ETF Rules

SEC Backs Nasdaq, CBOE, NYSE Push to Simplify Crypto ETF Rules

The US SEC on Wednesday approved new listing rules for major exchanges, paving the way for a surge of crypto spot exchange-traded funds. On Wednesday, the regulator voted to let Nasdaq, Cboe BZX and NYSE Arca adopt generic listing standards for commodity-based trust shares. The decision clears the final hurdle for asset managers seeking to launch spot ETFs tied to cryptocurrencies beyond Bitcoin and Ether. In July, the SEC outlined how exchanges could bring new products to market under the framework. Asset managers and exchanges must now meet specific criteria, but will no longer need to undergo drawn-out case-by-case reviews. Solana And XRP Funds Seen to Be First In Line Under the new system, the time from filing to launch can shrink to as little as 75 days, compared with up to 240 days or more under the old rules. “This is the crypto ETP framework we’ve been waiting for,” Bloomberg research analyst James Seyffart said on X, predicting a wave of new products in the coming months. The first filings likely to benefit are those tracking Solana and XRP, both of which have sat in limbo for more than a year. SEC Chair Paul Atkins said the approval reflects a commitment to reduce barriers and foster innovation while maintaining investor protections. The move comes under the administration of President Donald Trump, which has signaled strong support for digital assets after years of hesitation during the Biden era. New Standards Replace Lengthy Reviews And Repeated Denials Until now, the commission reviewed each application separately, requiring one filing from the exchange and another from the asset manager. This dual process often dragged on for months and led to repeated denials. Even Bitcoin spot ETFs, finally approved in Jan. 2024, arrived only after years of resistance and a legal battle with Grayscale. According to Bloomberg ETF analyst Eric Balchunas, the streamlined rules could apply to any cryptocurrency with at least six months of futures trading on the Coinbase Derivatives Exchange. That means more than a dozen tokens may now qualify for listing, potentially unleashing a new wave of altcoin ETFs. SEC Clears Grayscale Large Cap Fund Tracking CoinDesk 5 Index The SEC also approved the Grayscale Digital Large Cap Fund, which tracks the CoinDesk 5 Index, including Bitcoin, Ether, XRP, Solana and Cardano. Alongside this, it cleared the launch of options linked to the Cboe Bitcoin US ETF Index and its mini contract, broadening the set of crypto-linked derivatives on regulated US markets. Analysts say the shift shows how far US policy has moved. Where once regulators resisted digital assets, the latest changes show a growing willingness to bring them into the mainstream financial system under established safeguards
Share
CryptoNews2025/09/18 12:40
Bank of Canada cuts rate to 2.5% as tariffs and weak hiring hit economy

Bank of Canada cuts rate to 2.5% as tariffs and weak hiring hit economy

The Bank of Canada lowered its overnight rate to 2.5% on Wednesday, responding to mounting economic damage from US tariffs and a slowdown in hiring. The quarter-point cut was the first since March and met predictions from markets and economists. Governor Tiff Macklem, speaking in Ottawa, said the decision was unanimous. “With a weaker economy […]
Share
Cryptopolitan2025/09/17 23:09