The discussion compares BSGAL to existing methods, noting its efficient use of gradient cache to reduce runtime.The discussion compares BSGAL to existing methods, noting its efficient use of gradient cache to reduce runtime.

Future of GAL: Dynamic Filtering Strategies and Generative Model in the Loop

2025/12/09 10:57

Abstract and 1 Introduction

  1. Related work

    2.1. Generative Data Augmentation

    2.2. Active Learning and Data Analysis

  2. Preliminary

  3. Our method

    4.1. Estimation of Contribution in the Ideal Scenario

    4.2. Batched Streaming Generative Active Learning

  4. Experiments and 5.1. Offline Setting

    5.2. Online Setting

  5. Conclusion, Broader Impact, and References

    \

A. Implementation Details

B. More ablations

C. Discussion

D. Visualization

C. Discussion

C.1. Comparing with existing methods

\ We’ve drawn the Table 11, analyzing our setting compared to previous active learning or generative data filtration methods. We’ve conducted analysis from aspects of data scale, whether it’s oriented towards downstream tasks, label quality, labeling costs, and whether there exists domain difference (between generated and real data).

C.2. Analysis of the computational cost

We recorded the training duration and GPU memory usage for training 90,000 iterations with 4 Nvidia 4090 GPUs. It can be observed that our method based on Grad cache increases the GPU memory usage compared to Loss estimate, but it significantly reduces the training time. Compared with our Baseline, the additional time and memory overheads are within an acceptable range.

\ Table 12. Analysis of the computational cost of different algorithms

C.3. Future work

We hope that this paper can provide more inspiration to the academic community on how to utilize generated data and how to design better data analysis methods. It should be pointed out that our method is not limited to specific tasks or specific model architectures. In this work, for the convenience of comparison with the baseline, we use the same dataset and model architecture as the baseline. We hope that in future work, we can further verify it on more tasks and model architectures. At the same time, we can also design more flexible and controllable evaluation functions to better utilize generated data. For example, in this paper, when filtering the data with a gradient, there is a trade-off between diversity and consistency. For rare categories in the data, due to the small number of real data itself, diversity should be considered more, while for common categories, due to the large number of real data itself, consistency should be considered more. Therefore, in the future, we can consider adopting a dynamic strategy for different categories. In the long run, our current research is done under the premise of a fixed generative model. A more ideal situation is to involve the generative model in this loop, further optimizing the generative model based on the downstream model’s feedback, to achieve a true “generative model in the loop”.

\

:::info Authors:

(1) Muzhi Zhu, with equal contribution from Zhejiang University, China;

(2) Chengxiang Fan, with equal contribution from Zhejiang University, China;

(3) Hao Chen, Zhejiang University, China (haochen.cad@zju.edu.cn);

(4) Yang Liu, Zhejiang University, China;

(5) Weian Mao, Zhejiang University, China and The University of Adelaide, Australia;

(6) Xiaogang Xu, Zhejiang University, China;

(7) Chunhua Shen, Zhejiang University, China (chunhuashen@zju.edu.cn).

:::


:::info This paper is available on arxiv under CC BY-NC-ND 4.0 Deed (Attribution-Noncommercial-Noderivs 4.0 International) license.

:::

\

Sorumluluk Reddi: Bu sitede yeniden yayınlanan makaleler, halka açık platformlardan alınmıştır ve yalnızca bilgilendirme amaçlıdır. MEXC'nin görüşlerini yansıtmayabilir. Tüm hakları telif sahiplerine aittir. Herhangi bir içeriğin üçüncü taraf haklarını ihlal ettiğini düşünüyorsanız, kaldırılması için lütfen service@support.mexc.com ile iletişime geçin. MEXC, içeriğin doğruluğu, eksiksizliği veya güncelliği konusunda hiçbir garanti vermez ve sağlanan bilgilere dayalı olarak alınan herhangi bir eylemden sorumlu değildir. İçerik, finansal, yasal veya diğer profesyonel tavsiye niteliğinde değildir ve MEXC tarafından bir tavsiye veya onay olarak değerlendirilmemelidir.

Ayrıca Şunları da Beğenebilirsiniz

Tokenized Assets Shift From Wrappers to Building Blocks in DeFi

Tokenized Assets Shift From Wrappers to Building Blocks in DeFi

The post Tokenized Assets Shift From Wrappers to Building Blocks in DeFi appeared on BitcoinEthereumNews.com. RWAs are rapidly moving on-chain, unlocking new opportunities for investors and DeFi protocols, according to a new report from Dune and RWAxyz. Tokenized real-world assets (RWAs) are moving beyond digital versions of traditional securities to become key building blocks of decentralized finance (DeFi), according to the 2025 RWA Report from Dune and RWAxyz. The report notes that Treasuries, bonds, credit, and equities are now being used in DeFi as collateral, trading instruments, and yield products. This marks tokenization’s “real breakthrough” – composability, or the ability to combine and reuse assets across different protocols. Projects are already showing how this works in practice. Asset manager Maple Finance’s syrupUSDC, for example, has grown to $2.5 billion, with more than 30% placed in DeFi apps like Spark ($570 million). Centrifuge’s new deJAAA token, a wrapper for Janus Henderson’s AAA CLO fund, is already trading on Aerodrome, Coinbase and other exchanges, with Stellar planned next. Meanwhile, Aave’s Horizon RWA Market now lets institutional users post tokenized Treasuries and CLOs as collateral. This trend underscores a bigger shift: RWAs are no longer just copies of traditional assets; instead, they are becoming core parts of on-chain finance, powering lending, liquidity, and yield, and helping to close the gap between traditional finance (TradFi) and DeFi. “RWAs have crossed the chasm from experimentation to execution,” Sid Powell, CEO of Maple Finance, says in the report. “Our growth to $3.5B AUM reflects a broader shift: traditional financial services are adopting crypto assets while institutions seek exposure to on-chain markets.” Investor demand for higher returns and more diversified options is mainly driving this growth. Tokenized Treasuries proved there is strong demand, with $7.3 billion issued by September 2025 – up 85% year-to-date. The growth was led by BlackRock, WisdomTree, Ondo, and Centrifuge’s JTRSY (Janus Henderson Anemoy Treasury Fund). Spark’s $1…
Paylaş
BitcoinEthereumNews2025/09/18 06:10