This article presents an ablation study showing that the proposed IIL method performs well with larger networksThis article presents an ablation study showing that the proposed IIL method performs well with larger networks

Network Size and Task Number: Ablation Study on IIL Performance and Stability

2025/11/12 23:30

Abstract and 1 Introduction

  1. Related works

  2. Problem setting

  3. Methodology

    4.1. Decision boundary-aware distillation

    4.2. Knowledge consolidation

  4. Experimental results and 5.1. Experiment Setup

    5.2. Comparison with SOTA methods

    5.3. Ablation study

  5. Conclusion and future work and References

    \

Supplementary Material

  1. Details of the theoretical analysis on KCEMA mechanism in IIL
  2. Algorithm overview
  3. Dataset details
  4. Implementation details
  5. Visualization of dusted input images
  6. More experimental results

12. More experimental results

12.1. Ablation study on network size

\ To investigate the impact of network size on the proposed method, we compare the performance of ResNet-18, ResNet-34 and ResNet-50 on the ImageNet-100. As shown in Tab. 5, the proposed method performs well with bigger

\ Table 5. Impact of the network size on the proposed method.

\ Table 6. Performance of the proposed method with different IIL task numbers.

\ network size. When the network size is larger, more parameters can be utilized for new knowledge learning with the proposed decision boundary-aware distillation. Hence, consolidating knowledge from student to teacher causes less forgetting.

\ 12.2. Ablation study on the task number

\ As mentioned in Sec. 7, our method accumulates the error along with the consecutive IIL tasks. However, such a kind of error accumulates slowly and mainly affects the performance on old tasks, i.e. forgetting rate. We further study the impact of task length on the performance of the proposed method by splitting the incremental data into different number of subsets. As shown in Tab. 6, with the incremental of task number, the performance promotion changes less but the forgetting rate increased slightly. Minor variation of performance promotion reveals that the proposed method is stable in learning new knowledge, irrespective of the number of tasks. The acquisition of new knowledge primarily hinges on the volume of new data involved. Although we increase the task number in the experiments, the total number of new data utilized in IIL phase is the same. While increasing the task number will increase the EMA steps, which causes more forgetting on the old data. Experimental results in Tab. 6 well validate our analysis in Sec. 7.

\ Compared to the performance promotion, forgetting on the old data is negligible. Noteworthy, when the task number is relatively small, such as 5 in Tab. 6, the proposed method slightly boosts the model’s performance on the base data. This behavior is similar with full-data model, which demonstrates the capability of our method in accumulating knowledge from new data.

\ 12.3. Detailed comparison between the KC-EMA and vanilla EMA

\ The performance of vanilla EMA and the proposed KCEMA during training is shown in Fig. 11. As can be seen, the student model’s accuracy initially plummets due to the introduction of new data. However, around the 10th epoch,

\ Figure 11. Comparison between the proposed KC-EMA and vanilla EMA during training in the first IIL task, where t denotes the teacher model and s denotes the student model. Result is achieved on Cifar-100.

\ there’s a resurgence in accuracy for both the KC-EMA and vanilla EMA models. Therefore, we empirically set a freezing epoch of 10 in the proposed method.

\ When EMA is applied post the 10th epoch, the teacher model in the vanilla EMA is rapidly drawn towards the student model. This homogenization, however, doesn’t enhance either model. Instead, it leads to a decline in test accuracy due to overfitting to the new data. In contrast, with KC-EMA, both the teacher and student models exhibit gradual growth,, which indicates a knowledge accumulation in these two models. On one hand, consolidating new knowledge to the teacher model improves its test performance. On the other hand, a teacher model equipped with new knowledge liberates the student model to learn new data. That is, constraints from the teacher in distillation is alleviated.

\

:::info Authors:

(1) Qiang Nie, Hong Kong University of Science and Technology (Guangzhou);

(2) Weifu Fu, Tencent Youtu Lab;

(3) Yuhuan Lin, Tencent Youtu Lab;

(4) Jialin Li, Tencent Youtu Lab;

(5) Yifeng Zhou, Tencent Youtu Lab;

(6) Yong Liu, Tencent Youtu Lab;

(7) Qiang Nie, Hong Kong University of Science and Technology (Guangzhou);

(8) Chengjie Wang, Tencent Youtu Lab.

:::


:::info This paper is available on arxiv under CC BY-NC-ND 4.0 Deed (Attribution-Noncommercial-Noderivs 4.0 International) license.

:::

\

Disclaimer: The articles reposted on this site are sourced from public platforms and are provided for informational purposes only. They do not necessarily reflect the views of MEXC. All rights remain with the original authors. If you believe any content infringes on third-party rights, please contact service@support.mexc.com for removal. MEXC makes no guarantees regarding the accuracy, completeness, or timeliness of the content and is not responsible for any actions taken based on the information provided. The content does not constitute financial, legal, or other professional advice, nor should it be considered a recommendation or endorsement by MEXC.

You May Also Like

BDACS Launches KRW1 Stablecoin Backed by the Won

BDACS Launches KRW1 Stablecoin Backed by the Won

The post BDACS Launches KRW1 Stablecoin Backed by the Won appeared on BitcoinEthereumNews.com. BDACS Launches KRW1 Stablecoin Backed by South Korean Won Custody service provider BDACS has launched KRW1, a new stablecoin pegged 1:1 to the South Korean won (KRW). The regulated custodian focuses on institutional clients and offers services including crypto asset custody and transaction infrastructure supporting multiple blockchains. The KRW1 project recently completed its proof-of-concept (PoC) phase, with the stablecoin launching on the Avalanche blockchain. Each KRW1 token is fully backed by fiat currency, with reserves held at Woori Bank, one of South Korea’s largest financial institutions. Transparency and Platform Features BDACS emphasizes full transparency: holders can monitor reserves in real time via banking API integration, although no dedicated portal is currently available. According to the press release, “The KRW1 launch goes far beyond token issuance. BDACS has developed a comprehensive platform, including issuance and governance systems, as well as a user application supporting peer-to-peer transfers and transaction verification.” The stablecoin is positioned for global use, with potential expansion through new network integrations and collaborations with dollar-pegged stablecoins like USDC and USDT. BDACS also plans to integrate KRW1 into government initiatives, though negotiations or official involvement have not been confirmed. Current Status and Market Outlook KRW1 remains in the concept stage and is not yet publicly traded or available to retail consumers, as South Korea currently lacks a stablecoin framework. However, the launch is reportedly supported by the country’s new president, Lee Je-moon. In related news, Kakao is also reportedly considering a won-pegged stablecoin, highlighting growing interest in this emerging asset class. Source: https://coinpaper.com/11089/bdacs-launches-krw-1-stablecoin-backed-by-the-won
Share
BitcoinEthereumNews2025/09/18 21:28