Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Question about coreset selection #2

Open
zhangxin-xd opened this issue Aug 8, 2023 · 5 comments
Open

Question about coreset selection #2

zhangxin-xd opened this issue Aug 8, 2023 · 5 comments
Labels
good first issue Good for newcomers

Comments

@zhangxin-xd
Copy link

Thank you for sharing your excellent work! I have a question about coreset selection. I noticed that in Algorithm 1, all the samples are re-sorted according to dACS and then reconstituted in the subset. It appears that the coreset selection is dynamic, akin to dropping out some unimportant samples during the training phase (the dropped ones can be reselected). However, some of the comparison methods are static (the dropping is permanent). Is the comparison reasonable?

I'm looking forward to your reply!

@HuangOwen
Copy link
Owner

Thanks for the question and your interest in our work! We believe that this comparison is reasonable as all the coreset method has the same "coreset data fraction per epoch". Since our target is to improve the training efficiency, the training time reduction is the same across different methods. In addition, previous work [1][2] also adopts a similar adaptive coreset strategy and compares it with other fixed-coreset methods.
[1] Adaptive second order coresets for data-efficient machine learning, ICML 2022
[2] RETRIEVE: Coreset Selection for Efficient and Robust Semi-Supervised Learning, NeurIPS 2021

@HuangOwen HuangOwen added the good first issue Good for newcomers label Aug 8, 2023
@zhangxin-xd
Copy link
Author

Thanks for your reply!!

@zhangxin-xd
Copy link
Author

Hi, I have another question about the Error Vector Score which acts like EL2N.
In EL2N,
image.
The average is calculated across several models with different initialization weights.
In your ACS,
image.
When QAT starts, the weights at time t are fixed so how is the average calculated?

@HuangOwen
Copy link
Owner

Thanks for your question. Different from the GraNd score proposed in EL2N, the expectation of our ACS is computed on all logits $m \in M$ at a given training time $t$ (It is an average of these gradients instead of the sum). Since we then use $d_{\text{EVS}}$ to approximate it, the analysis still holds. We will correct the expectation equation in our manuscript later to avoid confusion.

@zhangxin-xd
Copy link
Author

Got that! Thanks for the reply!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
good first issue Good for newcomers
Projects
None yet
Development

No branches or pull requests

2 participants