Export Citations
Save this search
Please login to be able to save your searches and receive alerts for new content matching your search criteria.
QuickDrop: Efficient Federated Unlearning via Synthetic Data Generation
Middleware '24: Proceedings of the 25th International Middleware ConferencePages 266–278https://doi.org/10.1145/3652892.3700764Federated Unlearning (FU) aims to delete specific training data from an ML model trained using Federated Learning (FL). However, existing FU methods suffer from inefficiencies due to the high costs associated with gradient recomputation and storage. This ...
- research-articleNovember 2024
Tab-Distillation: Impacts of Dataset Distillation on Tabular Data For Outlier Detection
ICAIF '24: Proceedings of the 5th ACM International Conference on AI in FinancePages 804–812https://doi.org/10.1145/3677052.3698660Dataset distillation aims to replace large training sets with significantly smaller synthetic sets while preserving essential information. This method reduces the training costs of advanced deep learning models and is widely used in the image domain. ...
- research-articleOctober 2024
Diversified Semantic Distribution Matching for Dataset Distillation
MM '24: Proceedings of the 32nd ACM International Conference on MultimediaPages 7542–7550https://doi.org/10.1145/3664647.3680900Dataset distillation, also known as dataset condensation, offers a possibility for compressing a large-scale dataset into a small-scale one (i.e., distilled dataset) while achieving similar performance during model training. This method effectively ...
- short-paperOctober 2024
Distributed Boosting: An Enhancing Method on Dataset Distillation
CIKM '24: Proceedings of the 33rd ACM International Conference on Information and Knowledge ManagementPages 3689–3693https://doi.org/10.1145/3627673.3679897Dataset Distillation (DD) is a technique for synthesizing smaller, compressed datasets from large original datasets while retaining essential information to maintain efficacy. Efficient DD is a current research focus among scholars. Squeeze, Recover and ...
- research-articleAugust 2024
Federated Edge Learning with Blurred or Pseudo Data Sharing
ICPP '24: Proceedings of the 53rd International Conference on Parallel ProcessingPages 981–990https://doi.org/10.1145/3673038.3673084Edge servers and mobile devices are often assigned a large number of computing tasks. However, the data involved in computing tasks is often sensitive in terms of privacy. Our initial proposal is a federated edge learning strategy based on real-world ...
- research-articleAugust 2024
Distillation vs. Sampling for Efficient Training of Learning to Rank Models
ICTIR '24: Proceedings of the 2024 ACM SIGIR International Conference on Theory of Information RetrievalPages 51–60https://doi.org/10.1145/3664190.3672527In real-world search settings, learning to rank (LtR) models are trained and tuned repeatedly using large amounts of data, thus consuming significant time and computing resources, and raising efficiency and sustainability concerns. One way to address ...
- research-articleMay 2024
Fast Graph Condensation with Structure-based Neural Tangent Kernel
WWW '24: Proceedings of the ACM Web Conference 2024Pages 4439–4448https://doi.org/10.1145/3589334.3645694The rapid development of Internet technology has given rise to a vast amount of graph-structured data. Graph Neural Networks (GNNs), as an effective method for various graph mining tasks, incurs substantial computational resource costs when dealing with ...
- research-articleMarch 2024
Coreset Discovery for Machine Learning Problems
Cybernetics and Systems Analysis (KLU-CASA), Volume 60, Issue 2Pages 198–208https://doi.org/10.1007/s10559-024-00661-yAbstractThe authors review a coreset discovery problem and three main methods to solve it: geometric coreset estimation, coreset discovery using the genetic algorithm, and coreset discovery using neural networks. They analyze each of these methods and ...
- posterJuly 2022
DEvS: data distillation algorithm based on evolution strategy
GECCO '22: Proceedings of the Genetic and Evolutionary Computation Conference CompanionPages 292–295https://doi.org/10.1145/3520304.3528819The development of machine learning solutions often relies on training using large labeled datasets. This raises challenges in terms of data storage, data privacy protection, and longer model training time. One of the possible solutions to overcome these ...