Export Citations
Save this search
Please login to be able to save your searches and receive alerts for new content matching your search criteria.
- surveyJanuary 2025JUST ACCEPTED
Data-centric Artificial Intelligence: A Survey
Artificial Intelligence (AI) is making a profound impact in almost every domain. A vital enabler of its great success is the availability of abundant and high-quality data for building machine learning models. Recently, the role of data in AI has been ...
- research-articleJanuary 2025
GNNs also deserve editing, and they need it more than once
- Shaochen (Henry) Zhong,
- Duy Le,
- Zirui Liu,
- Zhimeng Jiang,
- Andrew Ye,
- Jiamu Zhang,
- Jiayi Yuan,
- Kaixiong Zhou,
- Zhaozhuo Xu,
- Jing Ma,
- Shuai Xu,
- Vipin Chaudhary,
- Xia Hu
ICML'24: Proceedings of the 41st International Conference on Machine LearningArticle No.: 2553, Pages 61727–61746Suppose a self-driving car is crashing into pedestrians, or a chatbot is instructing its users to conduct criminal wrongdoing; the stakeholders of such products will undoubtedly want to patch these catastrophic errors as soon as possible. To address such ...
- research-articleJanuary 2025
LLM Maybe LongLM: SelfExtend LLM context window without tuning
ICML'24: Proceedings of the 41st International Conference on Machine LearningArticle No.: 888, Pages 22099–22114It is well known that LLMs cannot generalize well to long contexts whose lengths are larger than the training sequence length. This poses challenges when employing LLMs for processing long input sequences during inference. In this work, we argue that ...
- research-articleMarch 2024
Towards Mitigating Dimensional Collapse of Representations in Collaborative Filtering
WSDM '24: Proceedings of the 17th ACM International Conference on Web Search and Data MiningPages 106–115https://doi.org/10.1145/3616855.3635832Contrastive Learning (CL) has shown promising performance in collaborative filtering. The key idea is to use contrastive loss to generate augmentation-invariant embeddings by maximizing the Mutual Information between different augmented views of the same ...
- research-articleJanuary 2025
Chasing fairness in graphs: a GNN architecture perspective
AAAI'24/IAAI'24/EAAI'24: Proceedings of the Thirty-Eighth AAAI Conference on Artificial Intelligence and Thirty-Sixth Conference on Innovative Applications of Artificial Intelligence and Fourteenth Symposium on Educational Advances in Artificial IntelligenceArticle No.: 2367, Pages 21214–21222https://doi.org/10.1609/aaai.v38i19.30115There has been significant progress in improving the performance of graph neural networks (GNNs) through enhancements in graph data, model architecture design, and training strategies. For fairness in graphs, recent studies achieve fair representations ...
- research-articleMay 2024
Fair graph distillation
NIPS '23: Proceedings of the 37th International Conference on Neural Information Processing SystemsArticle No.: 3535, Pages 80644–80660As graph neural networks (GNNs) struggle with large-scale graphs due to high computational demands, graph data distillation promises to alleviate this issue by distilling a large real graph into a smaller distilled graph while maintaining comparable ...
- research-articleMay 2024
Chasing fairness under distribution shift: a model weight perturbation approach
NIPS '23: Proceedings of the 37th International Conference on Neural Information Processing SystemsArticle No.: 2793, Pages 63931–63944Fairness in machine learning has attracted increasing attention in recent years. The fairness methods improving algorithmic fairness for in-distribution data may not perform well under distribution shifts. In this paper, we first theoretically ...
- research-articleMay 2024
Winner-take-all column row sampling for memory efficient adaptation of language model
- Zirui Liu,
- Guanchu Wang,
- Shaochen Zhong,
- Zhaozhuo Xu,
- Daochen Zha,
- Ruixiang Tang,
- Zhimeng Jiang,
- Kaixiong Zhou,
- Vipin Chaudhary,
- Shuai Xu,
- Xia Hu
NIPS '23: Proceedings of the 37th International Conference on Neural Information Processing SystemsArticle No.: 150, Pages 3402–3424As the model size grows rapidly, fine-tuning the large pre-trained language model has become increasingly difficult due to its extensive memory usage. Previous works usually focus on reducing the number of trainable parameters in the network. While the ...
- research-articleAugust 2023
Probabilistic masked attention networks for explainable sequential recommendation
- Huiyuan Chen,
- Kaixiong Zhou,
- Zhimeng Jiang,
- Chin-Chia Michael Yeh,
- Xiaoting Li,
- Menghai Pan,
- Yan Zheng,
- Xia Hu,
- Hao Yang
IJCAI '23: Proceedings of the Thirty-Second International Joint Conference on Artificial IntelligenceArticle No.: 230, Pages 2068–2076https://doi.org/10.24963/ijcai.2023/230Transformer-based models are powerful for modeling temporal dynamics of user preference in sequential recommendation. Most of the variants adopt the Softmax transformation in the self-attention layers to generate dense attention probabilities. However, ...
- research-articleJuly 2023
DIVISION: memory efficient training via dual activation precision
ICML'23: Proceedings of the 40th International Conference on Machine LearningArticle No.: 1496, Pages 36036–36057Activation compressed training provides a solution towards reducing the memory cost of training deep neural networks (DNNs). However, state-of-the-art work combines a search of quantization bitwidth with the training, which makes the procedure ...
- research-articleJuly 2023
Graph mixup with soft alignments
ICML'23: Proceedings of the 40th International Conference on Machine LearningArticle No.: 879, Pages 21335–21349We study graph data augmentation by mixup, which has been used successfully on images. A key operation of mixup is to compute a convex combination of a pair of inputs. This operation is straightforward for grid-like data, such as images, but challenging ...
- articleJuly 2023
Adaptive RiskAware Bidding with Budget Constraint in Display Advertising
ACM SIGKDD Explorations Newsletter (SIGKDD), Volume 25, Issue 1Pages 73–82https://doi.org/10.1145/3606274.3606281Real-time bidding (RTB) has become a major paradigm of display advertising. Each ad impression generated from a user visit is auctioned in real time, where demand-side plat- form (DSP) automatically provides bid price usually relying on the ad impression ...
- research-articleApril 2023
Hierarchy-Aware Multi-Hop Question Answering over Knowledge Graphs
WWW '23: Proceedings of the ACM Web Conference 2023Pages 2519–2527https://doi.org/10.1145/3543507.3583376Knowledge graphs (KGs) have been widely used to enhance complex question answering (QA). To understand complex questions, existing studies employ language models (LMs) to encode contexts. Despite the simplicity, they neglect the latent relational ...
- short-paperOctober 2022
BED: A Real-Time Object Detection System for Edge Devices
- Guanchu Wang,
- Zaid Pervaiz Bhat,
- Zhimeng Jiang,
- Yi-Wei Chen,
- Daochen Zha,
- Alfredo Costilla Reyes,
- Afshin Niktash,
- Gorkem Ulkar,
- Erman Okman,
- Xuanting Cai,
- Xia Hu
CIKM '22: Proceedings of the 31st ACM International Conference on Information & Knowledge ManagementPages 4994–4998https://doi.org/10.1145/3511808.3557168Deploying deep neural networks (DNNs) on edge devices provides efficient and effective solutions for the real-world tasks. Edge devices have been used for collecting a large volume of data efficiently in different domains. DNNs have been an effective ...
- research-articleApril 2022
Geometric Graph Representation Learning via Maximizing Rate Reduction
WWW '22: Proceedings of the ACM Web Conference 2022Pages 1226–1237https://doi.org/10.1145/3485447.3512170Learning discriminative node representations benefits various downstream tasks in graph analysis such as community detection and node classification. Existing graph representation learning methods (e.g., based on random walk and contrastive learning) ...