Xinyuan (Youki) Cao

Hi! I am Xinyuan Cao (曹馨元), a fourth-year PhD student in Machine Learning at Georgia Institute of Technology. I am fortunate to be advised by Prof. Santosh Vempala. Before joining Gatech, I received my Master's degree in Data Science from Columbia University, supervised by Prof. John Wright. I obtained my Bachelor's degree in Mathematics at Fudan University.

I have broad interest in machine learning, optimization and graph. My research mainly focuses on developing efficient and provable machine learning algorithms and build theories that inspire the machine learning practicers!

Email  /  Github  /  Google Scholar

profile photo
Talks
Contrastive Moments: Unsupervised Halfspace Learning in Polynomial Time
  • Nov 2023 - Georgia Tech ACO Student Seminar (Atlanta, GA)
  • Oct 2023 - LeT-All Mentorship Workshop (Virtual)
  • Sep 2023 - Georgia Tech ARC (Atlanta, GA)
  • Aug 2023 - SUFE TCS Seminar (Shanghai, China)
Provable Lifelong Learning of Representations
  • Mar 2023 - CMU Theory Lunch (Pittsburgh, PA)
  • Mar 2023 - Continual AI (CLAI) Seminar (Virtual)
  • Mar 2023 - AISTATS 2023 (Virtual)
  • Jan 2023 - Georgia Tech ACO Student Seminar (Atlanta, GA)
Publications
Machine Learning Theory
Contrastive Moments: Unsupervised Halfspace Learning in Polynomial Time
Xinyuan Cao, Santosh Vempala

NeurIPS 2023

We propose a provable and efficient unsupervised learning algorithm to learn the max margin classifier with linearly separable unlabeled data through a contrastive approach. The approach uses re-weighted first and second moments to compute the direction of the max margin classifier.

Provable Lifelong Learning of Representations
Xinyuan Cao, Weiyang Liu, Santosh Vempala

AISTATS 2022

We propose a lifelong learning algorithm that maintains and refines the internal feature representation and prove nearly matching upper and lower bounds on the total sample complexity. We also complement our analysis with an empirical study, where our method performs favorably on challenging realistic image datasets compared to state-of-the-art continual learning methods.

Graph Representation Learning
StructComp: Substituting propagation with Structural Compression in Training Graph Contrastive Learning
Shengzhong Zhang, Wenjie Yang, Xinyuan Cao, Hongwei Zhang, Zengfeng Huang

ICLR 2024

We propose a simple yet effective training framework called Structural Compression (StructComp) to do graph contrastive learning. Inspired by a sparse low-rank approximation on the diffusion matrix, StructComp trains the encoder with the compressed nodes. This allows the encoder not to perform any message passing during the training stage, and significantly reduces the number of sample pairs in the contrastive loss. 

Graph-Level Embedding for Time-Evolving Graphs
Lili Wang, Chenghan Huang, Weicheng Ma, Xinyuan Cao, Soroush Vosoughi

WWW 2023

We present a novel method for temporal graph-level embedding that involves constructing a multilayer graph and using a modified random walk with temporal backtracking to generate temporal contexts for the graph’s nodes.

Graph Embedding via Diffusion-Wavelets-Based Node Feature Distribution Characterization
Lili Wang, Chenghan Huang, Weicheng Ma, Xinyuan Cao, Soroush Vosoughi

CIKM 2021

We propose a novel unsupervised whole graph embedding method. Our method uses spectral graph wavelets to capture topological similarities on each k-hop sub-graph between nodes and uses them to learn embeddings for the whole graph.

Preprints
Learning-Augmented B-Trees
Xinyuan Cao, Jingbang Chen, Li Chen, Chris Lambert, Richard Peng, Daniel Sleator

We study learning-augmented binary search trees (BSTs) and B-Trees via Treaps with composite priorities. It also gives the first B-Tree data structure that can provably take advantage of localities in the access sequence via online self-reorganization. The data structure is robust to prediction errors and handles insertions, deletions, as well as prediction updates.

Towards Understanding Neural Collapse: The Effects of Batch Normalization and Weight Decay
Leyan Pan, Xinyuan Cao

We investigate the interrelationships between batch normalization (BN), weight decay, and proximity to the Neural Collapse (NC) structure. Experimental evidence substantiates our theoretical findings, revealing a pronounced occurrence of NC in models incorporating BN and appropriate weight-decay values.