Publications
(*) denotes equal contribution. (**) indicates alphabetical ordering authorship
2024
- arXivTowards Understanding How Transformer Perform Multi-step Reasoning with Matching OperationarXiv preprint arXiv:2405.15302, 2024
- arXivOn the Expressive Power of a Variant of the Looped TransformerarXiv preprint arXiv:2402.13572, 2024
- ICMLExact Conversion of In-Context Learning to Model Weights in Linearized-Attention TransformersInternational Conference on Machine Learning, 2024
- ICMLThe Surprising Effectiveness of Skip-Tuning in Diffusion SamplingInternational Conference on Machine Learning, 2024
- ICMLReferee Can Play: An Alternative Approach to Conditional Generation via Model InversionInternational Conference on Machine Learning, 2024
- CVPRAccelerating Diffusion Sampling with Optimized Time StepsProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2024
- ECCVJointDreamer: Ensuring Geometry Consistency and Text Congruence in Text-to-3D Generation via Joint Score DistillationEuropean Conference on Computer Vision, 2024
- ICLRElucidating The Design Space of Classifier-Guided Diffusion GenerationInternational Conference on Learning Representations, 2024
- JMLRRandom Smoothing Regularization in Kernel Gradient Descent LearningJournal of Machine Learning Research, 25(284), 2024
- IJCAIDeciphering the Projection Head: Representation Evaluation Self-supervised LearningInternational Joint Conference on Artificial Intelligence, 2024
- JMLRMinimax Optimal Deep Neural Network Classifiers Under Smooth Decision BoundaryJournal of Machine Learning Research, Accept after minor revision, 2024
2023
- NeurIPS
Spotlight Complexity Matters: Rethinking the Latent Space for Generative ModelingAdvances in Neural Information Processing Systems, 2023 - NeurIPSDiff-Instruct: A Universal Approach for Transferring Knowledge From Pre-trained Diffusion ModelsAdvances in Neural Information Processing Systems, 2023
- arXivTraining Energy-Based Models with Diffusion Contrastive DivergencesarXiv preprint arXiv:2307.01668, 2023
- arXivBoosting Visual-Language Models by Exploiting Hard SamplesarXiv preprint arXiv:2305.05208, 2023
- arXivConsistentNeRF: Enhancing Neural Radiance Fields with 3D Consistency for Sparse View SynthesisarXiv preprint arXiv:2305.11031, 2023
- CVPRContraNeRF: Generalizable Neural Radiance Fields for Synthetic-to-real Novel View Synthesis via Contrastive LearningIn Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023
- AISTATSInducing Neural Collapse in Deep Long-tailed LearningIn International Conference on Artificial Intelligence and Statistics, 2023
- ICMLExplore and Exploit the Diverse Knowledge in Model Zoo for Domain GeneralizationIn International Conference on Machine Learning, 2023
- UAIExact Count of Boundary Pieces of ReLU Classifiers: Towards the Proper Complexity Measure for ClassificationIn Conference on Uncertainty in Artificial Intelligence, 2023
- ICLRYour Contrastive Learning Is Secretly Doing Stochastic Neighbor EmbeddingInternational Conference on Learning Representations, 2023
- TMLRContinual Learning by Modeling Intra-Class VariationTransactions on Machine Learning Research, 2023
2022
- NeurIPSZooD: Exploiting Model Zoo for Out-of-Distribution GeneralizationAdvances in Neural Information Processing Systems, 2022
- NeurIPS
Spotlight Understanding Square Loss in Training Overparametrized Neural Network ClassifiersAdvances in Neural Information Processing Systems, 2022
2021
- AISTATSRegularization Matters: A Nonparametric Perspective on Overparametrized Neural NetworkIn International Conference on Artificial Intelligence and Statistics, 2021
2020
- Thesis
- arXivSharp Rate of Convergence for Deep Neural Network Classifiers Under the Teacher-student SettingarXiv preprint arXiv:2001.06892, 2020
- AJPEInter-rater Reliability of Web-based Calibrated Peer Review within a Pharmacy CurriculumAmerican Journal of Pharmaceutical Education, 2020
2018
- arXiv