CSE6392
|
| Lecture |
NH 202 | Friday 1:00-3:50 PM |
| Instructor |
Junzhou Huang | ERB 650 | Office hours: Friday 3:50-6:00 PM |
| Request |
Basic math and programming background; Basic learning and vision background preferred |
| Textbook |
None |
This course will provide an overview of the current state-of-the-art of machine learning techniques in computer vision, data mining and bioinformatics by studying a set of cutting-edge advanced topics in these areas. Several selected research topics reflect the current state in these fields. The main objective of this course is to review cutting-edge learning research in big data through lectures covering the underlying statistical & mathematical concepts and deep learning algorithms, paper reading, and implementation. The instructor will work with students on building ideas, performing experiments, and writing papers. Students can decide to submit his/her results to a learning/mining/vision related conference, or just play with funs.
The course is application-driven and includes advacnced topics
in machine learning, computer vision and bioinformatics, such as different learning
techniques and advanced vision tools in different applications. It will also
include selected topics relating to the machine learning theory and techniques.
The course will provide the participants with a thorough background in current
research in these areas, as well as to promote greater awareness and interaction
between multiple research groups within the university. The course material
is well suited for students in computer science, computer engineering, electrical
engineering and biomedical engineering.
| Fri Jan 23: Graph Neural Neural Networks (Slides) |
|
| Fri Jan 30: Graph Neural Neural Networks (Slides)
|
|
| Fri Feb 6: Zheng, Lin, et al. "A reparameterized discrete diffusion model for text generation." arXiv preprint arXiv:2302.05737 (2023). Presented by Hu xiao, Zheng Zheng L. Floridi and M. Chiriatti, "GPT-3: Its Nature, Scope, Limits, and Consequences", Minds and Machines, 2020. Presented by Harika Attipatla,Joel Ishika Reddy Kandukuri J. Devlin, et. al., "BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding", NAACL-HLT, 2019. Presented by Mathan Raj Arumugam Rajkumar, Praveen Anand |
|
| Fri Feb 13: B. Zhang, et. al., "Beyond Weisfeiler-Lehman: A Quantitative Framework for GNN Expressiveness", ICLR 2024. Presented by Humera Sabir, Fatima Farooq S. Yin, et. al., "Woodpecker: Hallucination Correction for Multimodal Large Language Models", arXiv:2310.16045, 2023. Presented by Muhan Zhang
|
|
|
W. Lin, et. al., "Generative Causal Explanations for Graph Neural Networks", ICML 2021. Presented by Dhanush Srinivas Y. Rong, et. al., "DropEdge: Towards Deep Graph Convolutional Networks on Node Classification", ICLR 2020. Presented by Nantha kumar Ashok Anand , Mukesh Eswaran E. Hu et. al., "LoRA: Low-Rank Adaptation of Large Language Models", ICLR 2022. Presented by Chaitanya Krishna Namburi, Kasi Rama Rao Sripalasetty |
|
| Fri Feb 27: Daniel Alexander Alber, et. al., "Medical Large Language Models are Vulnerable to Data-Poisoning Attacks", Nature Medicine, January 2025. Presented by Harmanpreet kaur, Vidhi Thummar H. Nori et. al., "Can Generalist Foundation Models Outcompete Special-Purpose Tuning? Case Study in Medicine", 2023. Presented by Sai Bhargav Varigunda, Lakshmi Prashanth Challa J. Li, et. al., "Towards Black‑Box Membership Inference for Diffusion Models", ICML 2025. Presented by Hongzhuo Chen, Shudong Lai. |
|
|
J. Li, et. al., "BLIP-2: Bootstrapping Language-Image Pre-training with Frozen Image Encoders and Large Language Models", ICML 2023. Presented by Jason Soundara Rajan, Jeran Joel Belavendran D. Kondratyuk et al., "VideoPoet: A Large Language Model for Zero-Shot Video Generation", ICML 2024. Presented by Arun Sabarish Krishnaswamy Ganesan, Kaushik Suresh S. Nie, et. al., "Large Language Diffusion Models", NeurIPS 2025. Presented Srinivasa Sai Abhijit Challapalli, Rozsa Zaruba |
|
|
P. Lewis et. al., "Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks", NeurIPS 2020. Presented by Ninghui Hao L. Ouyang et. al., "Training Language Models to Follow Instructions with Human Feedback", NeurIPS 2022. Presented by Ayesha Siddiqua R. Ying, et. al., "GNNExplainer: Generating Explanations for Graph Neural Networks", NeurIPS 2019. Presented by Bharani Bathula, Kabilan Rajendran |
|
|
W. Hamilton, et. al., "Inductive Representation Learning on Large Graphs", NIPS 2017. Presented by Likhitha Pericharla P. Veličković, et. al., "Graph Attention Networks", ICLR 2018. Presented by Akshay Varma Vegesna, Rohitha Sreya Namburi A. Vaswani, et. al., "Attention is All You Need", NIPS 2017. Presented by Roshan goud Rachakonda,Hemanth Reddy Majji |
|
Fri Apr 3: A. Jiang et al., "Mixtral of Experts", arXiv:2401.04088, January 2024. Presented by Xianhui Wu, Xiaoyan Shen "Structure-Aware Transformer for Graph Representation Learning", ICML 2022. Presented by Zeren li, Khan Fahim M. Henaff, et al., "Deep Convolutional Networks on Graph-Structured Data", 2015. Presented by Karthik Pellakuru , Siva Venkata Kaushik Pulipati |
|
| Fri Apr 10: D. Guo, et. al., "DeepSeek-R1 incentivizes reasoning in LLMs through reinforcement learning", Nature, 2025. Presented by Binamra Aryal, Jayanth Kotari T.Kipf and M.Welling, "Semi-Supervised Classification with Graph Convolutional Networks", ICLR 2017. Presented by Shreya Umesh Naidu, Kummathi Sai Likhith Reddy Y. Chebotar, et. al., "Q-Transformer: Scalable Offline Reinforcement Learning via Autoregressive Q-Functions". Presented by Prabhakaran Annadurai |
|
Fri Apr 17: N. Muennighoff, et. al., "Generative Representational Instruction Tuning", ICLR 2025. Presented by Songling Bai S. Semnani, et. al., "WikiChat: Stopping the Hallucination of Large Language Model Chatbots by Few-Shot Grounding on Wikipedia", EMNLP 2023. Presented by Sanjukktha Senthil Kumar, Chandrasekar Vasan P. Hase, et. al., "The Unreasonable Effectiveness of Easy Training Data for Hard Tasks", arXiv:2401.06751, January 2024 . Presented by Saakshi Jignesh Bhatt, Hrishitha Sai Kakarla |
|
|
Y. Li, et. al., "Evaluating Object Hallucination in Large Vision-Language Models", EMNLP 2023. Presented by Harshini Yallabandi, Arfan Basha Shaik C. Chen et al., “A Statistical and Multi-Perspective Revisiting of MIAs in LLMs”, ACL 2025. Presented by Yasho Mouli Karpurapu , Anusha Gadhiraju C. Ying et al., “Do Transformers Really Perform Bad for Graph Representation?”, NeurIPS 2021. Presented by Jatin Muddam, Anirudhda Yadav |
|
|
|
Each group has two members at most. Each group will select at least one paper from the following paper list and then be scheduled to present their selected papers in our class. Her/his final grade in this class will be mainly related with the peformance of her/his presentation.
Hallucination and Safety
If you do not believe a grade on a particular assignment is correct, you may appeal the grade in writing (email) within 5 class days. Grade appeals must be ppealed to the appropriate GTA firstly, then to your instructor if necessary. Please refer to the UTA Catalog for the detailed guide of grade appeals.