- Instructor: Kimon Fountoulakis
- Seminar Time: Tu/Th 1pm to 2:20pm
- Office hours: Monday 1pm to 2pm.
Learning from multi-modal datasets is currently one of the most prominent topics in artificial intelligence. The reason behind this trend is that many applications, such as recommendation systems and fraud detection, require the combination of different types of data. In addition, it is often the case that data exhibit relations that need to be captured for downstream applications. In this course, we are interested in multi-modal data that combine a graph—i.e., a set of nodes and edges—with attributes for each node and/or edge. The attributes of the nodes and edges capture information about the nodes and edges themselves, while the edges between the nodes capture relations among them. Capturing relations is particularly helpful for applications where we are trying to make predictions for nodes given neighborhood data.
One of the most prominent and principled ways of handling such multi-modal data for downstream tasks such as node classification is graph neural networks. Graph neural network models can mix hand-crafted or automatically learned attributes of the nodes while taking into account relational information among the nodes. Therefore, the output vector representation of the graph neural network contains global and local information about the nodes. This contrasts with neural networks that only learn from the attributes of entities.
This seminar-based course will cover seminal work in the space of graph neural networks. Below I provide the topics and architectures which we will study during the course.
Topics
- Generalization performance of graph neural networks
- Expressive power of graph neural networks
- Large language models and graphs
- Neural algorithmic reasoning
- Generative graph neural networks
- Self-supervised learning in graphs
- Oversmoothing
- Scalability
Architectures:
- Spectral and spatial convolutional graph neural networks
- Graph attention networks
- Invariant and equivariant graph neural networks
- General message passing graph neural networks
- Higher-order graph neural networks
- Graph neural networks for heterogeneous graphs
We will focus on both practical and theoretical aspects of machine learning on graphs. Practical aspects include, scalability and performance on real data. Examples of theoretical questions include: what does convolution do to the input data? Does convolution improve generalization compared to not using a graph? How do multiple convolutions change the data and how do they affect generalization?
Course structure: The seminar is based on weekly student presentations, discussions, a midterm and a final project.
The schedule below is subject to change:
- Geometric Deep Learning, Michael M. Bronstein, Joan Bruna, Taco Cohen, Petar Veličković, 2021
- Theory of Graph Neural Networks: Representation and Learning, Stefanie Jegelka, 2022
- Graph Representation Learning Book, William L. Hamilton, 2020
- Graph Neural Networks, Lingfei Wu, Peng Cui, Jian Pei, Liang Zhao, (2022)
- Machine Learning with Graphs, Jure Leskovec, Stanford
- Graph Representation Learning, William L. Hamilton, McGill
- Introduction to Graph Neural Networks, Xavier Bresson, Nanyang Techinical University and NYU
- Recent Developments in Graph Network Architectures, Xavier Bresson, Nanyang Techinical University
- Benchmarking GNNs, Xavier Bresson, Nanyang Techinical University
- Foundations of Graph Neural Networks, Petar Veličković, DeepMind
- Geometric Deep Learning Course
- Machine Learning for the Working Mathematician: Geometric Deep Learning, Geordie Williamson, The University of Syndney
- Advanced lectures on community detection, Laurent Massoulie, INRIA Paris
- Open Graph Benchmark
- CLRS Algorithmic Reasoning Benchmark
- The CLRS-Text Algorithmic Reasoning Language Benchmark
- PyTorch Geometric Datasets
- Open Graph Benchmark
- HyperGraphs
- TUDatasets
- Non Homophily Benchmarks
- Graph Learning Benchmarks
- Hetionet
- Heterogeneous graph benchmarks
- Long Range Graph Benchmark
- IGB-Datasets
- Class Participation: 15%
- Midterm Project: 20%
- Presentations: 25%
- Final Project: 40%
This is a 3-page paper, along with (if relevant) the source code of your project, including instructions on how to run it. You may use your midterm project as a foundation for your final project, which will be 6 pages. Please see the next section below for details.
Options for the midterm project:
- Option A (Empirical evaluation): Pick a problem that interests you. Implement and experiment with several graph neural network methods to tackle this problem.
- Option B (Method design): Identify a problem for which there are no satisfying approaches. Develop a new graph neural network architecture to tackle this problem. Analyze theoretically and/or empirically the performance of your technique.
- Option C (Theoretical analysis): Identify a problem or a graph neural network architecture for which theoretical performance (e.g., complexity, performance on random data, expressivity) is not well understood. Analyze the properties of this problem or technique.
Information about the project template and the source code is given below.
- Project Paper: The project papers will be 3 pages. You can have extra pages for the references and the appendix. They will be written in the two-column ICML format, using the ICML template which you can find in the corresponding website.
- Project Source Code: Please put your source code into github and include a link in your project writeup. On the github page, please document exactly how to run your source code.
There is one main deliverable of your project, a 6-page paper and (if relevant) the source code of your project with instructions to run your code. Note that you are allowed to use the midterm project (3 pages) as a foundation for this project.
Options for the final project:
- Option A (Empirical evaluation): Pick a problem that interests you. Implement and experiment with several graph neural network methods to tackle this problem.
- Option B (Method design): Identify a problem for which there are no satisfying approaches. Develop a new graph neural network architecture to tackle this problem. Analyze theoretically and/or empirically the performance of your technique.
- Option C (Theoretical analysis): Identify a problem or a graph neural network architecture for which theoretical performance (e.g., complexity, performance on random data, expressivity) is not well understood. Analyze the properties of this problem or technique.
Information about the project template and the source code is given below.
- Project Paper: The project papers will be 6 pages. You can have extra pages for the references and the appendix. They will be written in the two-column ICML format, using the ICML template which you can find in the corresponding website.
- Project Source Code: Please put your source code into github and include a link in your project writeup. On the github page, please document exactly how to run your source code.
Although not required for the course, keep in mind that I am more than happy to help you publish your final project. For example, in CS886 2024, Robert Wang published his final project at NeurIPS 2024.
Each student will be doing 2 presentations (estimated number, based on previous years) in the term. Each presentation will be about 40 to 50 minutes long plus questions. Here are the important points summarizing what you have to do for your presentations.
- You must present with slides. The content in your slides should be your own but you can use others' materials, e.g., figures from the paper we are reading, when necessary and by crediting your source on your slide.
- Please have a separate slide, or set of slides, for each of the 4 questions below:
- What is the problem?
- Why is it important?
- Why don't previous methods work on that problem?
- What is the solution to the problem the authors propose?
- What interesting research questions does the paper raise?
- It is very helpful to demonstrate the ideas in the paper through examples. So try to have examples in your presentation.
The University of Waterloo Senate Undergraduate Council has also approved the following message outlining University of Waterloo policy on academic integrity and associated policies.
In order to maintain a culture of academic integrity, members of the University of Waterloo community are expected to promote honesty, trust, fairness, respect and responsibility. Check the Office of Academic Integrity's website for more information. All members of the UW community are expected to hold to the highest standard of academic integrity in their studies, teaching, and research. This site explains why academic integrity is important and how students can avoid academic misconduct. It also identifies resources available on campus for students and faculty to help achieve academic integrity in, and our, of the classroom.
A student who believes that a decision affecting some aspect of his/her university life has been unfair or unreasonable may have grounds for initiating a grievance. Read Policy 70 - Student Petitions and Grievances, Section 4. When in doubt please be certain to contact the department's administrative assistant who will provide further assistance.
A student is expected to know what constitutes academic integrity, to avoid committing academic offenses, and to take responsibility for his/her actions. A student who is unsure whether an action constitutes an offense, or who needs help in learning how to avoid offenses (e.g., plagiarism, cheating) or about “rules” for group work/collaboration should seek guidance from the course professor, academic advisor, or the Undergraduate Associate Dean. For information on categories of offenses and types of penalties, students should refer to Policy 71-Student Discipline. For typical penalties check Guidelines for the Assessment of Penalties.
Most students are unaware of the line between acceptable and unacceptable academic behaviour, especially when discussing assignments with classmates and using the work of other students. For information on commonly misunderstood academic offenses and how to avoid them, students should refer to the Faculty of Mathematics Cheating and Student Academic Discipline Policy.
A decision made or a penalty imposed under Policy 70, Student Petitions and Grievances (other than a petition) or Policy 71, Student Discipline may be appealed if there is a ground. A student who believes he/she has a ground for an appeal should refer to Policy 72 - Student Appeals.
The AccessAbility Services Office (AAS), located in Needles Hall, Room 1401, collaborates with all academic departments to arrange appropriate accommodations for students with disabilities without compromising the academic integrity of the curriculum. If you require academic accommodations to lessen the impact of your disability, please register with the AAS at the beginning of each academic term.