Spring 2025 talk schedule

Date Talk by February 14 Christine Lee February 28 Christine Lee March 28 Geonhee Cho April 11 Jeremiah Birrell April 25 Jeremiah Birrell

Talk by Ivan Ojeda-Ruiz on Friday 4/5

Title:  An Overview of Graph Neural Network Propagation Modules Abstract: Graph Neural Networks (GNNs) offer powerful methods to analyze the complex relationships inherent in graph-structured data. Propagation modules sit at the core of GNNs, driving the process of learning representations by aggregating information across nodes. This talk explores the landscape of GNN propagation modules. We’ll…

Talk February 9 by Suho Oh: Intro to graph neural networks

Abstract: We’ll go over graph neural networks (GNN) that has been gaining popularity over the last few years. We will cover the definition, some different models of the similar flavor, and if time permits some coding examples on how one can use them for your purpose. Resources: https://web.stanford.edu/class/cs224w/slides/03-GNN1.pdf here is the slides going over basics of…

11/6/2023 Difference between SGD and ADAM by Xiaoxi Shen and Jialong Li

Abstract: With the growth of dimensionality of the data we encounter nowadays, the classical gradient descent methods need to speed up. In the presentation, we will start by reviewing the gradient descent and stochastic gradient descent and discuss their theoretical guarantees. After that, more commonly used optimization algorithms in deep learning will be introduced and…

Talk 11/3/23 Gradient descent by Usufu Nyakoojo and Humu Mohammed

Abstract: In training an artificial neural network, we need to find model parameters that optimize the cost function while ensuring sufficient learning rate of the machine. Gradient descent and in particular the stochastic gradient descent, comes in handy as an iterative optimization approach to address this need. In this talk, we shall present how this can be achieved i.e. how the cost function…

Talk 9/22/23: Linear regression by Xiaoxi Shen

Abstract: In this presentation, the framework of empirical risk minimization will be introduced. It will turn out that all the commonly used supervised machine learning models such as linear regression, logistic regression, support vector machines and neural networks are special examples under this framework. The excess risk of a regression problem will be computed, followed…