Talk by Ivan Ojeda-Ruiz on Friday 4/5

Title:  An Overview of Graph Neural Network Propagation Modules Abstract: Graph Neural Networks (GNNs) offer powerful methods to analyze the complex relationships inherent in graph-structured data. Propagation modules sit at the core of GNNs, driving the process of learning representations by aggregating information across nodes. This talk explores the landscape of GNN propagation modules. We’ll…

Talk February 9 by Suho Oh: Intro to graph neural networks

Abstract: We’ll go over graph neural networks (GNN) that has been gaining popularity over the last few years. We will cover the definition, some different models of the similar flavor, and if time permits some coding examples on how one can use them for your purpose. Resources: https://web.stanford.edu/class/cs224w/slides/03-GNN1.pdf here is the slides going over basics of…

11/6/2023 Difference between SGD and ADAM by Xiaoxi Shen and Jialong Li

Abstract: With the growth of dimensionality of the data we encounter nowadays, the classical gradient descent methods need to speed up. In the presentation, we will start by reviewing the gradient descent and stochastic gradient descent and discuss their theoretical guarantees. After that, more commonly used optimization algorithms in deep learning will be introduced and…

Talk 11/3/23 Gradient descent by Usufu Nyakoojo and Humu Mohammed

Abstract: In training an artificial neural network, we need to find model parameters that optimize the cost function while ensuring sufficient learning rate of the machine. Gradient descent and in particular the stochastic gradient descent, comes in handy as an iterative optimization approach to address this need. In this talk, we shall present how this can be achieved i.e. how the cost function…

Back To Top
Skip to toolbar