1 code implementation • NeurIPS 2023 • Namyong Park, Ryan Rossi, Xing Wang, Antoine Simoulin, Nesreen Ahmed, Christos Faloutsos
The choice of a graph learning (GL) model (i. e., a GL algorithm and its hyperparameter settings) has a significant impact on the performance of downstream tasks.
1 code implementation • 16 Mar 2024 • Namyong Park, Xing Wang, Antoine Simoulin, Shuai Yang, Grey Yang, Ryan Rossi, Puja Trivedi, Nesreen Ahmed
To address these limitations, the forward-forward algorithm (FF) was recently proposed as an alternative to BP in the image classification domain, which trains NNs by performing two forward passes over positive and negative data.
1 code implementation • 14 Feb 2024 • Nadav Schneider, Niranjan Hasabnis, Vy A. Vo, Tal Kadosh, Neva Krien, Mihai Capotă, Guy Tamir, Ted Willke, Nesreen Ahmed, Yuval Pinter, Timothy Mattson, Gal Oren
This study first investigates the performance of state-of-the-art language models in generating MPI-based parallel programs.
no code implementations • 28 Jan 2024 • Le Chen, Arijit Bhattacharjee, Nesreen Ahmed, Niranjan Hasabnis, Gal Oren, Vy Vo, Ali Jannesari
Large language models (LLMs), as epitomized by models like ChatGPT, have revolutionized the field of natural language processing (NLP).
2 code implementations • 20 Dec 2023 • Tal Kadosh, Niranjan Hasabnis, Vy A. Vo, Nadav Schneider, Neva Krien, Mihai Capota, Abdul Wasay, Nesreen Ahmed, Ted Willke, Guy Tamir, Yuval Pinter, Timothy Mattson, Gal Oren
Specifically, we start off with HPC as a domain and build an HPC-specific LM, named MonoCoder, that is orders of magnitude smaller than existing LMs but delivers similar, if not better performance, on non-HPC and HPC tasks.
2 code implementations • 18 Aug 2023 • Tal Kadosh, Niranjan Hasabnis, Vy A. Vo, Nadav Schneider, Neva Krien, Abdul Wasay, Nesreen Ahmed, Ted Willke, Guy Tamir, Yuval Pinter, Timothy Mattson, Gal Oren
With easier access to powerful compute resources, there is a growing trend in the field of AI for software development to develop larger and larger language models (LLMs) to address a variety of programming tasks.
no code implementations • 2 Feb 2023 • Leonardo Cotta, Beatrice Bevilacqua, Nesreen Ahmed, Bruno Ribeiro
Existing causal models for link prediction assume an underlying set of inherent node factors -- an innate characteristic defined at the node's birth -- that governs the causal evolution of links in the graph.
no code implementations • 28 Dec 2022 • Ryan Aponte, Ryan A. Rossi, Shunan Guo, Jane Hoffswell, Nedim Lipka, Chang Xiao, Gromit Chan, Eunyee Koh, Nesreen Ahmed
In this work, we introduce a hypergraph representation learning framework called Hypergraph Neural Networks (HNN) that jointly learns hyperedge embeddings along with a set of hyperedge-dependent embeddings for each node in the hypergraph.
1 code implementation • 18 Jun 2022 • Namyong Park, Ryan Rossi, Nesreen Ahmed, Christos Faloutsos
In this work, we develop the first meta-learning approach for evaluation-free graph learning model selection, called MetaGL, which utilizes the prior performances of existing methods on various benchmark graph datasets to automatically select an effective model for the new graph, without any model training or evaluations.
no code implementations • 8 Jun 2022 • Xinyi Zheng, Ryan A. Rossi, Nesreen Ahmed, Dominik Moritz
Challenges arise as networks are often used across different domains (e. g., network science, physics, etc) and have complex structures.
1 code implementation • 5 Apr 2022 • Namyong Park, Ryan Rossi, Eunyee Koh, Iftikhar Ahamath Burhanuddin, Sungchul Kim, Fan Du, Nesreen Ahmed, Christos Faloutsos
Especially, deep graph clustering (DGC) methods have successfully extended deep clustering to graph-structured data by learning node representations and cluster assignments in a joint optimization framework.
no code implementations • 29 Nov 2021 • Aravind Reddy, Ryan A. Rossi, Zhao Song, Anup Rao, Tung Mai, Nedim Lipka, Gang Wu, Eunyee Koh, Nesreen Ahmed
In this paper, we introduce the online and streaming MAP inference and learning problems for Non-symmetric Determinantal Point Processes (NDPPs) where data points arrive in an arbitrary order and the algorithms are constrained to use a single-pass over the data as well as sub-linear memory.
no code implementations • NeurIPS Workshop LMCA 2020 • Aaron Zweig, Nesreen Ahmed, Theodore L. Willke, Guixiang Ma
The application of deep reinforcement learning (RL) to graph learning and meta-learning admits challenges from both topics.
no code implementations • 16 Jan 2020 • Ryan Rossi, Somdeb Sarkhel, Nesreen Ahmed
We propose causal inference models for this problem that leverage both the graph topology and time-series to accurately estimate local causal effects of nodes.