2% of ICML papers desk rejected because the authors used LLM in their reviews
https://blog.icml.cc/2026/03/18/on-violations-of-llm-review-policies/
#HackerNews #ICML2023 #LLM #Ethics #ResearchPublishing #AIinResearch #DeskRejection
2% of ICML papers desk rejected because the authors used LLM in their reviews
https://blog.icml.cc/2026/03/18/on-violations-of-llm-review-policies/
#HackerNews #ICML2023 #LLM #Ethics #ResearchPublishing #AIinResearch #DeskRejection
Aloha! If you’re looking for a simple, accurate and efficient approach for distributed GNN training that works for web-scale data with billions of edges, let’s chat 4:15pm today at Room 310 at #ICML2023 Localized Learning Workshop!
📑 Simplifying Distributed Neural Network Training on Massive Graphs: Randomized Partitions Improve Model Aggregation (https://arxiv.org/abs/2305.09887)
🕓 Sat 29 Jul, 4:15-5:00pm HDT
Shout out to my colleagues at Amazon Science -- it is an exciting internship ride!
Distributed training of GNNs enables learning on massive graphs (e.g., social and e-commerce networks) that exceed the storage and computational capacity of a single machine. To reach performance comparable to centralized training, distributed frameworks focus on maximally recovering cross-instance node dependencies with either communication across instances or periodic fallback to centralized training, which create overhead and limit the framework scalability. In this work, we present a simplified framework for distributed GNN training that does not rely on the aforementioned costly operations, and has improved scalability, convergence speed and performance over the state-of-the-art approaches. Specifically, our framework (1) assembles independent trainers, each of which asynchronously learns a local model on locally-available parts of the training graph, and (2) only conducts periodic (time-based) model aggregation to synchronize the local models. Backed by our theoretical analysis, instead of maximizing the recovery of cross-instance node dependencies -- which has been considered the key behind closing the performance gap between model aggregation and centralized training -- , our framework leverages randomized assignment of nodes or super-nodes (i.e., collections of original nodes) to partition the training graph such that it improves data uniformity and minimizes the discrepancy of gradient and loss function across instances. In our experiments on social and e-commerce networks with up to 1.3 billion edges, our proposed RandomTMA and SuperTMA approaches -- despite using less training data -- achieve state-of-the-art performance and 2.31x speedup compared to the fastest baseline, and show better robustness to trainer failures.
"Causal Deep Learning"
Jeroen Berrevoets, Krzysztof Kacprzyk, Zhaozhi Qian, Mihaela van der Schaar
Prof Mihaela van der Schaar just gave a very exciting talk at the #counterfactuals workshop at #icml2023 on her lab's CDL framework dealing explicitly and with concepts from #causality and #machine-learning. Definitely going on my reading list.
Causality has the potential to truly transform the way we solve a large number of real-world problems. Yet, so far, its potential largely remains to be unlocked as causality often requires crucial assumptions which cannot be tested in practice. To address this challenge, we propose a new way of thinking about causality -- we call this causal deep learning. Our causal deep learning framework spans three dimensions: (1) a structural dimension, which incorporates partial yet testable causal knowledge rather than assuming either complete or no causal knowledge among the variables of interest; (2) a parametric dimension, which encompasses parametric forms that capture the type of relationships among the variables of interest; and (3) a temporal dimension, which captures exposure times or how the variables of interest interact (possibly causally) over time. Causal deep learning enables us to make progress on a variety of real-world problems by leveraging partial causal knowledge (including independencies among variables) and quantitatively characterising causal relationships among variables of interest (possibly over time). Our framework clearly identifies which assumptions are testable and which ones are not, such that the resulting solutions can be judiciously adopted in practice. Using our formulation we can combine or chain together causal representations to solve specific problems without losing track of which assumptions are required to build these solutions, pushing real-world impact in healthcare, economics and business, environmental sciences and education, through causal deep learning.
Excited to present my poster about regression on Grassmann manifolds for the analysis of multi-condition single-cell data at #ICML2023 in the TAG workshop at 11:10am :)
Read the extended abstract at openreview.net/pdf?id=MrE4jL0… and check out the software at https://github.com/const-ae/lemur
Check out this amazing Google DeepMind paper presented at [ICML] Int'l Conference on Machine Learning #ICML2023!
"Can Neural Network Memorization be localized?"
Memorization in neural networks is a process where a specific set of neurons in different layers of the model store and remember specific patterns or data.
Hanie Sedghi et al conducted groundbreaking research using a technique called "Example-tied Dropout." This approach allows them to pinpoint and limit memorization to certain fixed neurons that can be discarded during the testing phase.
Read more here>https://arxiv.org/pdf/2307.09542.pdf