Longer write-up on the #NeurIPS2021 consistency experiment now up on arxiv: https://arxiv.org/abs/2306.03262

Some takeaways:

- There was noise in 2014; there's noise now. Conference growth hasn't changed this much.

- Spotlights/orals are especially noisy. Being more selective would make decisions more arbitrary.

- The amount of noise in which papers get flagged for ethics review is also striking.

Has the Machine Learning Review Process Become More Arbitrary as the Field Has Grown? The NeurIPS 2021 Consistency Experiment

We present the NeurIPS 2021 consistency experiment, a larger-scale variant of the 2014 NeurIPS experiment in which 10% of conference submissions were reviewed by two independent committees to quantify the randomness in the review process. We observe that the two committees disagree on their accept/reject recommendations for 23% of the papers and that, consistent with the results from 2014, approximately half of the list of accepted papers would change if the review process were randomly rerun. Our analysis suggests that making the conference more selective would increase the arbitrariness of the process. Taken together with previous research, our results highlight the inherent difficulty of objectively measuring the quality of research, and suggest that authors should not be excessively discouraged by rejected work.

arXiv.org
Thinking about a great conversation starter for tomorrow? How about discussing how authors rank their own papers before and after they see their #NeurIPS2021 reviews https://blog.neurips.cc/2022/11/27/how-do-authors-perceptions-of-their-papers-compare-with-co-authors-perceptions-and-peer-review-decisions/
How do Authors’ Perceptions of their Papers Compare with Co-authors’ Perceptions and Peer-review Decisions? – NeurIPS Blog

Announcing the public release of the #̶N̶e̶u̶r̶I̶P̶S̶2̶0̶2̶2̶ #NeurIPS2021 (😅) RETINA Benchmark:

A suite of tasks evaluating the reliability of uncertainty quantification methods like Deep Ensembles, MC Dropout, Parameter- and Function-Space VI, and more.

Paper: https://arxiv.org/abs/2211.12717
Code+Checkpoints: https://rebrand.ly/retina-benchmark

#NewPaper #arxiv #PaperThread
🧵 below👇🏾 [0/N]

Benchmarking Bayesian Deep Learning on Diabetic Retinopathy Detection Tasks

Bayesian deep learning seeks to equip deep neural networks with the ability to precisely quantify their predictive uncertainty, and has promised to make deep learning more reliable for safety-critical real-world applications. Yet, existing Bayesian deep learning methods fall short of this promise; new methods continue to be evaluated on unrealistic test beds that do not reflect the complexities of downstream real-world tasks that would benefit most from reliable uncertainty quantification. We propose the RETINA Benchmark, a set of real-world tasks that accurately reflect such complexities and are designed to assess the reliability of predictive models in safety-critical scenarios. Specifically, we curate two publicly available datasets of high-resolution human retina images exhibiting varying degrees of diabetic retinopathy, a medical condition that can lead to blindness, and use them to design a suite of automated diagnosis tasks that require reliable predictive uncertainty quantification. We use these tasks to benchmark well-established and state-of-the-art Bayesian deep learning methods on task-specific evaluation metrics. We provide an easy-to-use codebase for fast and easy benchmarking following reproducibility and software design principles. We provide implementations of all methods included in the benchmark as well as results computed over 100 TPU days, 20 GPU days, 400 hyperparameter configurations, and evaluation on at least 6 random seeds each.

arXiv.org

RT @[email protected]

The WHY'21 workshop "Causal Inference & Machine Learning: Why now?" will take place this Monday at #NeurIPS2021. Our goal is to bring CI & ML researchers together to discuss the nextgen AI! Program https://why21.causalai.net/ (joint w/ @[email protected], @[email protected], Y Bengio, T Sejnowski)

🐦🔗: https://twitter.com/eliasbareinboim/status/1469811587741536258

WHY21 - Causal Inference & Machine Learning: Why now?

What implications do cultural evolution & cooperation have for cooperative AI? #NeurIPS2021 talk: https://t.co/EGNw8jjynh

Insights include diffs in constraints b/w genetic & cultural evolution vs machine learning; interaction b/w scales of cooperation.

Live Q&A Tues 14th 7 PST

NeurIPS 2021

Conference Platform

Cyber Valley again at the top of the list at NeurIPS

A total of 50 papers from Cyber Valley researchers will be presented at the 35th Conference on Neural Information Processing Systems (NeurIPS). With the COVID-19 pandemic still ongoing, the world’s leading machine learning conference will once again take place online from December 6 to 12.

1/12 Como falei outro dia, tivemos um artigo do nosso grupo aceito no #WHY21, o workshop de Inferência Causal do #NeurIPS2021. O título é "Reliable causal discovery based on mutual information supremum principle for finite datasets " e pode ser lido aqui: https://why21.causalai.net/papers/WHY21_24.pdf

Para quem tiver interesse, os papers aceitos no WHY 2021 (o workshop de inferência causal do NeurIPS 2021) já estão disponíveis para todos lerem!

https://why21.causalai.net/papers.html #NeurIPS2021

WHY21 - Causal Inference & Machine Learning: Why now?