On the bird site, @nicolaspapernot posted a nice summary of @timnitGebru #SaTML thought-provoking keynote "Eugenics and the Promise of Utopia through Artificial General Intelligence". Video will become available online at conference web site.

My main take away: targeting (promising) AGI is inherently dangerous -- we should focus on "well-scoped systems" instead.

https://twitter.com/NicolasPapernot/status/1623885631930658816

#AGI #ArtificialGeneralIntelligence #TimnitGebru

Nicolas Papernot on Twitter

“A thread about @timnitGebru 's thought-provoking keynote "Eugenics and the Promise of Utopia through Artificial General Intelligence" @satml_conf (Note that we will release the recording to the keynote soon with the rest of the @satml_conf talks)”

Twitter

Getting ready for @timnitGebru ‘s (remote) keynote at the IEEE conference on Secure and Trustworthy Machine Learning.

Title: “Eugenics and the Promise of Utopia through Artificial General Intelligence”

#satml #agi #ArtificialGeneralIntelligence

Since some people have been asking, here's a preprint:
https://arxiv.org/abs/2208.13904

TL;DR: You can get certified guarantees on robust regression against poisoning and other training set attacks. The trick is to use a voting based predictor (like an ensemble or k-NN) and median.

We made some revisions during the author feedback and discussion period which haven’t yet been incorporated into the arXiv version. I’ll post again when we have the camera-ready version.
#SaTML #AdversarialML #NewPaper

Reducing Certified Regression to Certified Classification

Adversarial training instances can severely distort a model's behavior. This work investigates certified regression defenses, which provide guaranteed limits on how much a regressor's prediction may change under a training-set attack. Our key insight is that certified regression reduces to certified classification when using median as a model's primary decision function. Coupling our reduction with existing certified classifiers, we propose six new provably-robust regressors. To the extent of our knowledge, this is the first work that certifies the robustness of individual regression predictions without any assumptions about the data distribution and model architecture. We also show that existing state-of-the-art certified classifiers often make overly-pessimistic assumptions that can degrade their provable guarantees. We introduce a tighter analysis of model robustness, which in many cases results in significantly improved certified guarantees. Lastly, we empirically demonstrate our approaches' effectiveness on both regression and classification data, where the accuracy of up to 50% of test predictions can be guaranteed under 1% training-set corruption and up to 30% of predictions under 4% corruption. Our source code is available at https://github.com/ZaydH/certified-regression.

arXiv.org

Our paper on adversarially-robust regression was accepted to SaTML 2023 (https://satml.org) -- the first ever IEEE Conference on Secure and Trustworthy Machine Learning!

I'm really excited about this conference and hoping to see it take off. There's so much important work to do in this area.
#SaTML #AdversarialML

IEEE SaTML

IEEE Conference on Secure and Trustworthy Machine Learning