The challenge with AI in open source security has transitioned from an AI slop tsunami into more of a ... plain security report tsunami. Less slop but lots of reports. Many of them really good.

I'm spending hours per day on this now. It's intense.

This trend is seen elsewhere as well. Mentioned by Willy here: https://lwn.net/Articles/1065620/
Significant raise of reports [LWN.net]

@bagder

this is both fascinating and somewhat terrifying

"we want fewer slop reports"

monkey paw curls

@bagder So it is good? Kind of?
@feyter lots of high quality reports is the good part. The challenge is the load they generate.

@bagder Since I feel there isn't much I could do to help in this, I will at least use this opportunity to thank you for doing all this work.

Thank You 🙂

@bagder “Now most of these reports are correct, to the point that we had to bring in more maintainers to help us.”

Well that’s good to see. At least it’s not the 100% made up garbage anymore.

Rough situation. Maybe this is where 2 levels of people would help? First for general + find dupes, second for bigger/trickier, passed on from first?

IDK. Such a sea change. Going to be interesting to see how the world sorts all this out.

@bagder
At least it's spending time on stuff that's somewhat worth while?
But still, it's a DDOS on maintainers
@dirkhh it's actually really hard to complain when the reports are good, but yeah we're still humans who need to deal with all this
@bagder @dirkhh do those reports come also with a way to fix them? Are patches attached too? In what percentage of them?
@pemensik @dirkhh very few have any attempts at fixes
@bagder @dirkhh which is my opinion the most important problem with AI reports. If they can use AI to find issues, it should come with fixes as well. Ideally a test case attached for it as well. Then the burden on human maintainers would be only the review part, not everything else.
@pemensik @dirkhh the AIs are still better at finding problems than fixing them, in my experience
@bagder
Oh yes, by a wide margin...
They actually are also better at introducing subtle bugs than fixing them 🤷‍♂️
@pemensik
@bagder @dirkhh sure, that is expected. But flood of many reports need to be processed by maintainers. If they came together with a fix, is should be less work to the maintainer, in theory. If the proposal is decent enough. Cloning good maintainers is much harder than cloning AI analyzers. Increasing number of reports won't help without increasing number of entities fixing bugs. If you increased the first, please try also the latter.
@bagder wait a minute, that post comes a day too late? ;)
@bagder well at least it's not a waste of time anymore then.
@bagder Looking at the dashboard, Curl is at an all time low of Cyclomatic Complexity and has few open issues compared to previous years. Lines of docs and testcases per KLOC are at an all time high. I think now is a good time to work on security. If there are good reports, even better. I think curl might be in the best state it has ever been.
@bagder has the ratio of vulns to reports improved compared to the slop age?
@wolf480pl yes, I think so. We still get quite a few reports we conclude are not vulnerabilities, but there are very few of the strong slop kinds now.
@bagder here is a similar experience on lwn. https://lwn.net/Articles/1065620/ something is really shifting. Thanks for keeping us up to date with the ai and security developments.
Significant raise of reports [LWN.net]

@bagder
Perhaps it's because LLMs learn very quickly. And in that way they make fewer mistakes because they have fewer biases than humans.

@bagder any feeling what the biggest change is?

Just more advanced models, or people learning to use them as proper tools instead of just prompting "find me bugs"?

@skaverat the tools are getting better, no doubt
@bagder imagine if people joined and contributed to projects instead of inundating maintainers they dont compensate with requests to do more work for free please
@bagder I'm hearing similar things from both #OpenSSL and #GnuTLS.

@neverpanic @bagder

If that hat is any indication than both yours and mine (SUSE) colleagues (and me) can testify the same. Number of CVEs (some of them spurious but most not) is completely crazy.

@bagder any thoughts what that means for AI in software development in general, not necessarily only bug/security reports?