Why Tree-Sitter Is Inadequate for Program Analysis

A super simple tool tells you who called the function and who implemented it

who-ast는 함수 호출자와 구현자를 알려주는 간단한 코드 분석 도구입니다. whocall은 특정 함수를 호출하는 코드를, whoimpl은 특정 트레이트를 구현한 코드를 찾아줍니다. Rust, Python, TypeScript, Go, Java 등 다양한 언어 지원을 계획 중이며, CLI 도구로 npm과 Homebrew를 통해 설치할 수 있습니다. AI 에이전트와 인간 개발자 모두에게 유용한 코드 탐색 기능을 제공합니다.

https://github.com/meloalright/who-ast

#codeanalysis #cli #rust #opensource #developertools

GitHub - meloalright/who-ast: ✨ A super simple code analysis tool for both humans and AI agents that tells you who called the function and who implemented it.

✨ A super simple code analysis tool for both humans and AI agents that tells you who called the function and who implemented it. - meloalright/who-ast

GitHub

Praveen Koka (@praveenkoka)

HTML 설명 수준은 이제 기본 역량이며, Claude가 난독화된 익스플로잇 코드를 파싱할 수 있다는 점이 실제 능력의 전환점이라고 언급한다. 보안 코드 분석과 악성 코드 이해에서 모델의 고급 추론 능력이 중요해졌음을 시사한다.

https://x.com/praveenkoka/status/2052995504372367839

#claude #cybersecurity #codeanalysis #llm #ai

Praveen Koka (@praveenkoka) on X

@simonw Explaining HTML is baseline now. Claude parsing obfuscated exploit code is the real capability inflection.

X (formerly Twitter)

Debt Behind the AI Boom: A Large-Scale Study of AI-Generated Code in the Wild

이 논문은 AI 코딩 어시스턴트가 실제 소프트웨어 개발 현장에서 생성한 코드가 장기적으로 기술 부채를 유발하는지를 대규모로 분석했다. 6,299개 GitHub 저장소에서 30만 건 이상의 AI 생성 커밋을 추적해 코드 냄새, 정확성 문제, 보안 이슈 등 48만 건 이상의 문제를 발견했으며, 이 중 22.7%는 최신 버전까지도 해결되지 않고 남아있음을 확인했다. AI 생성 코드는 생산성 향상에 기여하지만, 품질 보증과 유지보수 비용 증가라는 과제도 함께 존재함을 시사한다.

https://arxiv.org/abs/2603.28592

#aigeneratedcode #technicaldebt #softwarequality #github #codeanalysis

Debt Behind the AI Boom: A Large-Scale Empirical Study of AI-Generated Code in the Wild

AI coding assistants are now widely used in software development. Software developers increasingly integrate AI-generated code into their codebases to improve productivity. Prior studies have shown that AI-generated code may contain code quality issues under controlled settings. However, we still know little about the real-world impact of AI-generated code on software quality and maintenance after it is introduced into production repositories. In other words, it remains unclear whether such issues are quickly fixed or persist and accumulate over time as technical debt. In this paper, we conduct a large-scale empirical study on the technical debt introduced by AI coding assistants in the wild. To achieve that, we built a dataset of 302.6k verified AI-authored commits from 6,299 GitHub repositories, covering five widely used AI coding assistants. For each commit, we run static analysis before and after the change to precisely attribute which code smells, correctness issues, and security issues the AI introduced. We then track each introduced issue from the introducing commit to the latest repository revision to study its lifecycle. Our results show that we identified 484,366 distinct issues, and that code smells are by far the most common type, accounting for 89.3% of all issues. We also find that more than 15% of commits from every AI coding assistant introduce at least one issue, although the rates vary across tools. More importantly, 22.7% of tracked AI-introduced issues still survive at the latest version of the repository. These findings show that AI-generated code can introduce long-term maintenance costs into real software projects and highlight the need for stronger quality assurance in AI-assisted development.

arXiv.org

Dante (@thedntx)

코드 자동 정리만 할 줄 알았는데, 먼저 코드의 이상한 부분을 설명해 주는 방식이 더 똑똑하다고 평가했다. 개발자 도구나 AI 코딩 보조 기능이 단순 리포맷보다 설명 중심으로 진화하고 있음을 시사한다.

https://x.com/thedntx/status/2052866464906985623

#codingassistant #aidev #codeanalysis #developerexperience #llm

Dante (@thedntx) on X

@simonw i expected it to just reformat the code, but explaining the weird bits first is actually way smarter.

X (formerly Twitter)

Anthropic stellt Claude Security als Public Beta für Enterprise-Kunden vor.

Das auf Opus 4.7 basierende Modell führt dateiübergreifende Code-Analysen durch und verfolgt Datenflüsse direkt im Quellcode, ohne klassische Signaturen. Generierte Patches enthalten Konfidenzwerte und Schweregrade. Ergebnisse lassen sich per Webhook an Jira übermitteln oder als CSV/Markdown exportieren.

#ClaudeSecurity #Anthropic #DevSecOps #CodeAnalysis #AIGeneratedImage

https://www.all-ai.de/news/news26/claude-security-public-beta

Claude Security mit Verbesserungen in der Public Beta

Mit der Integration von Opus 4.7 erhalten Entwickler tiefgehende Fehleranalysen samt Schweregrad und Konfidenzwert.

All-AI.de

田中義弘 | taziku CEO / AI × Creative (@taziku_co)

Claude Security는 기존 스캐너가 놓친 프로덕션 코드의 취약점을 탐지한다. 단순 탐지가 아니라 각 결과를 검증해 오탐을 줄이고, 검토 가능한 패치까지 생성해 보안팀의 1차 대응 도구로 진화하는 흐름을 보여준다.

https://x.com/taziku_co/status/2049965961592312108

#claude #security #vulnerability #codeanalysis #cybersecurity

田中義弘 | taziku CEO / AI × Creative (@taziku_co) on X

既存スキャナが見逃した脆弱性を、Claude Securityが本番コードから検出。 重要なのは検出だけではない。 各findingを検証し、誤検知を削り、レビュー可能なパッチまで出す。 AIがセキュリティチームの一次対応者になる未来、かなり近い。 詳細は🧵

X (formerly Twitter)
Release 1.1.0 · rife2/bld-spotbugs

Summary Refactor parsing logic Add nullability contracts What's Changed in 1.1.0 Refactor to extract parsing and enforce nullability contracts in 9721edd Extract XML/SARIF parsing from SpotBugs...

GitHub
🎮📚"Tempest vs. Tempest" is a riveting saga where someone decided to write an entire book dissecting 40-year-old and 30-year-old code like it’s the Rosetta Stone of gaming. 🧠💤 Because, clearly, nothing screams "must-read" like an in-depth analysis of archaic assembly language. 🙄
https://tempest.homemade.systems #TempestVsTempest #GamingHistory #CodeAnalysis #RetroGaming #AssemblyLanguage #HackerNews #ngated
Tempest vs Tempest

The Making and Remaking of Atari's Iconic Video Game

Tempest vs Tempest