We've just released a tool that fixes a particularly annoying problem for those trying to fuzz HTTP/3.
The issue is that QUIC is designed to prevent network bottlenecks (HOL blocking), which is beneficial, but it disrupts the fundamental timing required for exploiting application-level race conditions. We tried all the obvious solutions, but QUIC's RFC essentially blocks fragmentation and other low-level network optimizations. π€·ββοΈ
So, we figured out a way to synchronize things at the QUIC stream layer using a technique we call Quic-Fin-Sync.
The gist:
This one packet forces the server to "release" all the requests into processing near-simultaneously. It worked way better than existing methods in our testsβwe successfully raced a vulnerable Keycloak setup over 40 times.
If you are pentesting HTTP/3, grab the open-source tool and let us know what you break with it. The full write-up is below.
Whatβs the most frustrating thing youβve run into trying to test QUIC/HTTP/3?
Disclosure: I work at CyberArk
The research shows that Chromeβs AppBound cookie encryption relies on a key derivation process with limited entropy and predictable inputs. By systematically generating possible keys based on known parameters, an attacker can brute-force the correct encryption key without any elevated privileges or code execution. Once recovered, this key can decrypt any AppBound-protected cookies, completely undermining the isolation AppBound was intended to provide in enterprise environments.
We broke Chromeβs AppBound cookie encryption used in enterprise environments. Using timing-based attacks against the WebView lifecycle, we were able to recover encryption keys and decrypt secure cookies β no root or sandbox escape required. This affects managed Android profiles using MDM and AppBound mode.
[Disclosure: I work at CyberArk and was involved in this research]
We've completed a security evaluation of the Model Context Protocol and discovered several concerning attack patterns relevant to ML practitioners integrating external tools with LLMs.
Background: MCP standardizes how AI applications access external resources - essentially creating a plugin ecosystem for LLMs. While this enables powerful agentic behaviors, it introduces novel security considerations.
Technical Findings:
ML-Specific Implications: For researchers using tools like Claude Desktop or Cursor with MCP servers, these vulnerabilities could lead to:
Best Practices:
This highlights the importance of security-by-design as we build more sophisticated AI systems.