AISecHub
2.09K subscribers
714 photos
38 videos
286 files
1.9K links
https://linktr.ee/aisechub managed by AISecHub. Sponsored by: innovguard.com
Download Telegram
redteam-indirect-web-pwn - Indirect Prompt Injection in Web-Browsing Agents https://www.promptfoo.dev/blog/indirect-prompt-injection-web-agents/

AI agents that can browse the web are increasingly common. Tools like web_fetch, MCP browser servers, and built-in browsing capabilities let agents pull in external content, summarize pages, and take action on what they find.

This is also one of the easiest ways to attack them.

An attacker doesn't need access to your system. They just need to put malicious instructions on a web page that your agent will visit. If the agent follows those instructions, you have a problem.

We built a test harness, which we call indirect-web-pwn, to test exactly this.
Agentic AI Risk-Management Standards Profile

A new paper authored by researchers from the Center for Long-Term Cybersecurity’s Artificial Intelligence Security Initiative (AISI) focuses on “AI agents” or “agentic AI,” AI systems that can autonomously pursue goals and take actions with little to no human oversight, often through interaction with external environments and tools.

https://cltc.berkeley.edu/2026/02/11/new-cltc-report-on-managing-risks-of-agentic-ai/
Trail of Bits Claude Code Config - https://github.com/trailofbits/claude-code-config

Opinionated defaults, documentation, and workflows for Claude Code at Trail of Bits. Covers sandboxing, permissions, hooks, skills, MCP servers, and usage patterns we've found effective across security audits, development, and research.
The Definitive Guide to AgentSecOps -https://info.straiker.ai/hubfs/Ebooks-Whitepapers/Straiker-AgenstSecOps-Ebook-2026.pdf

Traditional CI/CD pipelines test for deterministic security flaws like SQL injection and dependency vulnerabilities, but AI agents and applications introduce a different challenge. They reason, make decisions, and take actions autonomously. Risks such as prompt injection, tool misuse and context leakage emerge during execution, not in code. These behavioral vulnerabilities require a different approach to security.

AgentSecOps, which extends DevSecOps to test cognitive behavior alongside code. At its foundation is Autonomous Attack Simulation (AAS), where adversarial agents probe target agents in controlled environments. It integrates into your existing pipeline as a new test stage, similar to how fuzz testing works for code paths.
CVE-2026-25253: How Malicious Links Can Steal Authentication Tokens and Compromise OpenClaw AI Systems
https://hackers-arise.com/cve-2026-25253-how-malicious-links-can-steal-authentication-tokens-and-compromise-openclaw-ai-systems/
Anthropic just published a “Sabotage Risk Report” on Claude Opus 4.6. - https://anthropic.com/claude-opus-4-6-risk-report
“Google says attackers used 100,000+ prompts to try to clone AI chatbot Gemini
“ - https://cloud.google.com/blog/topics/threat-intelligence/distillation-experimentation-integration-ai-adversarial-use
secureclaw - Automated security hardening for OpenClaw AI agents - https://github.com/adversa-ai/secureclaw

51 audit checks. 12 behavioral rules. 9 scripts. 4 pattern databases. Full OWASP ASI Top 10 coverage.

SecureClaw audits your OpenClaw installation for misconfigurations and known vulnerabilities, applies automated hardening fixes, and gives your agent behavioral security rules that protect against prompt injection, credential theft, supply chain attacks, and privacy leaks.

What Problem Does SecureClaw Solve?

AI agents with access to your files, credentials, email, and the internet are a fundamentally different security surface than traditional software. An agent that can read your .env file and send HTTP requests can exfiltrate your API keys in a single tool call. An agent that trusts instructions embedded in a web page or email can be hijacked to act against your interests.

SecureClaw addresses this by operating on three layers:

Layer 1 - Audit. 51 automated checks across 8 categories scan your OpenClaw installation for known misconfigurations: exposed gateway ports, weak file permissions, missing authentication, plaintext credentials outside .env, disabled sandboxing, and more.

Layer 2 -Hardening. Automated fixes for the most critical findings: binding the gateway to localhost, locking down file permissions, adding privacy and injection-awareness directives to your agent's core identity file, and creating cryptographic baselines for tamper detection.

Layer 3 - Behavioral rules. 12 rules loaded into your agent's context that govern how it handles external content, credentials, destructive commands, privacy, and inter-agent communication. These rules cost approximately 1,150 tokens of context window and provide defense against prompt injection, data exfiltration, and social engineering -- attacks that cannot be prevented by infrastructure configuration alone.
🔥4
AI Security Guide and Risk Assessment Tool - https://www.rand.org/pubs/tools/TLA4174-1/ai-security/guide.html by RAND

This guide is a practical, risk-based resource for developers, security experts, and policy professionals navigating the AI security landscape.⁠1 The guide addresses security of AI systems broadly, including machine learning (ML) models and other AI-enabled architectures. Certain sections, such as the threat landscape and model weight protection sections, focus more specifically on statistical, ML-based models. Building on industry best practices and expert insights, the guide helps you understand and manage the security risks associated with AI systems across their lifecycle—from design and development to deployment and operation.