101010.pl is one of the many independent Mastodon servers you can use to participate in the fediverse.
101010.pl czyli najstarszy polski serwer Mastodon. Posiadamy wpisy do 2048 znaków.

Server stats:

485
active users

#aisecurity

0 posts0 participants0 posts today

Hello World! #introduction

Work in cybersec for 25+ years. Big OSS proponent.

Latest projects:

VectorSmuggle is acomprehensive proof-of-concept demonstrating vector-based data exfiltration techniques in AI/ML environments. This project illustrates potential risks in RAG systems and provides tools and concepts for defensive analysis.
github.com/jaschadub/VectorSmu

SchemaPin protocol for cryptographically signing and verifying AI agent tool schemas to prevent supply-chain attacks (aka MCP Rug Pulls).
github.com/ThirdKeyAI/SchemaPin

GitHubGitHub - jaschadub/VectorSmuggle: Testing platform for covert data exfiltration techniques where sensitive documents are embedded into vector representations and tunneled out under the guise of legitimate RAG operations — bypassing traditional security controls and evading detection through semantic obfuscation.Testing platform for covert data exfiltration techniques where sensitive documents are embedded into vector representations and tunneled out under the guise of legitimate RAG operations — bypassing...

AI-powered features are the new attack surface! Check out our new blog in which LMG Security’s Senior Penetration Tester Emily Gosney @baybedoll shares real-world strategies for testing AI-driven web apps against the latest prompt injection threats.

From content smuggling to prompt splitting, attackers are using natural language to manipulate AI systems. Learn the top techniques—and why your web app pen test must include prompt injection testing to defend against today’s AI-driven threats.

Read now: lmgsecurity.com/are-your-ai-ba

LMG SecurityAre Your AI-Backed Web Apps Secure? Why Prompt Injection Testing Belongs in Every Web App Pen Test | LMG SecurityDiscover how prompt injection testing reveals hidden vulnerabilities in AI-enabled web apps. Learn real-world attack examples, risks, and why your pen test must include LLM-specific assessments.

Microsoft Copilot for SharePoint just made recon a whole lot easier. 🚨
 
One of our Red Teamers came across a massive SharePoint, too much to explore manually. So, with some careful prompting, they asked Copilot to do the heavy lifting...
 
It opened the door to credentials, internal docs, and more.
 
All without triggering access logs or alerts.
 
Copilot is being rolled out across Microsoft 365 environments, often without teams realising Default Agents are already active.
 
That’s a problem.
 
Jack, our Head of Red Team, breaks it down in our latest blog post, including what you can do to prevent it from happening in your environment.
 
📌Read it here: pentestpartners.com/security-b

Replied in thread

@LukaszOlejnik

The assertion is, that one can not guarantee safety of #AI #LLM Models at all.

I'd like to get to the bottom of this, not that I doubt it by #AIsecurity is going to be increasingly more important, especially now that Old school #infosec has proven to be susceptible to the "Walk in and Seize control" exploit 😉

Edit;

Sometimes my friends tell me; Wulfy; /no_think

😁

Man, this whole AI hype train... Yeah, sure, the tools are definitely getting sharper and faster, no doubt about it. But an AI pulling off a *real* pentest? Seriously doubt that's happening anytime soon. Let's be real: automated scans are useful, but they just aren't the same beast as a genuine penetration test.

Honestly, I think security needs to be woven right into the fabric of a company from the get-go. It can't just be an afterthought you tack on when alarms are already blaring.

Now, don't get me wrong, AI definitely brings its own set of dangers – disinformation is a big one that springs to mind. But here's the thing: we absolutely *have* to get our heads around these tools and figure them out. If we don't keep pace, we risk becoming irrelevant pretty quick.

So, curious to hear what you all think – where do the greatest pitfalls lie with AI in the security field? What keeps you up at night?

Continued thread

◆ Hallucination and factual accuracy
◆ Bias and fairness
◆ Resistance to adversarial attacks
◆ Harmful content prevention

The LLM Benchmark incorporates diverse linguistic and cultural contexts to ensure comprehensiveness, and representative samples will be open-source.

Read about our methodology, and early findings: gisk.ar/3CRFdeB

We will be sharing more results in the coming months 👀

gisk.arGiskard announces a new LLM Evaluation Benchmark during the Paris AI SummitGiskard partners with Google DeepMind to launch an independent multilingual LLM benchmark, evaluating hallucinations and AI security risks.

Open Source AI Models are a growing cybersecurity risk.

Organizations are increasingly using AI models from repositories like Hugging Face and TensorFlow Hub—but are they considering the hidden cybersecurity risks? Attackers are slipping malicious code into AI models, bypassing security checks, and exploiting vulnerabilities.

New research shows that bad actors are leveraging open-source AI models to introduce backdoors, execute arbitrary code, and even manipulate model outputs. If your team is developing AI solutions, now is the time to secure your AI supply chain by:

🔹 Vetting model sources rigorously
🔹 Avoiding vulnerable data formats like Pickle
🔹 Using safer alternatives like Safetensors
🔹 Managing AI models like any other open-source dependency

As AI adoption skyrockets, you must proactively safeguard your models against supply chain threats. Check out the full article to learn more: darkreading.com/cyber-risk/ope

www.darkreading.comOpen Source AI Models: Big Risks for Malicious Code, VulnsCompanies pursing internal AI development using models from Hugging Face and other repositories need to focus on supply chain security and checking for vulnerabilities.

Ready to Secure AI Systems? Join Our 3-Day Hands-On Training at OWASP Global AppSec EU 2025!

Dive into AI/ML Whiteboard Hacking with expert Sebastien Deleersnyder from May 26-28, 2025 in Barcelona.

Designed for AI engineers, developers, architects, and security professionals, this intermediate-level training will equip you with practical skills to identify AI-specific threats.

owasp.glueup.com/event/123983/

I am reading up on abliterations:
huggingface.co/blog/mlabonne/a

Still trying to wrap my head around the consequences of this. But...

...I kinda feel like abliterations have implications also for prompt injections?

As in, it feels like abliterations could mean that it is simply impossible to secure an LLM from prompt injection?

I'm sure I am misunderstanding stuff here. Anyone any input on this?

huggingface.coUncensor any LLM with abliterationA Blog post by Maxime Labonne on Hugging Face