Queen of PETs: Part 1

A Netflix story from the past that holds important lessons about privacy risks in today's world of AI and Big Data.
Welcome to Part 1 of Queen of PETs (Privacy Enhancing Technologies).








Search for a command to run...

A Netflix story from the past that holds important lessons about privacy risks in today's world of AI and Big Data.
Welcome to Part 1 of Queen of PETs (Privacy Enhancing Technologies).








Are you building your next LLM integration? Please consider this: Integrations for data retrievals can introduce vulnerabilities in your LLM, allowing attackers to inject malicious prompts. This type of vulnerability is known as Indirect Prompt Injec...

Terraform is a go-to for infrastructure as code (IaC), letting developers easily set up and manage infrastructure. But, as with any tool, it brings security challenges. Here are five key tips to keep your Terraform setups safe and solid 👇

Guard your LLM against prompt injection with these powerful tools: - https://github.com/protectai/llm-guard - https://github.com/protectai/rebuff - https://github.com/NVIDIA/NeMo-Guardrails - https://github.com/amoffat/HeimdaLLM - https://github.com/...

What could go wrong during the ML model deployment lifecycle (Part 2)? Continuing the example threat model from last time. It is based on the talk "Kubernetes MLSec: Securing AI in Space" by Francesco Beltramini and James Callaghan of ControlPlane. L...

What could go wrong during the ML model development lifecycle? Here is an example threat model based on the talk "Kubernetes MLSec: Securing AI in Space" by Francesco Beltramini and James Callaghan of ControlPlane. Link: [https://www.youtube.com/watc...
