Papers and resources related to the security and privacy of LLMs 🤖
-
Updated
Nov 27, 2024 - Python
Papers and resources related to the security and privacy of LLMs 🤖
The one-stop repository for large language model (LLM) unlearning. Supports TOFU, MUSE, WMDP, and many unlearning methods. All features: benchmarks, methods, evaluations, models etc. are easily extensible.
Python package for measuring memorization in LLMs.
The fastest Trust Layer for AI Agents
An Execution Isolation Architecture for LLM-Based Agentic Systems
It is a comprehensive resource hub compiling all LLM papers accepted at the International Conference on Learning Representations (ICLR) in 2024.
LLM security and privacy
LLM Platform Security: Applying a Systematic Evaluation Framework to OpenAI's ChatGPT Plugins
Make Zettelkasten-style note-taking the foundation of interactions with Large Language Models (LLMs).
Example of running last_layer with FastAPI on vercel
Add a description, image, and links to the llm-privacy topic page so that developers can more easily learn about it.
To associate your repository with the llm-privacy topic, visit your repo's landing page and select "manage topics."