Skip to content
Merged
Changes from all commits
Commits
File filter

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
29 changes: 28 additions & 1 deletion docs/security-privacy-considerations.md
Original file line number Diff line number Diff line change
Expand Up @@ -37,9 +37,14 @@ These capabilities enable powerful user experiences but also create new risks th
- [1. Prompt Injection Attacks](#1-prompt-injection-attacks)
- [Metadata / Description Attacks (Tool Poisoning)](#1-metadata--description-attacks-tool-poisoning)
- [Output Injection Attacks](#2-output-injection-attacks)
- [Tool Implementation as Attack Targets](#3-tool-implementation-as-attack-targets)
- [2. Misrepresentation of Intent](#2-misrepresentation-of-intent)
- [3. Privacy Leakage Through Over-Parameterization](#3-privacy-leakage-through-over-parameterization)
- [Open Questions for Discussion](#open-questions-for-discussion)
- [Mitigations](#mitigations)
- [Proposed Mitigations](#proposed-mitigations)
- [Consider restricting maximum input lengths](#consider-restricting-maximum-input-lengths)
- [Supporting interoperable probabilistic defense structures through shared attack eval datasets](#supporting-interoperable-probabilistic-defense-structures-through-shared-attack-eval-datasets)
- [Next Steps](#next-steps)
- [References](#references)

Expand Down Expand Up @@ -335,6 +340,28 @@ To advance the security and privacy posture of WebMCP, we need community input o
- Should some tool categories require elevated permissions or review processes?
- Related: [Issue #44 - Action-specific permission](https://github.com/webmachinelearning/webmcp/issues/44)

## Mitigations

### Proposed Mitigations

#### [Consider restricting maximum input lengths](https://github.com/webmachinelearning/webmcp/issues/73)

**What:** Restrict the maximum amount of characters

**Threats addressed:** Metadata / Description Attacks (Tool Poisoning)

**How:** This restriction would not fully solve prompt injection attacks but helps shrink the possible universe of attacks, preventing longer prompts that leverage e.g. repetition and [sockpuppeting](https://arxiv.org/pdf/2601.13359) to convince agents of malicious tasks.

#### [Supporting interoperable probabilistic defense structures through shared attack eval datasets](https://github.com/webmachinelearning/webmcp/issues/106)

**What:** Shared evals for prompt injection attacks against WebMCP

**Threats addressed:** Prompt Injection Attacks (potentially Privacy Leakage Through Over-Parameterization)

**How:** Ensuring an interoperable basis for prompt injection defense, by requiring any implementer to protect against at least the attacks in that dataset

... add more issues here

## Next Steps

This document is intended to spark discussion and collaboration on WebMCP security and privacy considerations. We invite the community to:
Expand All @@ -352,4 +379,4 @@ This document is intended to spark discussion and collaboration on WebMCP securi

## Acknowledgment

This document was initially drafted based on discussion points from [Victor Huang](https://github.com/victorhuangwq), [Khushal Sagar](https://github.com/khushalsagar), [Johann Hofmann](https://github.com/johannhof), [Emily Lauber](https://github.com/EmLauber), [Dave Risney](https://github.com/david-risney), [Luis Flores](https://github.com/lflores-ms), and [Andrew Nolan](https://github.com/annolanmsft).
This document was initially drafted based on discussion points from [Victor Huang](https://github.com/victorhuangwq), [Khushal Sagar](https://github.com/khushalsagar), [Johann Hofmann](https://github.com/johannhof), [Emily Lauber](https://github.com/EmLauber), [Dave Risney](https://github.com/david-risney), [Luis Flores](https://github.com/lflores-ms), and [Andrew Nolan](https://github.com/annolanmsft).