AI Plugins

AI plugins are specialized software components that enable AI systems to interface with external applications, services, and data sources. They function as an adapter layer — abstracting away each external system’s complexity so the AI can call a standardized interface rather than learning every system’s native API.

The Adapter Pattern

The fundamental architecture is:

  • AI model issues a high-level request (“search the web”, “query the database”, “send a message”)
  • Plugin translates that request into system-specific commands using the target system’s native protocols
  • External system executes and returns results, which the plugin relays back to the AI

This indirection keeps the AI’s core reasoning general-purpose while new capabilities are added without retraining the model.

Historical Origin

OpenAI introduced the plugin paradigm in March 2023 with ChatGPT Plugins — the first large-scale attempt to connect a language model to live external systems. This created a marketplace dynamic: developers published plugins (web browsing, code execution, shopping) that users could install on demand. The model couldn’t directly know which plugin to use; it was guided by plugin descriptions embedded in the context window.

From Plugins to Standardized Protocols

The early plugin ecosystem suffered from fragmentation — each provider implemented a proprietary interface. Anthropic’s Model-Context-Protocol (2024) addresses this by defining a standard protocol that any tool can implement, making plugins interoperable across models and clients. MCP servers are, in effect, plugins with a standardized interface. The distinction matters: plugins are the general concept; MCP is one protocol that implements them at scale.

Claude Plugins: The Productized Form

Anthropic’s Claude Plugins directory extends the concept further — plugins bundle tools, skills, and integrations into a single one-click installation. A Claude Plugin may combine MCP tools, AGENTS-md-Files behavior instructions, and pre-built agent skills into a cohesive capability extension.

Security Surface

Plugins significantly expand the AI’s attack surface. Indirect prompt injection — where malicious content in a retrieved document hijacks the AI’s tool-use behavior — is a critical risk. Plugins can be exploited to exfiltrate data, execute unintended actions, or bypass safety guardrails. Mitigations include sandboxing, minimal-permission scopes, and human-in-the-loop confirmation for destructive operations.

Where Plugins Sit in the Harness

Within Harness-Engineering, plugins are a key integration layer: they extend what the AI can perceive and affect without expanding the model itself. They work alongside Context-Engineering (plugins provide fresh, dynamic context) and AI-Coding-Agent-Architecture (agents require plugins to act on external systems).

Sources

  • OpenAI (2023). “ChatGPT plugins.” OpenAI Blog. March 23, 2023. Available: https://openai.com/index/chatgpt-plugins/

    • Original announcement defining the plugin paradigm for LLMs; introduced the plugin marketplace and context-guided tool selection model
  • Moveworks (2025). “AI Plugin.” Moveworks AI Terms Glossary. Available: https://www.moveworks.com/us/en/resources/ai-terms-glossary/ai-plugin

    • Defines plugins as adapter components; explains the translation mechanism between AI and external systems; covers automation and scalability benefits
  • Anthropic (2025). “Claude Plugins.” claude.com. Available: https://claude.com/plugins

    • Illustrates the productized plugin form: bundled tools, skills, and integrations for one-click installation; MCP-server architecture
  • Anthropic (2024). “Introducing the Model Context Protocol.” Anthropic News. November 2024. Available: https://www.anthropic.com/news/model-context-protocol

    • Positions MCP as the standardization layer addressing plugin ecosystem fragmentation; contextualizes the plugin-to-protocol evolution
  • Greshake, Kai, Sahar Abdelnabi, Shailesh Mishra, Christoph Endres, Thorsten Holz, and Mario Fritz (2023). “Not What You’ve Signed Up For: Compromising Real-World LLM-Integrated Applications with Indirect Prompt Injections.” arXiv preprint arXiv:2302.12173. Available: https://arxiv.org/abs/2302.12173

    • Seminal paper on indirect prompt injection attacks via LLM plugins; foundational security reference for plugin-enabled AI systems

Note

This content was drafted with assistance from AI tools for research, organization, and initial content generation. All final content has been reviewed, fact-checked, and edited by the author to ensure accuracy and alignment with the author’s intentions and perspective.