llm-moat

npm version TypeScript GitHub

Project Overview

llm-moat is an open-source security toolkit for production LLM systems. It is built to help developers detect prompt injection attempts, protect tool execution paths, and enforce safer trust boundaries around model input and output.

Why

LLM applications often connect models to tools, memory, retrieval systems, and sensitive business workflows. That creates a new security surface:

  • Prompt injection can override system instructions
  • Untrusted input can manipulate downstream tools
  • Tool chains can expose data or execute unintended actions
  • Teams need security controls that fit real-time, streaming applications

llm-moat exists to make those defenses practical in application code.

Key Features

  • Prompt injection detection for production-facing LLM workflows
  • Hybrid analysis using rule-based checks and semantic classification
  • Trust-boundary enforcement before tool execution
  • Input sanitization to reduce prompt poisoning and unsafe tool calls
  • Streaming-aware inspection for large or real-time model inputs
  • Adapter-friendly design for multiple model providers
  • Telemetry hooks for logging, monitoring, and incident review

Built With

  • TypeScript
  • OpenAI
  • Anthropic
  • Ollama

Installation

npm install llm-moat

Use Cases

  • Protecting AI copilots with tool access
  • Securing retrieval-augmented generation pipelines
  • Blocking unsafe prompt patterns before model execution
  • Reducing risk of data exfiltration and privilege escalation

Design Goals

  • Lightweight enough to integrate into existing app stacks
  • Flexible enough to support different providers and policies
  • Practical enough for real product teams, not only demos