OPEN SOURCE SECURITY FOR LLM INFERENCE

> Self-hosted. Provider-agnostic. Free forever.

✓ OPEN SOURCE ✓ SELF-HOSTED ✓ ZERO-TRUST

// Launch: March 2026 | MIT Licensed

🚀

v0.1 MVP Release

Current: API key-based integrations (OpenAI, Anthropic, Google)

Coming in v0.2: OAuth device flows for IDE integrations (GitHub Copilot, Cursor, Windsurf) — Track progress on GitHub →

infershield-proxy@localhost:8000
[17:42:03] INFO Proxy started on port 8000
[17:42:05] INFO Connected to OpenAI API
[17:42:08] PASS Request ID: req_abc123 | Risk: 0 | Latency: 42ms
[17:42:12] DETECT Prompt injection attempt detected
└─> Pattern: "Ignore previous instructions"
└─> Risk Score: 85/100
└─> Action: BLOCKED
[17:42:15] BLOCK Request ID: req_def456 | Risk: 85 | Threat: PROMPT_INJECTION
[17:42:18] PASS Request ID: req_ghi789 | Risk: 5 | Latency: 38ms

// THE LLM SECURITY GAP

GARTNER PREDICTION
60%
of AI enterprises will face security incidents by 2027
ATTACK SURFACE
10x
deployment speed vs. security implementation
COMPLIANCE GAP
73%
of CISOs can't answer basic AI audit questions
[THREAT_001]

Prompt Injection

Attackers manipulate LLM behavior through crafted inputs. Your WAF won't catch it.

[THREAT_002]

Data Exfiltration

Sensitive credentials leak through LLM responses. No audit trail exists.

[THREAT_003]

Jailbreak Attempts

Users bypass safety guardrails. Your SIEM has no visibility.

[THREAT_004]

Compliance Violations

Auditors demand proof. You have no logs, no controls, no answers.

// HOW INFERSHIELD WORKS

YOUR APP
app.py / server.js
INFERSHIELD PROXY
localhost:8000
✓ Threat Detection
✓ Policy Enforcement
✓ Audit Logging
ANY LLM
OpenAI | Anthropic | Google
quick-start.sh
# Install InferShield (< 60 seconds)
$ docker pull infershield/proxy:latest
$ docker run -p 8000:8000 \\
-e OPENAI_API_KEY=sk-xxx \\
infershield/proxy
 
# Update your code (one line change)
client = OpenAI(base_url="https://api.openai.com/v1")
client = OpenAI(base_url="http://localhost:8000/v1")
 
✓ Done. Every request now protected.

🛡️ Real-Time Protection

  • Prompt injection detection
  • Data exfiltration blocking
  • Jailbreak prevention
  • Custom policy rules

📊 Complete Visibility

  • Every request logged
  • Risk scores calculated
  • Threat patterns tracked
  • Compliance-ready exports

🚀 Zero Friction

  • Self-hosted (your infra)
  • Provider-agnostic
  • < 1ms latency overhead
  • No code changes needed

🔓 Truly Open

  • MIT licensed forever
  • Full source transparency
  • No telemetry/tracking
  • Community-driven
enterprise.md

INFERSHIELD ENTERPRISE

Advanced security for regulated industries

// Advanced Detection
  • ML-based threat analysis
  • Custom detection rules
  • Zero-day protection
  • Behavioral anomaly detection
// Compliance
  • SOC 2 / HIPAA / GDPR reports
  • Audit log retention
  • Compliance dashboard
  • Policy templates (finance/health)
// Integrations
  • SSO / SAML / RBAC
  • Slack / PagerDuty alerts
  • Splunk / DataDog export
  • Custom webhooks
// Support
  • 24/7 security hotline
  • Dedicated Slack channel
  • SLA guarantees
  • Managed hosting option
join-waitlist

SECURE YOUR LLM INFRASTRUCTURE

Early access opens March 2026. Lock in your spot.

// No spam. Just launch updates and early access.