Skip to main content
AI Tool Directory

AI Infrastructure Tools

Curated directory of AI infrastructure tools reviewed by practitioners. LLM security, observability, orchestration, RAG, agents, and MLOps.

27 Tools Reviewed

In-depth technical analysis — not marketing copy.

All (27)LLM Security (6)Vector Databases (5)AI Observability (5)AI Orchestration (1)RAG Platforms (2)AI Agent Frameworks (2)AI Gateways (2)MLOps (4)
LLM SecurityFree tier available, Enterprise custom

Lakera Guard

Real-time LLM security and prompt injection defense

Lakera Guard provides real-time protection against prompt injection, data leakage, and harmful content in LLM applications. Deploys as a middleware layer between your application and the LLM provider.

Use Cases
  • Prompt injection detection
  • PII/data leakage prevention
  • Content moderation for LLM outputs
  • Compliance enforcement
REST APIPython SDKDockerKubernetes
LLM SecurityOpen-source

Rebuff

Self-hardening prompt injection detection

Rebuff is a prompt injection detection tool that uses multiple layers — heuristics, LLM-based analysis, and a vector database of known attacks — to protect LLM applications.

Use Cases
  • Multi-layer prompt injection defense
  • Attack pattern learning
  • API gateway integration
  • Red team testing
PythonVector DBLLM-based detection
LLM SecurityOpen-source, Enterprise plans

Guardrails AI

Input/output validation for LLM applications

Guardrails AI provides validators for LLM inputs and outputs — enforce structure, detect toxicity, check factuality, and ensure compliance in production LLM systems.

Use Cases
  • Output format validation (JSON, XML)
  • Toxicity and bias filtering
  • Factuality checking
  • Custom business rule enforcement
PythonValidator HubLLM agnostic
Architecture Guides
LLM Security🚀 StartupFlat predictable pricing, Foundation / Growth / Strategic tiers

SlashLLM

Integrated Service Provider for AI Security — platform, operations, and governance

SlashLLM is the ISP for AI Security — a fully integrated platform that sits between your applications and any LLM provider. Combines API gateway, guardrails, observability, red-teaming, and governance into one service with 24/7 AI-SOC monitoring and compliance evidence generation.

Use Cases
  • End-to-end LLM security with gateway + guardrails + observability
  • 24/7 AI-SOC monitoring for prompt injection and data exfiltration
  • Compliance automation (SOC 2, ISO 27001, HIPAA, GDPR, EU AI Act)
  • Automated red-teaming and jailbreak testing in CI/CD
DockerKubernetesMulti-model gatewayCI/CD integrationSIEM/IAM integration
LLM SecurityOpen-source tools, Enterprise platform

Protect AI

ML supply chain security and model scanning

Protect AI provides security scanning for ML models, pipelines, and supply chains. Detects vulnerabilities in model artifacts, serialized objects, and AI/ML dependencies before deployment.

Use Cases
  • ML model vulnerability scanning
  • Supply chain security for AI pipelines
  • Malicious model detection
  • CI/CD security gates for ML
PythonCLICI/CD pluginsModelScan
Architecture Guides
LLM SecurityEnterprise

Robust Intelligence

AI firewall and continuous model validation

Robust Intelligence provides an AI firewall that continuously validates model inputs and outputs in production. Detects adversarial attacks, data drift, and model degradation in real time.

Use Cases
  • Real-time AI firewall for production models
  • Adversarial attack detection
  • Model stress testing and red-teaming
  • Continuous validation and monitoring
Python SDKREST APIKubernetesCloud-native
Architecture Guides
Vector Databases⭐ FeaturedFree tier, Standard and Enterprise plans

Pinecone

Managed vector database for similarity search at scale

Pinecone is a fully managed vector database built for high-performance similarity search. Handles billions of vectors with low-latency queries, automatic scaling, and zero operational overhead.

Use Cases
  • Production RAG vector storage
  • Semantic search at scale
  • Recommendation engines
  • Anomaly detection with embeddings
PythonREST APIgRPCServerless / Pod-based
Architecture Guides
Vector Databases⭐ FeaturedOpen-source, Weaviate Cloud managed service

Weaviate

AI-native vector database with built-in vectorization

Weaviate is an open-source vector database with built-in vectorization modules, hybrid search (vector + keyword), and multi-tenancy support. Extensible via modules for different ML models.

Use Cases
  • Hybrid search applications
  • Multi-tenant RAG systems
  • Generative search with built-in LLM integration
  • Knowledge graph augmented retrieval
GoREST/GraphQL APIDockerKubernetesHelm
Architecture Guides
Vector Databases🔥 TrendingOpen-source (Apache 2.0), Qdrant Cloud managed

Qdrant

High-performance open-source vector search engine

Qdrant is a Rust-based vector search engine optimized for speed and efficiency. Features advanced filtering, payload indexing, and quantization for production-scale similarity search.

Use Cases
  • Low-latency vector search
  • Filtered similarity queries
  • Multi-vector and sparse vector support
  • Edge deployment with quantization
RustPython/JS/Go SDKsgRPCDockerKubernetes
Architecture Guides
Vector DatabasesOpen-source (Apache 2.0)

ChromaDB

Lightweight open-source embedding database

Chroma is a lightweight, developer-friendly embedding database designed for LLM applications. Runs in-memory or persistent mode with zero-config setup — ideal for prototyping and small-scale RAG.

Use Cases
  • Rapid RAG prototyping
  • Local development and testing
  • Small-scale embedding storage
  • Notebook-friendly vector search
PythonJavaScript/TypeScriptSQLite backendREST API
Vector DatabasesOpen-source, LanceDB Cloud (beta)

LanceDB

Serverless vector database built on Lance format

LanceDB is a serverless vector database built on the Lance columnar format. Supports multi-modal data (text, images, video), automatic versioning, and zero-copy integration with ML pipelines.

Use Cases
  • Serverless vector search
  • Multi-modal embedding storage
  • Data versioning for ML experiments
  • Cost-efficient large-scale storage
PythonRustLance formatS3/GCS compatible
AI Observability🔥 TrendingOpen-source (self-hosted free), Cloud plans available

Langfuse

Open-source LLM observability and analytics

Langfuse is an open-source observability platform for LLM applications. It provides tracing, evaluation, prompt management, and cost analytics for production AI systems.

Use Cases
  • LLM call tracing and debugging
  • Prompt versioning and A/B testing
  • Cost tracking per model/feature
  • Quality evaluation pipelines
TypeScriptPython SDKOpenTelemetryPostgreSQLSelf-hosted or Cloud
AI Observability⭐ FeaturedOpen-source

Arize Phoenix

AI observability for LLMs, embeddings, and RAG

Phoenix by Arize provides deep observability into LLM and ML systems including retrieval analysis for RAG, embedding drift detection, and trace-level debugging of AI pipelines.

Use Cases
  • RAG retrieval quality analysis
  • LLM trace visualization
  • Embedding drift monitoring
  • Hallucination detection
PythonOpenTelemetryJupyter integrationSelf-hosted
AI ObservabilityFree tier, Growth and Enterprise plans

WhyLabs

AI observability with data and model monitoring

WhyLabs provides AI observability focused on data quality monitoring, model performance tracking, and LLM security. Built on the open-source whylogs profiling library for lightweight data monitoring.

Use Cases
  • Data drift and quality monitoring
  • LLM guardrails and content safety
  • Model performance degradation alerts
  • Embedding and feature monitoring
PythonwhylogsREST APIIntegrations (Spark, Airflow)
AI Observability🔥 TrendingFree tier, Pro and Enterprise plans

Braintrust

End-to-end LLM evaluation and observability platform

Braintrust provides evaluation, logging, and prompt playground for LLM applications. Features CI-integrated eval scoring, dataset management, and real-time production tracing.

Use Cases
  • LLM evaluation with custom scoring
  • A/B testing for prompts and models
  • Production logging and tracing
  • Dataset curation for fine-tuning
PythonTypeScriptREST APICI/CD integration
Architecture Guides
AI ObservabilityEnterprise

Fiddler AI

Enterprise AI observability and model monitoring

Fiddler provides enterprise-grade AI observability with explainability, drift detection, fairness monitoring, and LLM analytics. Designed for regulated industries requiring model governance.

Use Cases
  • Model explainability and bias detection
  • Data drift monitoring at scale
  • LLM token and cost analytics
  • Regulatory compliance and audit trails
Python SDKREST APICloud-nativeSOC 2 compliant
AI OrchestrationOpen-source, LangSmith paid plans

LangChain

Framework for building LLM-powered applications

LangChain provides composable building blocks for LLM application development — chains, agents, retrieval, memory, and tool use. The most widely adopted orchestration framework in the LLM ecosystem.

Use Cases
  • Conversational AI with memory
  • RAG pipelines with vectorstores
  • Multi-step agent workflows
  • Tool-augmented LLM systems
PythonTypeScript/JSLangSmith (observability)LangGraph (agents)
RAG PlatformsOpen-source, deepset Cloud enterprise

Haystack

Production-ready framework for RAG and NLP pipelines

Haystack by deepset is a framework for building production-grade RAG, search, and NLP pipelines. It provides a pipeline-based architecture with built-in document processing, retrieval, and generation.

Use Cases
  • Document search and retrieval
  • Question answering systems
  • Semantic search pipelines
  • Multi-modal RAG
PythonPipeline APIElasticsearchOpenSearchWeaviate
Architecture Guides
Related Tools
RAG PlatformsOpen-source, LlamaCloud paid

LlamaIndex

Data framework for LLM applications and RAG

LlamaIndex provides the data infrastructure for LLM applications — data ingestion, indexing, retrieval, and query engines. Optimized for connecting LLMs with enterprise data sources.

Use Cases
  • Enterprise knowledge bases
  • Multi-source data ingestion
  • Structured + unstructured RAG
  • Query planning over complex data
PythonTypeScriptLlamaCloudVector stores
Architecture Guides
Related Tools
AI Agent FrameworksOpen-source

CrewAI

Framework for orchestrating multi-agent AI systems

CrewAI enables building teams of AI agents that collaborate to accomplish complex tasks. Agents have roles, goals, and backstories, and work together through defined processes.

Use Cases
  • Multi-agent research workflows
  • Automated content pipelines
  • Code review and analysis agents
  • Business process automation
PythonLLM agnosticTool integration API
Related Tools
AI Agent FrameworksOpen-source (MIT)

AutoGen

Multi-agent conversational AI framework by Microsoft

AutoGen enables building multi-agent systems where agents can converse with each other to solve tasks. Supports human-in-the-loop patterns and complex conversation flows.

Use Cases
  • Collaborative problem solving
  • Code generation and execution
  • Task decomposition with agent teams
  • Human-AI collaborative workflows
PythonMulti-model supportCode execution sandbox
Related Tools
AI GatewaysFree tier, Growth and Enterprise plans

Portkey

AI gateway with guardrails, caching, and observability

Portkey is a full-featured AI gateway that sits between your application and LLM providers. Provides unified API, semantic caching, guardrails, fallbacks, load balancing, and cost tracking across 25+ providers.

Use Cases
  • Multi-provider LLM routing and fallbacks
  • Semantic caching for cost reduction
  • Guardrails and content filtering
  • LLM spend analytics and budgeting
REST APIPython/JS SDKsOpenAI-compatibleCloud or self-hosted
Related Tools
AI GatewaysOpen-source (MIT), Enterprise support

LiteLLM

Lightweight open-source LLM proxy and gateway

LiteLLM is a lightweight proxy that provides a unified OpenAI-compatible interface to 100+ LLM providers. Features model fallbacks, spend tracking, rate limiting, and virtual API keys.

Use Cases
  • Unified API across LLM providers
  • Cost tracking and budget enforcement
  • Rate limiting and access control
  • Provider failover and load balancing
PythonDockerOpenAI-compatible APIPostgreSQL
Related Tools
MLOpsOpen-source (Apache 2.0)

MLflow

Open-source platform for ML lifecycle management

MLflow manages the full ML lifecycle — experiment tracking, model registry, deployment, and monitoring. Now with LLM tracking and evaluation features for the AI/LLM era.

Use Cases
  • Experiment tracking and comparison
  • Model versioning and registry
  • LLM evaluation and benchmarking
  • Model deployment and serving
PythonREST APISQL backendDockerKubernetes
MLOpsOpen-source (Apache 2.0)

Kubeflow

ML toolkit for Kubernetes

Kubeflow provides a portable, scalable ML platform on Kubernetes. Includes pipeline orchestration, notebook servers, model training, serving, and experiment tracking.

Use Cases
  • ML pipeline orchestration
  • Distributed model training
  • Model serving at scale
  • Jupyter notebook management on K8s
KubernetesPythonIstioKnativeTensorFlow/PyTorch
MLOps🔥 TrendingPay-per-token, volume discounts

Together AI

Inference and fine-tuning cloud for open-source models

Together AI provides a cloud platform for running, fine-tuning, and deploying open-source LLMs. Offers fast inference via custom hardware, serverless endpoints, and fine-tuning APIs.

Use Cases
  • Fast inference for open-source LLMs
  • Custom fine-tuning with your data
  • Serverless model endpoints
  • Batch processing for embeddings
REST APIPython SDKOpenAI-compatibleNVIDIA GPUs
MLOpsPay-as-you-go compute, Enterprise plans

Anyscale

Scalable AI compute platform built on Ray

Anyscale provides a managed platform for Ray — the distributed computing framework for ML and AI workloads. Simplifies scaling training, serving, and data processing across GPU clusters.

Use Cases
  • Distributed model training
  • Scalable model serving with Ray Serve
  • Data preprocessing at scale
  • Reinforcement learning workloads
RayPythonKubernetesAWS/GCPNVIDIA GPUs

Submit Your AI Infrastructure Tool

Building an AI infrastructure tool? Get listed in our directory. We review and feature tools used by DevOps engineers and AI platform teams.

Submissions are reviewed for technical quality, production readiness, and relevance to AI infrastructure. Accepted tools receive a directory listing, optional technical review, and cross-linking to architecture guides.

Building an AI Infrastructure Tool?

Get featured in our directory. We write technical reviews, architecture guides, and comparison content reaching thousands of engineers and enterprise teams.