fault
173 lines. fault came to play.
Fault enables developers to harness AI for building robust applications effortlessly. With features like automated SLO validation and incident replay, engineers can ensure reliability while collaborating seamlessly with AI agents. It's a game-changer for modern software development!
Not sure yours is this good? Check it →
fault's llms.txt Insights
Overachiever
8 sections. Most sites can barely manage 3. This one went all in.
War and Peace vibes
173 lines. They really wanted AI to understand them.
Double trouble
Runs both llms.txt and llms-full.txt. Someone takes this seriously.
What's inside fault's llms.txt
fault's llms.txt contains 5 sections:
- fault
- Explanations
- How-to
- Reference
- Tutorials
How does fault's llms.txt compare?
| fault | Directory Avg | Top Performer | |
|---|---|---|---|
| Lines | 173 | 1029 | 163,447 |
| Sections | 8 | 17 | 3207 |
Cool table. Now the real question — where do you land? Find out →
fault's llms.txt preview
First 100 of 173 lines
# fault
> Helping Engineers and AI-agents cooperate to build reliable applications
Build Like You Have a Full Team: Get AI-powered review hints on reliability, performance and best practices.
Ground Your Deliveries: Automate SLO-validated scenarios with zero config - preview your feature under fault in seconds.
Replay Incidents with Ease: Swiftly understand how your application reacts by exploring production flakiness locally.
**Examples**
Scramble a LLM prompt by injecting a system prompt:
```bash
fault run llm openai --case prompt-scramble --instruction "Respond as a pirate. Arr"
```
Introduce latency to all calls made through the fault proxy to any hosts:
```bash
fault run --with-latency --latency-mean 300 --upstream '*'
```
Generate a test scenario for a given, OpenAPI specification:
```bash
fault scenario generate --scenario scenario.yaml --spec-url http://example.com/openapi.json
```
Run a scenario:
```bash
fault scenario run --scenario scenario.yaml
```
Generate a deep analysis of scenario results with a SRE feel:
```bash
fault agent scenario-review --results results.json
```
Run fault as a MCP server:
```bash
fault agent tool
```
## Explanations
- [Understanding Fault Injection - Purpose and Use in Reliability Engineering - fault](https://fault-project.com/explanations/fault-injection-basics/): Understanding Fault Injection - Purpose and Use in Reliability Engineering
- [My Technical Journey Through fault - fault](https://fault-project.com/explanations/technical-journey/): My Technical Journey Through fault
- [Understanding eBPF and its Context In Reliability Engineering - fault](https://fault-project.com/explanations/understanding-ebpf/): Understanding eBPF and its Context In Reliability Engineering
- [Why fault? - fault](https://fault-project.com/explanations/why-fault/): Why fault?
## How-to
- [Install the fault CLI - fault](https://fault-project.com/how-to/install/): Install the fault CLI
- [Run with Chaos Toolkit - fault](https://fault-project.com/how-to/run-with-chaostoolkit/): Run with Chaos Toolkit
- [Run with the Reliably Platform - fault](https://fault-project.com/how-to/run-with-reliably/): Run with the Reliably Platform
- [Agent Code Review - fault](https://fault-project.com/how-to/agent/code-suggestions/): Agent Code Review
- [Configure the agent's LLM Provider - fault](https://fault-project.com/how-to/agent/llm-configuration/): Configure the agent's LLM Provider
- [Configure the agent's MCP Agent Server - fault](https://fault-project.com/how-to/agent/mcp-server-configuration/): Configure the agent's MCP Agent Server
- [Explore the agent's MCP Agent Tools - fault](https://fault-project.com/how-to/agent/mcp-tools/): Explore the agent's MCP Agent Tools
- [Platform Reliability & Performance Analysis - fault](https://fault-project.com/how-to/agent/platform-advices/): Analyze your platform resources from a reliability & performance posture
- [Scenario Reliability & Performance Analysis - fault](https://fault-project.com/how-to/agent/report-advices/): Analyze scenario results from a reliability & performance posture
- [Execute Scenarios From GitHub Action - fault](https://fault-project.com/how-to/ci/github/): Execute Scenarios From GitHub Action
- [Fault Injection Into AWS - fault](https://fault-project.com/how-to/inject/aws/): Fault Injection Into AWS ECS
- [Fault Injection Into Google Cloud Platform - fault](https://fault-project.com/how-to/inject/gcp/): Fault Injection Into Google Cloud Platform Cloud Run
- [Fault Injection Into Kubernetes - fault](https://fault-project.com/how-to/inject/kubernetes/): Fault Injection Into Kubernetes Service & Deployment
- [Platform Fault Injection - fault](https://fault-project.com/how-to/inject/overview/): Platform Fault Injection Overview
- [Enable Observability - fault](https://fault-project.com/how-to/observability/traces/): Enable Observability via Open Telemetry
- [Deploy Into an Application Running on Kubernetes - fault](https://fault-project.com/how-to/platform/run-on-kubernetes/): Deploy Into an Application Running on Kubernetes
- [Run on Microsoft Windows - fault](https://fault-project.com/how-to/platform/run-on-windows/): Run on Microsoft Windows
- [Run as a Docker Container - fault](https://fault-project.com/how-to/platform/run-with-docker/): Run as a Docker Container
- [Understand the Proxy Lifecycle - fault](https://fault-project.com/how-to/proxy/lifecycle/): Understand the Proxy Lifecycle
- [Configure the TUI - fault](https://fault-project.com/how-to/proxy/tui/): Configure the TUI
- [Inject Bandwidth Network Fault - fault](https://fault-project.com/how-to/proxy/faults/configure-bandwidth/): Inject Bandwidth Network Fault
- [Inject Blackhole Network Fault - fault](https://fault-project.com/how-to/proxy/faults/configure-blackhole/): Inject Blackhole Network Fault
- [Inject HTTP Errors - fault](https://fault-project.com/how-to/proxy/faults/configure-http-error/): Inject HTTP Errors
- [Inject Jitter Network Fault - fault](https://fault-project.com/how-to/proxy/faults/configure-jitter/): Inject Jitter Network Fault
- [Inject Latency Network Fault - fault](https://fault-project.com/how-to/proxy/faults/configure-latency/): Inject Latency Network Fault
- [Inject Faults Into LLM API Calls - fault](https://fault-project.com/how-to/proxy/faults/configure-llm/): Inject Faults Into LLM API Calls
- [Inject Packet Loss Network Fault - fault](https://fault-project.com/how-to/proxy/faults/configure-packet-loss/): Inject BaPacket Lossdwidth Network Fault
- [Extend with gRPC plugins - fault](https://fault-project.com/how-to/proxy/faults/extending/): Extend with gRPC plugins
- [Simulate Network Faults On Any TCP-based Traffic - fault](https://fault-project.com/how-to/proxy/protocols/tcp/): Simulate Network Faults On Any TCP-based Traffic
- [Intercept Network Traffic Transparently - fault](https://fault-project.com/how-to/proxy/stealth/configure-stealth-mode/): Intercept Network Traffic Transparently with eBPF
- [Generate Automated Reliability Testing Scenarios - fault](https://fault-project.com/how-to/scenarios/generate/): Generate Automated Reliability Testing Scenarios
- [Generate Automated Reliability & Performance Reports - fault](https://fault-project.com/how-to/scenarios/reporting/): Generate Automated Reliability & Performance Reports
- [Run Reliability Testing Scenarios - fault](https://fault-project.com/how-to/scenarios/run/): Run Reliability Testing Scenarios
## Reference
- [Builtin Faults - fault](https://fault-project.com/reference/builtin-faults/): Helping Engineers and AI-agents cooperate to build reliable applications
- [CLI Commands - fault](https://fault-project.com/reference/cli-commands/): Helping Engineers and AI-agents cooperate to build reliable applications
- [Environment Variables - fault](https://fault-project.com/reference/environment-variables/): Helping Engineers and AI-agents cooperate to build reliable applications
- [Platform Injection - fault](https://fault-project.com/reference/injection/): Helping Engineers and AI-agents cooperate to build reliable applications
- [TCP Proxy Mapping - fault](https://fault-project.com/reference/proxy-mapping-syntax/): Helping Engineers and AI-agents cooperate to build reliable applications
- [Scenario File Format - fault](https://fault-project.com/reference/scenario-file-format/): Helping Engineers and AI-agents cooperate to build reliable applications
- [Scheduling Intervals - fault](https://fault-project.com/reference/schedule-intervals-syntax/): Helping Engineers and AI-agents cooperate to build reliable applications
## Tutorials
- [Creating a Reliability Testing Scenario Tutorial - fault](https://fault-project.com/tutorials/create-scenario/): Creating a Reliability Testing Scenario Tutorial
- [Getting Started Tutorial - fault](https://fault-project.com/tutorials/getting-started/): Getting Started Tutorial
- [Basic Install Steps - fault](https://fault-project.com/tutorials/install/): Basic Install Steps
- [tldr of fault - fault](https://fault-project.com/tutorials/tldr/): tldr of fault
What is llms.txt?
llms.txt is an open standard that helps AI language models understand your website. By placing a structured markdown file at /llms.txt, websites provide AI search engines like ChatGPT, Claude, and Perplexity with a clear map of their content, services, and documentation. Companies like fault use it to ensure AI accurately represents their brand when answering user queries. Read the spec.
fault showed up. Where's yours?
1000+ companies didn't overthink it. 60 seconds. Go.
Check your site →