# resq

[docs.resq.software](https://docs.resq.software)

- **Overall score:** 86/100 (Grade B)
- **Checks passed:** 17 / 29
- **Last computed:** 2026-05-11

## Components

### Content Discoverability

- **Score:** 80/100 · **Status:** fail
- **Summary:** 2 failed and 2 warnings across 6 AFDocs checks.
- **Rationale:** Agents need a clear entry point and crawl map before they can reliably discover the right pages.
- **Reference:** [AFDocs reference](https://afdocs.dev)

**Checks**

- ✅ **LLMS TXT Exists** — llms.txt found at 1 location(s)
- ⚠️ **LLMS TXT Valid** — llms.txt contains parseable links but doesn't fully follow the proposed structure: https://docs.resq.software/llms.txt: No blockquote summary found Your llms.txt contains parseable links but doesn't follow the standard structure. Add an H1 title as the first line and a blockquote summary (lines starting with >) to improve agent parsing.
- ⚠️ **LLMS TXT Size** — llms.txt is 97,275 characters (between 50,000 and 100,000; consider splitting) Your llms.txt is 97,275 characters, which may be truncated on some agent platforms. If it grows further, split into nested llms.txt files with a root index under 50,000 characters.
- ❌ **LLMS TXT Links Resolve** — Only 11/15 same-origin sampled links resolve (73%); 4 broken 4 of 20 links in your llms.txt return errors. A stale llms.txt with broken links is worse than no llms.txt at all because it sends agents down dead ends with high confidence.
- ❌ **LLMS TXT Links Markdown** — Same-origin links point to HTML and no markdown alternatives detected (5 external links excluded) Your llms.txt links point to HTML pages. Update them to .md URL variants so agents receive markdown instead of converted HTML.
- ✅ **LLMS TXT Directive** — llms.txt directive found in all 14 sampled pages, near the top of content; 1 failed to fetch

### Markdown Availability

- **Score:** 100/100 · **Status:** pass
- **Summary:** 2 AFDocs checks pass.
- **Rationale:** When markdown is available directly, agents spend less effort stripping presentation markup and guessing structure.
- **Reference:** [AFDocs reference](https://afdocs.dev)

**Checks**

- ✅ **Markdown Url Support** — 9/9 sampled pages support .md URLs (100%)
- ✅ **Content Negotiation** — 9/9 sampled pages support content negotiation (100%)

### Page Size and Truncation Risk

- **Score:** 74/100 · **Status:** fail
- **Summary:** 2 failed across 4 AFDocs checks.
- **Rationale:** Large pages and delayed primary content increase truncation risk and make retrieval less reliable.
- **Reference:** [AFDocs reference](https://afdocs.dev)

**Checks**

- ✅ **Rendering Strategy** — All 15 sampled pages contain server-rendered content
- ✅ **Page Size Markdown** — All 9 pages under 50K chars (median 2K, max 6K)
- ❌ **Page Size Html** — 15 of 15 sampled pages convert to over 100K chars (max 685K, 8% boilerplate) 15 of 15 pages convert to over 100K characters of markdown. Reduce inline CSS/JS, break large pages, or provide markdown versions as a smaller alternative.
- ❌ **Content Start Position** — 1 of 15 sampled pages have content starting past 50% (worst 100%) 1 of 15 pages have content starting past 50% of the converted output. Agents may never see the documentation content. Move or remove inline CSS/JS that precedes the content area.

### Content Structure

- **Score:** 100/100 · **Status:** pass
- **Summary:** 3 AFDocs checks pass.
- **Rationale:** Predictable sections, valid code fences, and serialized tabs make the content easier for agents to parse correctly.
- **Reference:** [AFDocs reference](https://afdocs.dev)

**Checks**

- ✅ **Tabbed Content Serialization** — No tabbed content detected across 15 sampled pages
- ✅ **Section Header Quality** — No tabbed content found; header quality check not applicable
- ✅ **Markdown Code Fence Validity** — All 17 code fences properly closed across 10 pages

### URL Stability and Redirects

- **Score:** 100/100 · **Status:** pass
- **Summary:** 2 AFDocs checks pass.
- **Rationale:** Stable URLs and sane redirect behavior prevent retrieval drift and broken tool references.
- **Reference:** [AFDocs reference](https://afdocs.dev)

**Checks**

- ✅ **Http Status Codes** — All 15 sampled pages return proper error codes for bad URLs
- ✅ **Redirect Behavior** — No redirects detected across 15 sampled pages

### Observability and Content Health

- **Score:** 83/100 · **Status:** fail
- **Summary:** 2 failed across 3 AFDocs checks.
- **Rationale:** Coverage, parity, and cache behavior determine whether agents can trust the content they retrieve.
- **Reference:** [AFDocs reference](https://afdocs.dev)

**Checks**

- ❌ **LLMS TXT Freshness** — llms.txt covers 427/616 sitemap doc pages (69%); 189 missing; 300 llms.txt links not in sitemap (may indicate stale links or incomplete sitemap) Your llms.txt covers less than 80% of your site's pages. 189 live pages are missing from the index. Regenerate llms.txt from your sitemap or build pipeline.
- ❌ **Markdown Content Parity** — 1 of 9 pages have substantive content differences between markdown and HTML (avg 4% missing) 1 pages have substantive content differences between markdown and HTML (avg 4% missing). Agents receiving the markdown version are getting outdated or incomplete content. Regenerate markdown from source or fix the build pipeline.
- ✅ **Cache Header Hygiene** — All 16 endpoints have appropriate cache headers

### Authentication and Access

- **Score:** 100/100 · **Status:** partial
- **Summary:** 1 skipped across 2 AFDocs checks.
- **Rationale:** Agents need either public access or a clear alternative path when documentation is gated behind auth.
- **Reference:** [AFDocs reference](https://afdocs.dev)

**Checks**

- ✅ **Auth Gate Detection** — All 15 sampled pages are publicly accessible
- ⏭️ **Auth Alternative Access** — All docs pages are publicly accessible; no alternative access paths needed

### Full Content Discoverability

- **Score:** 75/100 · **Status:** fail
- **Summary:** llms-full.txt has 1 failing check.
- **Rationale:** A full-document snapshot gives long-context agents a single canonical corpus to ingest without repeated crawling.
- **Reference:** [llms-full.txt guide](https://www.mintlify.com/docs/ai/llmstxt#llms-full-txt)

**Checks**

- ✅ **LLMS Full Exists** — Found llms-full.txt.
- ✅ **LLMS Full Size** — llms-full.txt size is within the expected range.
- ✅ **LLMS Full Valid** — llms-full.txt has a recognizable markdown structure.
- ❌ **LLMS Full Links Resolve** — Some llms-full.txt links do not resolve successfully.

### Agent Skills

- **Score:** 100/100 · **Status:** pass
- **Summary:** skill.md passes all checks.
- **Rationale:** Agent skills provide product-specific operating guidance that plain documentation pages do not encode on their own.
- **Reference:** [skill.md guide](https://www.mintlify.com/docs/ai/skillmd)

**Checks**

- ✅ **Skill MD** — Found an agent skill definition.

### MCP Server

- **Score:** 0/100 · **Status:** fail
- **Summary:** MCP has 1 failing check.
- **Rationale:** A discoverable MCP server lets agents use first-class tools instead of scraping pages and inferring behavior.
- **Reference:** [MCP guide](https://www.mintlify.com/docs/ai/model-context-protocol)

**Checks**

- ❌ **MCP Server Discoverable** — No MCP server was discovered at the expected endpoints.
- ⏭️ **MCP Tool Count** — Skipped because the MCP server was not discoverable.
