236 lines
8.2 KiB
TypeScript
236 lines
8.2 KiB
TypeScript
import { c as _c } from "react/compiler-runtime";
|
||
import { marked, type Token, type Tokens } from 'marked';
|
||
import React, { Suspense, use, useMemo, useRef } from 'react';
|
||
import { useSettings } from '../hooks/useSettings.js';
|
||
import { Ansi, Box, useTheme } from '../ink.js';
|
||
import { type CliHighlight, getCliHighlightPromise } from '../utils/cliHighlight.js';
|
||
import { hashContent } from '../utils/hash.js';
|
||
import { configureMarked, formatToken } from '../utils/markdown.js';
|
||
import { stripPromptXMLTags } from '../utils/messages.js';
|
||
import { MarkdownTable } from './MarkdownTable.js';
|
||
type Props = {
|
||
children: string;
|
||
/** When true, render all text content as dim */
|
||
dimColor?: boolean;
|
||
};
|
||
|
||
// Module-level token cache — marked.lexer is the hot cost on virtual-scroll
|
||
// remounts (~3ms per message). useMemo doesn't survive unmount→remount, so
|
||
// scrolling back to a previously-visible message re-parses. Messages are
|
||
// immutable in history; same content → same tokens. Keyed by hash to avoid
|
||
// retaining full content strings (turn50→turn99 RSS regression, #24180).
|
||
const TOKEN_CACHE_MAX = 500;
|
||
const tokenCache = new Map<string, Token[]>();
|
||
|
||
// Characters that indicate markdown syntax. If none are present, skip the
|
||
// ~3ms marked.lexer call entirely — render as a single paragraph. Covers
|
||
// the majority of short assistant responses and user prompts that are
|
||
// plain sentences. Checked via indexOf (not regex) for speed.
|
||
// Single regex: matches any MD marker or ordered-list start (N. at line start).
|
||
// One pass instead of 10× includes scans.
|
||
const MD_SYNTAX_RE = /[#*`|[>\-_~]|\n\n|^\d+\. |\n\d+\. /;
|
||
function hasMarkdownSyntax(s: string): boolean {
|
||
// Sample first 500 chars — if markdown exists it's usually early (headers,
|
||
// code fence, list). Long tool outputs are mostly plain text tails.
|
||
return MD_SYNTAX_RE.test(s.length > 500 ? s.slice(0, 500) : s);
|
||
}
|
||
function cachedLexer(content: string): Token[] {
|
||
// Fast path: plain text with no markdown syntax → single paragraph token.
|
||
// Skips marked.lexer's full GFM parse (~3ms on long content). Not cached —
|
||
// reconstruction is a single object allocation, and caching would retain
|
||
// 4× content in raw/text fields plus the hash key for zero benefit.
|
||
if (!hasMarkdownSyntax(content)) {
|
||
return [{
|
||
type: 'paragraph',
|
||
raw: content,
|
||
text: content,
|
||
tokens: [{
|
||
type: 'text',
|
||
raw: content,
|
||
text: content
|
||
}]
|
||
} as Token];
|
||
}
|
||
const key = hashContent(content);
|
||
const hit = tokenCache.get(key);
|
||
if (hit) {
|
||
// Promote to MRU — without this the eviction is FIFO (scrolling back to
|
||
// an early message evicts the very item you're looking at).
|
||
tokenCache.delete(key);
|
||
tokenCache.set(key, hit);
|
||
return hit;
|
||
}
|
||
const tokens = marked.lexer(content);
|
||
if (tokenCache.size >= TOKEN_CACHE_MAX) {
|
||
// LRU-ish: drop oldest. Map preserves insertion order.
|
||
const first = tokenCache.keys().next().value;
|
||
if (first !== undefined) tokenCache.delete(first);
|
||
}
|
||
tokenCache.set(key, tokens);
|
||
return tokens;
|
||
}
|
||
|
||
/**
|
||
* Renders markdown content using a hybrid approach:
|
||
* - Tables are rendered as React components with proper flexbox layout
|
||
* - Other content is rendered as ANSI strings via formatToken
|
||
*/
|
||
export function Markdown(props) {
|
||
const $ = _c(4);
|
||
const settings = useSettings();
|
||
if (settings.syntaxHighlightingDisabled) {
|
||
let t0;
|
||
if ($[0] !== props) {
|
||
t0 = <MarkdownBody {...props} highlight={null} />;
|
||
$[0] = props;
|
||
$[1] = t0;
|
||
} else {
|
||
t0 = $[1];
|
||
}
|
||
return t0;
|
||
}
|
||
let t0;
|
||
if ($[2] !== props) {
|
||
t0 = <Suspense fallback={<MarkdownBody {...props} highlight={null} />}><MarkdownWithHighlight {...props} /></Suspense>;
|
||
$[2] = props;
|
||
$[3] = t0;
|
||
} else {
|
||
t0 = $[3];
|
||
}
|
||
return t0;
|
||
}
|
||
function MarkdownWithHighlight(props) {
|
||
const $ = _c(4);
|
||
let t0;
|
||
if ($[0] === Symbol.for("react.memo_cache_sentinel")) {
|
||
t0 = getCliHighlightPromise();
|
||
$[0] = t0;
|
||
} else {
|
||
t0 = $[0];
|
||
}
|
||
const highlight = use(t0);
|
||
let t1;
|
||
if ($[1] !== highlight || $[2] !== props) {
|
||
t1 = <MarkdownBody {...props} highlight={highlight} />;
|
||
$[1] = highlight;
|
||
$[2] = props;
|
||
$[3] = t1;
|
||
} else {
|
||
t1 = $[3];
|
||
}
|
||
return t1;
|
||
}
|
||
function MarkdownBody(t0) {
|
||
const $ = _c(7);
|
||
const {
|
||
children,
|
||
dimColor,
|
||
highlight
|
||
} = t0;
|
||
const [theme] = useTheme();
|
||
configureMarked();
|
||
let elements: React.ReactNode[];
|
||
if ($[0] !== children || $[1] !== dimColor || $[2] !== highlight || $[3] !== theme) {
|
||
const tokens = cachedLexer(stripPromptXMLTags(children));
|
||
elements = [];
|
||
let nonTableContent = "";
|
||
const flushNonTableContent = function flushNonTableContent() {
|
||
if (nonTableContent) {
|
||
elements.push(<Ansi key={elements.length} dimColor={dimColor}>{nonTableContent.trim()}</Ansi>);
|
||
nonTableContent = "";
|
||
}
|
||
};
|
||
for (const token of tokens) {
|
||
if (token.type === "table") {
|
||
flushNonTableContent();
|
||
elements.push(<MarkdownTable key={elements.length} token={token as Tokens.Table} highlight={highlight} />);
|
||
} else {
|
||
nonTableContent = nonTableContent + formatToken(token, theme, 0, null, null, highlight);
|
||
nonTableContent;
|
||
}
|
||
}
|
||
flushNonTableContent();
|
||
$[0] = children;
|
||
$[1] = dimColor;
|
||
$[2] = highlight;
|
||
$[3] = theme;
|
||
$[4] = elements;
|
||
} else {
|
||
elements = $[4] as React.ReactNode[];
|
||
}
|
||
const elements_0 = elements;
|
||
let t1;
|
||
if ($[5] !== elements_0) {
|
||
t1 = <Box flexDirection="column" gap={1}>{elements_0}</Box>;
|
||
$[5] = elements_0;
|
||
$[6] = t1;
|
||
} else {
|
||
t1 = $[6];
|
||
}
|
||
return t1;
|
||
}
|
||
type StreamingProps = {
|
||
children: string;
|
||
};
|
||
|
||
/**
|
||
* Renders markdown during streaming by splitting at the last top-level block
|
||
* boundary: everything before is stable (memoized, never re-parsed), only the
|
||
* final block is re-parsed per delta. marked.lexer() correctly handles
|
||
* unclosed code fences as a single token, so block boundaries are always safe.
|
||
*
|
||
* The stable boundary only advances (monotonic), so ref mutation during render
|
||
* is idempotent and safe under StrictMode double-rendering. Component unmounts
|
||
* between turns (streamingText → null), resetting the ref.
|
||
*/
|
||
export function StreamingMarkdown({
|
||
children
|
||
}: StreamingProps): React.ReactNode {
|
||
// React Compiler: this component reads and writes stablePrefixRef.current
|
||
// during render by design. The boundary only advances (monotonic), so
|
||
// the ref mutation is idempotent under StrictMode double-render — but the
|
||
// compiler can't prove that, and memoizing around the ref reads would
|
||
// break the algorithm (stale boundary). Opt out.
|
||
'use no memo';
|
||
|
||
configureMarked();
|
||
|
||
// Strip before boundary tracking so it matches <Markdown>'s stripping
|
||
// (line 29). When a closing tag arrives, stripped(N+1) is not a prefix
|
||
// of stripped(N), but the startsWith reset below handles that with a
|
||
// one-time re-lex on the smaller stripped string.
|
||
const stripped = stripPromptXMLTags(children);
|
||
const stablePrefixRef = useRef('');
|
||
|
||
// Reset if text was replaced (defensive; normally unmount handles this)
|
||
if (!stripped.startsWith(stablePrefixRef.current)) {
|
||
stablePrefixRef.current = '';
|
||
}
|
||
|
||
// Lex only from current boundary — O(unstable length), not O(full text)
|
||
const boundary = stablePrefixRef.current.length;
|
||
const tokens = marked.lexer(stripped.substring(boundary));
|
||
|
||
// Last non-space token is the growing block; everything before is final
|
||
let lastContentIdx = tokens.length - 1;
|
||
while (lastContentIdx >= 0 && tokens[lastContentIdx]!.type === 'space') {
|
||
lastContentIdx--;
|
||
}
|
||
let advance = 0;
|
||
for (let i = 0; i < lastContentIdx; i++) {
|
||
advance += tokens[i]!.raw.length;
|
||
}
|
||
if (advance > 0) {
|
||
stablePrefixRef.current = stripped.substring(0, boundary + advance);
|
||
}
|
||
const stablePrefix = stablePrefixRef.current;
|
||
const unstableSuffix = stripped.substring(stablePrefix.length);
|
||
|
||
// stablePrefix is memoized inside <Markdown> via useMemo([children, ...])
|
||
// so it never re-parses as the unstable suffix grows
|
||
return <Box flexDirection="column" gap={1}>
|
||
{stablePrefix && <Markdown>{stablePrefix}</Markdown>}
|
||
{unstableSuffix && <Markdown>{unstableSuffix}</Markdown>}
|
||
</Box>;
|
||
}
|