markdown-it-ts
A TypeScript-first, markdown-it compatible Markdown parser and renderer with streaming/incremental parsing and async render.
English | 简体中文
Quick links: Docs index · Stream optimization · Performance report · Compatibility report
A TypeScript migration of markdown-it with modular architecture for tree-shaking and separate parse/render imports.
🚀 Migration Status: 100% Complete
Port from markdown-it to TypeScript is complete and maintained with the following goals:
- ✅ Full TypeScript type safety
- ✅ Modular architecture (separate parse/render imports)
- ✅ Tree-shaking support
- ✅ Ruler-based rule system
- ✅ API compatibility with original markdown-it
What's Implemented
✅ Core System (100%)
- All 7 core rules (normalize, block, inline, linkify, replacements, smartquotes, text_join)
- CoreRuler with enable/disable/getRules support
- Full parsing pipeline
✅ Block System (100%)
- All 11 block rules:
- table (GFM tables)
- code (indented code blocks)
- fence (fenced code blocks)
- blockquote (block quotes)
- hr (horizontal rules)
- list (bullet and ordered lists with nesting)
- reference (link reference definitions)
- html_block (raw HTML blocks)
- heading (ATX headings
#)
- lheading (Setext headings
===)
- paragraph (paragraphs)
- StateBlock with full line tracking (200+ lines)
- BlockRuler implementation (80 lines)
- ParserBlock refactored with Ruler pattern
✅ Inline System (100%)
- All 12 inline rules (text, escape, linkify, strikethrough, etc.) with full post-processing coverage
- StateInline with 18 properties, 3 methods
- InlineRuler implementation mirroring markdown-it behavior
✅ Renderer & Infrastructure (100%)
- Renderer ported from markdown-it with attribute handling & highlight support
- Type definitions with Token interface and renderer options
- Helper functions (parseLinkLabel, parseLinkDestination, parseLinkTitle)
- Common utilities (html_blocks, html_re, utils)
markdownit() instances expose render, renderInline, and renderer for plugin compatibility
Installation
npm install markdown-it-ts
Usage
Basic Parsing (Current State)
import markdownIt from 'markdown-it-ts'
const md = markdownIt()
const tokens = md.parse('# Hello World')
console.log(tokens)
Rendering Markdown
Use the built-in renderer for full markdown-it compatibility:
import markdownIt from 'markdown-it-ts'
const md = markdownIt()
const html = md.render('# Hello World')
console.log(html)
Large Inputs
For normal usage, keep the original markdown-it-compatible API:
const md = markdownIt()
const tokens = md.parse(hugeMarkdown)
const html = md.render(hugeMarkdown)
Those default parse / render calls now auto-activate the internal large-input path once the document crosses the large-document threshold, so users do not need a separate API just to benefit from huge-text optimizations. If needed, this can be disabled with autoUnbounded: false.
Use the explicit stream-oriented APIs only when your upstream input already arrives as chunks and you do not want to join it into one giant string first:
import MarkdownIt, { UnboundedBuffer } from 'markdown-it-ts'
const md = MarkdownIt()
const tokens = md.parseIterable(fileChunks)
const buffer = new UnboundedBuffer(md, { mode: 'stream' })
for await (const chunk of logChunks) {
buffer.feed(chunk)
buffer.flushAvailable()
}
const finalTokens = buffer.flushForce()
parseIterable / parseAsyncIterable are advanced entry points for explicit Iterable<string> / AsyncIterable<string> inputs. UnboundedBuffer is the advanced append-only path for real chunk streams and only keeps a bounded tail in memory instead of the whole historical source string.
If you also need bounded output memory for explicit chunk-stream inputs, use the sink form instead of retaining a full token array:
md.parseIterableToSink(fileChunks, (tokens, info) => {
consumeTokenChunk(tokens, info)
})
For arbitrary in-place edits, use EditableBuffer. It stores the source in a piece table and reparses only from an anchor before the affected block instead of flattening and reparsing the whole document every time. Internally, both the full parse and the localized reparse paths now hand a PieceTableSourceView straight to md.core.parseSource(...), so the selected range no longer needs to be materialized as one giant intermediate string first.
Need async renderer rules (for example, asynchronous syntax highlighting)? Use renderAsync which awaits async rule results:
const md = markdownIt()
const html = await md.renderAsync('# Hello World', {
highlight: async (code, lang) => {
const highlighted = await someHighlighter(code, lang)
return highlighted
},
})
If you initially import core-only and want to attach rendering (to keep bundles smaller when only parse is needed elsewhere), use the provided helper:
import markdownIt, { withRenderer } from 'markdown-it-ts'
const md = withRenderer(markdownIt())
const html = md.render('# Hello World')
console.log(html)
Documentation
Why render with markdown-it-ts?
- Compared with markdown-it: same API/plugin surface, but rewritten in TypeScript with a modular architecture that can be tree-shaken and that ships streaming/chunked strategies. Normal
parse / render usage stays unchanged, large finite strings now auto-enable internal large-input optimizations, and editor-style flows can additionally opt into stream, streamChunkedFallback, etc., to re-parse only appended content instead of reprocessing entire documents.
- Compared with markdown-exit: both projects target speed, but markdown-it-ts stays 100% compatible with markdown-it plugins, offers typed APIs plus async rendering (
renderAsync), and exposes richer tuning knobs (fence-aware chunking, hybrid fallback modes). In our 5k–100k measurements, markdown-it-ts consistently leads one-shot parse latency (see “Parse ranking”), and its streaming path keeps append latency far lower than re-running a full parse per keystroke.
- Compared with remark: remark’s strength is AST transforms, yet rendering Markdown → HTML usually requires a rehype/rehype-stringify pipeline, which adds significant overhead (our measurements show ~29× slower HTML rendering at 20k chars). markdown-it-ts produces HTML directly, keeps markdown-it renderer semantics, and still supports async highlighting or token post-processing, which makes it a better fit for real-time preview, SSR, or any latency-sensitive render workload.
- Compared with micromark: micromark is a CommonMark-oriented reference implementation that can render Markdown → HTML directly. markdown-it-ts targets markdown-it’s plugin API and renderer semantics while keeping end-to-end render throughput competitive (see “Render vs micromark” below).
- Developer experience: Type definitions and tuning helpers ship in the package (
docs/stream-optimization.md, recommend*Strategy APIs, StreamBuffer, chunkedParse, etc.), so teams can build adaptive streaming pipelines quickly. The repository’s benchmark scripts (perf:generate, perf:update-readme) keep comparison data up to date in CI, reducing the risk of unnoticed regressions.
- Drop-in compatibility: markdown-it-ts preserves the ruler system, Token shape, and plugin hooks, so most existing markdown-it plugins just work after changing the import. For parse-only bundles you can opt into rendering later via
withRenderer, enabling incremental migrations.
- Production readiness: async render, Token-level post-processing, streaming buffers, and chunked fallbacks serve SSR, collaborative editors, and large batch pipelines alike. With
docs/perf-report.md plus long-term history (docs/perf-history/*.json) you can track performance trends over time and catch regressions early.
Customization
You can customize the parser and renderer by enabling or disabling specific rules:
import markdownIt from 'markdown-it-ts'
const md = markdownIt()
.enable(['linkify', 'typographer'])
.disable('html')
const result = md.render('Some markdown content')
console.log(result)
Demo
Build the demo site into ./demo and open it in your browser.
Note: the demo build uses the current project's published build artifact (the files in dist/). The demo script runs npm run build before bundling, so the demo reflects the current repo source.
This ensures demo/markdown-it.js is produced from the most recent dist/index.js output.
Generating API docs
You can generate API documentation into ./apidoc using the built-in script. The script will attempt to use pnpm dlx or npx if available, otherwise it uses the locally-installed ndoc from node_modules.
npm run build
npm run doc
open apidoc/index.html
xdg-open apidoc/index.html
Continuous Integration
This repository includes a GitHub Actions workflow that runs on push and pull requests to main. The CI job verifies the TypeScript build, linting, API docs generation and demo build to help catch regressions early.
Files to inspect: .github/workflows/ci-docs.yml
Deploying to Netlify
You can deploy both the generated API docs (apidoc/) and the demo site (demo/) to Netlify. There are two supported workflows:
- Manual / CLI deploy (local)
- Create two Netlify sites (one for docs and one for demo), or use two separate site IDs under the same account.
- Install
netlify-cli locally or use the helper scripts included in package.json.
Deploy docs locally:
export NETLIFY_AUTH_TOKEN=your_token_here
export NETLIFY_SITE_ID_DOCS=your_docs_site_id
pnpm run netlify:deploy:docs
Deploy demo locally:
export NETLIFY_AUTH_TOKEN=your_token_here
export NETLIFY_SITE_ID_DEMO=your_demo_site_id
pnpm run netlify:deploy:demo
- CI-driven deploy (recommended)
The repo contains two GitHub Actions workflows, one for docs and one for demo. Each workflow will only run if you add the required secrets to the repository:
- NETLIFY_AUTH_TOKEN — a Netlify Personal Access Token with deploy permissions
- NETLIFY_SITE_ID_DOCS — the Site ID for the docs site
- NETLIFY_SITE_ID_DEMO — the Site ID for the demo site
Add these as GitHub Secrets for the repository (Settings → Secrets and variables → Actions). When pushed to main, the workflows will run and deploy to the corresponding Netlify site.
Files to inspect: .github/workflows/deploy-netlify-docs.yml and .github/workflows/deploy-netlify-demo.yml
Automatic CI deploy: when you push to main, the CI workflow will build the project, generate docs, and build the demo. After a successful build the workflow attempts to deploy both apidoc/ and demo/ to Netlify automatically — but only if the corresponding GitHub Actions secrets are set:
NETLIFY_AUTH_TOKEN — Netlify Personal Access Token
NETLIFY_SITE_ID_DOCS — Netlify Site ID for the docs site
NETLIFY_SITE_ID_DEMO — Netlify Site ID for the demo site
If those secrets exist, the CI will publish both sites. If not, the CI will skip publishing and still report build/lint/docs/demo status.
npm run gh-demo
If you only want to build the demo (skip publishing) you can run:
npm run demo
To publish the demo automatically set GH_PAGES_REPO to your target repo (you must have push access):
export GH_PAGES_REPO='git@github.com:youruser/markdown-it.github.io.git'
npm run gh-demo
Subpath exports
For advanced or tree-shaken imports you can target subpaths directly:
import { Token } from 'markdown-it-ts/common/token'
import { withRenderer } from 'markdown-it-ts/plugins/with-renderer'
import Renderer from 'markdown-it-ts/render/renderer'
import { StreamBuffer } from 'markdown-it-ts/stream/buffer'
import { chunkedParse } from 'markdown-it-ts/stream/chunked'
import { DebouncedStreamParser, ThrottledStreamParser } from 'markdown-it-ts/stream/debounced'
Plugin Authoring (Type-Safe)
Plugins are regular functions that receive the markdown-it-ts instance. For full type-safety use the exported MarkdownItPlugin type:
import markdownIt, { MarkdownItPlugin } from 'markdown-it-ts'
const plugin: MarkdownItPlugin = (md) => {
md.core.ruler.after('block', 'my_rule', (state) => {
})
}
const md = markdownIt().use(plugin)
Performance tips
For large documents or append-heavy editing flows, you can enable the stream parser and an optional chunked fallback. See the detailed guide in docs/stream-optimization.md.
Quick start:
import markdownIt from 'markdown-it-ts'
const md = markdownIt({
stream: true,
streamChunkedFallback: true,
streamChunkSizeChars: 10_000,
streamChunkSizeLines: 200,
streamChunkFenceAware: true,
})
let src = '# Title\n\nHello'
md.parse(src, {})
src += '\nworld!'
md.parse(src, {})
Try the quick benchmark (build first):
npm run build
node scripts/quick-benchmark.mjs
More:
- Full performance matrix across modes and sizes:
npm run perf:matrix
- Non-stream chunked sweep to tune thresholds:
npm run perf:sweep
- Parser family hotspot report:
pnpm run perf:families
- Long-text default strategy matrix:
pnpm run perf:strategies
- Independent default-strategy perf gate:
pnpm run perf:gate
- See detailed findings in
docs/perf-report.md.
- See long-text strategy docs in
docs/stream-optimization.md and docs/parse-strategy-matrix.md.
Adaptive chunk sizing
- Non-stream full fallback now chooses chunk size automatically by default (
fullChunkAdaptive: true), targeting ~8 chunks and clamping sizes into practical ranges.
- Stream chunked fallback also uses adaptive sizing by default (
streamChunkAdaptive: true).
- You can restore fixed sizes by setting the respective
*Adaptive: false flags or by providing explicit *SizeChars/*SizeLines values.
Programmatic recommendations
If you want to display or persist the suggested chunk settings without enabling auto-tune, you can query them directly:
import markdownIt, { recommendFullChunkStrategy, recommendStreamChunkStrategy } from 'markdown-it-ts'
const size = 50_000
## Streaming performance recommendations (summary)
Short summary of interactive/workflow guidance (see `docs/perf-latest.md` for full details):
- Enable `stream` + caching for append-heavy editors — it gives the best append throughput in most sizes.
- Prefer paragraph-level batching when feeding edits; line-by-line updates are more expensive and reduce the streaming speedup.
- For non-append edits (in-place paragraph edits), expect full reparses; the baseline parser often outperforms incremental approaches for these cases.
### Key performance summary (selected winners)
Quick winners from the latest run (see `docs/perf-latest.md` for full tables):
- Best one-shot parse (by document size):
- 5,000 chars: **S3** (stream ON, cache ON, chunk ON) — 0.0002ms
- 20,000 chars: **S2** (stream ON, cache ON, chunk OFF) — 0.0002ms
- 50,000 chars: **S3** (stream ON, cache ON, chunk ON) — 0.0004ms
- 100,000 chars: **S1** (stream ON, cache OFF, chunk ON) — 0.0006ms
- 200,000 chars: **S2** (stream ON, cache ON, chunk OFF) — 12.18ms
- Best append (paragraph-level) throughput:
- 5,000 chars: **S3** — 0.3560ms
- 20,000 chars: **S3** — 1.2651ms
- 50,000 chars: **S3** — 3.3976ms
- 100,000 chars: **S2** — 6.8648ms
- 200,000 chars: **S2** — 25.56ms
- Best append (line-level, finer-grained):
- 5,000 chars: **S3** — 0.8666ms
- 20,000 chars: **S2** — 5.4193ms
- 50,000 chars: **S2** — 5.6287ms
- 100,000 chars: **S2** — 9.7292ms
- 200,000 chars: **S3** — 42.30ms
- Best replace (in-place paragraph edits): baseline `markdown-it` often wins for larger docs:
- 5,000 chars: **S3** — 0.2964ms
- 20,000 chars: **M1** (markdown-it) — 0.8474ms
- 50,000 chars: **M1** — 2.0403ms
- 100,000 chars: **M1** — 4.0348ms
- 200,000 chars: **M1** — 8.3294ms
Notes: these numbers are from the most recent run and included as illustrative guidance. For exact, per-scenario numbers and environment details, consult `docs/perf-latest.json`.
### Example: per-scenario timings at 20,000 chars
The table below shows a compact, side-by-side comparison for a 20,000-char document (numbers taken from `docs/perf-latest.md` / `docs/perf-latest.json`). Columns are: one-shot parse time, paragraph-level append workload, line-level append workload, and replace-paragraph workload (all times in milliseconds). Lower is better.
| Scenario | Config summary | One-shot | Append (paragraph) | Append (line) | Replace (paragraph) |
|:--|:--|---:|---:|---:|---:|
| S1 | stream ON, cache OFF, chunk ON | 0.0003ms | 3.9113ms | 10.91ms | 1.1784ms |
| S2 | stream ON, cache ON, chunk OFF | **0.0002ms** | 1.3094ms | **5.4193ms** | 0.8797ms |
| S3 | stream ON, cache ON, chunk ON | 0.0002ms | **1.2651ms** | 6.5309ms | 1.1191ms |
| S4 | stream OFF, chunk ON | 1.2229ms | 3.9489ms | 10.68ms | 1.2995ms |
| S5 | stream OFF, chunk OFF | 0.9306ms | 3.2370ms | 8.6026ms | 1.1024ms |
| M1 | markdown-it (baseline) | 0.8803ms | 2.8267ms | 7.7509ms | **0.8474ms** |
Notes: bolded values indicate the best (lowest) time in that column for this document size.
Reproduce the measurements:
```bash
pnpm run build
node scripts/perf-generate-report.mjs
Report outputs: docs/perf-latest.md and docs/perf-latest.json.
const fullRec = recommendFullChunkStrategy(size)
// { strategy: 'plain', fenceAware: true }
const streamRec = recommendStreamChunkStrategy(size)
// { strategy: 'discrete', maxChunkChars: 16_000, maxChunkLines: 250, fenceAware: true }
These mirror the same mappings used internally when `autoTuneChunks: true` and no explicit sizes are provided.
### Performance regression checks
To make sure each change is not slower than the previous run at any tested size/config, we ship a tiny perf harness and a comparator:
- Generate the latest report and snapshot:
- `npm run perf:generate` → writes `docs/perf-latest.md` and `docs/perf-latest.json`
- Also archives `docs/perf-history/perf-<shortSHA>.json` when git is available
- Compare two snapshots (fail on regressions beyond threshold):
- `node scripts/perf-compare.mjs docs/perf-latest.json docs/perf-history/perf-<baselineSHA>.json --threshold=0.10`
- Accept the latest run as the new baseline (after manual review):
- `pnpm run perf:accept`
- Run the regression check against the most recent baseline (same harness):
- `pnpm run perf:check:latest`
- Run the per-token-type render benchmark against `markdown-it`:
- `pnpm run perf:render-rules`
- Add `--include-noise` to also show zero-token / sub-signal categories
- Use `pnpm run perf:render-rules:check` to fail if any meaningful category regresses beyond the threshold
- Run the parser rule-family hotspot benchmark:
- `pnpm run perf:families`
- Writes `docs/perf-family-hotspots.md` and `docs/perf-family-hotspots.json`
- Run the long-text default-strategy benchmark and gate:
- `pnpm run perf:strategies`
- `pnpm run perf:strategy:check`
- `pnpm run perf:gate`
- Writes `docs/perf-large-defaults.*` and `docs/parse-strategy-matrix.md`
- Inspect detailed deltas by size/scenario (sorted by worst):
- `pnpm run perf:diff`
See `docs/perf-regression.md` for details and CI usage.
## Upstream Test Suites (optional)
This repo can run a subset of the original markdown-it tests and pathological cases. They are disabled by default because they require:
- A sibling checkout of the upstream `markdown-it` repo (referenced by relative path in tests)
- Network access for fetching reference scripts
To enable upstream tests locally:
```bash
# Ensure directory layout like:
# ../markdown-it/ # upstream repo with index.mjs and fixtures
# ./markdown-it-ts/ # this repo
RUN_ORIGINAL=1 pnpm test
Notes
- Pathological tests are heavy and use worker threads and network; enable only when needed.
- CI keeps these disabled by default.
Alternative: set a custom upstream path without sibling layout
MARKDOWN_IT_DIR=/absolute/path/to/markdown-it RUN_ORIGINAL=1 pnpm test
Convenience scripts
pnpm run test:original
pnpm run test:original:network
Parse performance vs markdown-it
Latest one-shot parse results on this machine (Node.js v23): markdown-it-ts is roughly at parity with upstream markdown-it in the 5k–100k range.
Examples from the latest run (avg over 20 iterations):
- 5,000 chars: 0.1654ms vs 0.2015ms → ~1.2× faster, ~18% less time
- 20,000 chars: 0.6545ms vs 0.7885ms → ~1.2× faster, ~17% less time
- 100,000 chars: 4.0118ms vs 5.3204ms → ~1.3× faster, ~25% less time
- 500,000 chars: 24.42ms vs 28.78ms → ~1.2× faster, ~15% less time
- 1,000,000 chars: 54.82ms vs 57.37ms → ~1× faster, ~4% less time
- Notes
- Numbers vary by Node version, CPU, and content shape; see
docs/perf-latest.md for the full table and environment details.
- Streaming/incremental mode is correctness-first by default. For editor-style input, using
StreamBuffer to flush at block boundaries can yield meaningful wins on append-heavy workloads.
We also compare parse-only performance against remark (parse-only). The following figures are taken from docs/perf-latest.json and show one-shot parse times and append-workload times reported by the harness.
One-shot parse (oneShotMs) — markdown-it-ts vs remark (lower is better):
- 5,000 chars: 0.1654ms vs 5.2773ms → 31.9× faster
- 20,000 chars: 0.6545ms vs 23.81ms → 36.4× faster
- 100,000 chars: 4.0118ms vs 161.03ms → 40.1× faster
Append workload (appendWorkloadMs) — markdown-it-ts vs remark:
- 5,000 chars: 0.2487ms vs 16.34ms → 65.7× faster
- 20,000 chars: 1.0456ms vs 81.83ms → 78.3× faster
- 100,000 chars: 5.4090ms vs 530.05ms → 98× faster
Parse performance vs micromark
We also compare parse-only performance against micromark (scenario MM1), measured via its preprocess+parse+postprocess pipeline (no HTML compile). Numbers are taken from docs/perf-latest.json.
One-shot parse (oneShotMs) — markdown-it-ts vs micromark-based parse:
- 5,000 chars: 0.1654ms vs 4.1778ms → 25.3× faster
- 20,000 chars: 0.6545ms vs 18.12ms → 27.7× faster
- 100,000 chars: 4.0118ms vs 103.48ms → 25.8× faster
Append workload (appendWorkloadMs) — markdown-it-ts vs micromark-based parse:
- 5,000 chars: 0.2487ms vs 13.89ms → 55.9× faster
- 20,000 chars: 1.0456ms vs 61.35ms → 58.7× faster
- 100,000 chars: 5.4090ms vs 348.15ms → 64.4× faster
Parse performance vs markdown-exit
The following shows one-shot parse times (oneShotMs) comparing the best markdown-it-ts scenario against markdown-exit (E1) from the latest perf snapshot.
| 5,000 | 0.1654ms | 0.2662ms |
| 20,000 | 0.6545ms | 1.0446ms |
| 50,000 | 1.7409ms | 2.7569ms |
| 100,000 | 4.0118ms | 6.4988ms |
| 200,000 | 10.50ms | 13.83ms |
Notes: markdown-it-ts remains substantially faster for small one-shot parses due to streaming/chunk strategies; for very large documents (200k+) raw one-shot times are closer between implementations. See docs/perf-latest.json for full details.
Notes on interpretation
- These numbers compare parse-only times produced by the project's perf harness.
remark workflows often include additional tree transforms/plugins; real-world workloads may differ.
- Results are machine- and content-dependent. For reproducible comparisons run the local harness and compare
docs/perf-latest.json or the archived docs/perf-history/*.json files.
- Source:
docs/perf-latest.json (one-shot and appendWorkload values).
Render performance (markdown → HTML)
We also profile end-to-end md.render API throughput (parse + render) across markdown-it-ts, upstream markdown-it, and a remark+rehype pipeline. This section is intentionally about the full render(markdown) call, not the lower-level renderer-only hot path. Numbers below come from the latest pnpm run perf:generate snapshot.
For large finite strings, these numbers already include the default automatic large-input path; users do not need to switch to parseIterable / UnboundedBuffer to get those optimizations. Those advanced APIs are reserved for explicit chunk-stream inputs.
vs markdown-it render API
- 5,000 chars: 0.1877ms vs 0.2377ms → ~1.3× faster
- 20,000 chars: 0.7574ms vs 0.9580ms → ~1.3× faster
- 100,000 chars: 5.0904ms vs 6.3302ms → ~1.2× faster
- 500,000 chars: 37.12ms vs 41.10ms → ~1.1× faster
- 1,000,000 chars: 69.30ms vs 89.72ms → ~1.3× faster
- 5,000 chars: 0.1877ms vs 4.7880ms → ~25.5× faster
- 20,000 chars: 0.7574ms vs 27.23ms → ~36× faster
- 100,000 chars: 5.0904ms vs 175.89ms → ~34.6× faster
vs micromark (CommonMark reference)
- 5,000 chars: 0.1877ms vs 3.9821ms → ~21.2× faster
- 20,000 chars: 0.7574ms vs 22.72ms → ~30× faster
- 100,000 chars: 5.0904ms vs 123.07ms → ~24.2× faster
Reproduce locally
pnpm build
node scripts/quick-benchmark.mjs
This will update docs/perf-latest.md and refresh the snippet above.
vs markdown-exit render API
- 5,000 chars: 0.1877ms vs 0.3051ms → ~1.6× faster
- 20,000 chars: 0.7574ms vs 1.2242ms → ~1.6× faster
- 50,000 chars: 2.0238ms vs 3.2443ms → ~1.6× faster
- 100,000 chars: 5.0904ms vs 7.6242ms → ~1.5× faster
- 200,000 chars: 12.23ms vs 18.20ms → ~1.5× faster
Contributing
Contributions are welcome! Please open an issue or submit a pull request for any enhancements or bug fixes.
Acknowledgements
markdown-it-ts is a TypeScript re-implementation that stands on the shoulders of
markdown-it. We are deeply grateful to
the original project and its maintainers and contributors (notably Vitaly Puzrin and
the markdown-it community). Many ideas, algorithms, renderer behaviors, specs, and
fixtures originate from markdown-it; this project would not exist without that work.
License
This project is licensed under the MIT License. See the LICENSE file for more details.