Skip to content

Commit 4aee7aa

Browse files
committed
fix(docs): add language tags to 22 bare code fences across 13 pages
1 parent 80faf8f commit 4aee7aa

File tree

13 files changed

+22
-22
lines changed

13 files changed

+22
-22
lines changed

content/docs/api/serve.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -6,7 +6,7 @@ bookToc: true
66

77
# Package serve
88

9-
```
9+
```go
1010
import "github.com/zerfoo/zerfoo/serve"
1111
```
1212

content/docs/architecture/gpu-setup.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -32,7 +32,7 @@ This means:
3232

3333
The detection flow:
3434

35-
```
35+
```text
3636
1. dlopen("libcudart.so.12") or dlopen("libcudart.so")
3737
2. dlsym each required symbol (cudaMalloc, cudaFree, cudaMemcpy, ...)
3838
3. Optionally resolve CUDA graph symbols (cudaStreamBeginCapture, ...)
@@ -55,7 +55,7 @@ nvidia-smi
5555

5656
Expected output shows your GPU model, driver version, and CUDA version:
5757

58-
```
58+
```text
5959
+-----------------------------------------------------------------------------------------+
6060
| NVIDIA-SMI 560.35.03 Driver Version: 560.35.03 CUDA Version: 12.6 |
6161
|-----------------------------------------+------------------------+----------------------+
@@ -182,7 +182,7 @@ rocm-smi
182182

183183
Expected output:
184184

185-
```
185+
```text
186186
========================= ROCm System Management Interface =========================
187187
================================ Concise Info =======================================
188188
GPU Temp AvgPwr SCLK MCLK Fan Perf PwrCap VRAM% GPU%

content/docs/architecture/overview.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -168,7 +168,7 @@ statement (with special-case detection -- e.g., Mistral models report
168168
Most decoder-only architectures share the same transformer body. The shared
169169
logic lives in `buildTransformerGraph()`, which constructs:
170170

171-
```
171+
```text
172172
Embed -> [RMSNorm -> GQA -> Add -> RMSNorm -> FFN(SiLU-gate) -> Add] x N -> RMSNorm -> LMHead
173173
```
174174

@@ -433,7 +433,7 @@ flowchart TD
433433

434434
A quick reference for where to find things:
435435

436-
```
436+
```text
437437
zerfoo/
438438
cmd/ CLI entry points (run, serve, pull, predict, tokenize)
439439
inference/

content/docs/blog/03-architecture-deep-dive.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@ Zerfoo runs LLM inference in Go at 245 tokens/second — 20% faster than Ollama.
1212

1313
When you call `zerfoo.Load("google/gemma-3-4b")` followed by `m.Chat("Hello")`, the following pipeline executes:
1414

15-
```
15+
```text
1616
GGUF file on disk
1717
-> Parse GGUF header + tensors
1818
-> Map tensor names to canonical form
@@ -86,7 +86,7 @@ func init() {
8686

8787
The `general.architecture` field in the GGUF metadata determines which builder is invoked. Most decoder-only architectures share the same transformer body through `buildTransformerGraph()`, which constructs:
8888

89-
```
89+
```text
9090
Embed -> [RMSNorm -> GQA -> Add -> RMSNorm -> FFN(SiLU-gate) -> Add] x N -> RMSNorm -> LMHead
9191
```
9292

content/docs/blog/how-we-beat-ollama-cuda-graph-capture.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -38,7 +38,7 @@ This is a hard problem for a real inference pipeline. Token embeddings require l
3838

3939
Zerfoo solves this by splitting the execution plan into three regions:
4040

41-
```
41+
```text
4242
[Pre-capture: CPU-touching ops] [Capture region: GPU-only ops] [Post-capture: CPU-touching ops]
4343
```
4444

content/docs/contributing/overview.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -27,7 +27,7 @@ Zerfoo is an ecosystem of six independent repositories (each with its own `go.mo
2727

2828
**Dependency graph:**
2929

30-
```
30+
```text
3131
float16 --+
3232
float8 --+--> ztensor --> zerfoo
3333
ztoken --+
@@ -96,7 +96,7 @@ go tool cover -html=coverage.out -o coverage.html
9696

9797
We use [Conventional Commits](https://www.conventionalcommits.org/) for automated versioning with release-please.
9898

99-
```
99+
```text
100100
<type>(<scope>): <description>
101101
```
102102

@@ -112,7 +112,7 @@ We use [Conventional Commits](https://www.conventionalcommits.org/) for automate
112112

113113
Examples:
114114

115-
```
115+
```text
116116
feat(inference): add Qwen 2.5 architecture support
117117
fix(generate): correct KV cache eviction for sliding window attention
118118
perf(layers): fuse SiLU and gate projection into single kernel

content/docs/deployment/enterprise.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1027,7 +1027,7 @@ readinessProbe:
10271027

10281028
Zerfoo logs every request with structured fields:
10291029

1030-
```
1030+
```text
10311031
method=POST path=/v1/chat/completions model=gemma-3-1b prompt_tokens=0 completion_tokens=0 latency_ms=142 status_code=200
10321032
```
10331033

@@ -1093,7 +1093,7 @@ sidecar. SHA-256 is computed and stored on upload.
10931093

10941094
#### Directory Layout
10951095

1096-
```
1096+
```text
10971097
/models/
10981098
llama-3-7b-q4_k_m/
10991099
model.gguf
@@ -1176,7 +1176,7 @@ a new load would exceed the budget.
11761176

11771177
### Architecture
11781178

1179-
```
1179+
```text
11801180
Request -> ModelManager.Get("model-id")
11811181
|
11821182
+-- Already loaded? -> promote to MRU, return handle

content/docs/ecosystem/_index.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -11,7 +11,7 @@ Zerfoo is a family of Go modules that together form a complete ML inference and
1111

1212
## Dependency Graph
1313

14-
```
14+
```text
1515
float16 ──┐
1616
├──► ztensor ──► zerfoo
1717
float8 ──┘ ▲

content/docs/getting-started/first-inference.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -80,7 +80,7 @@ Start a chat session with `zerfoo run`:
8080
zerfoo run gemma-3-1b-q4
8181
```
8282

83-
```
83+
```text
8484
Model loaded. Type your message (Ctrl-D to quit).
8585
8686
> What is the capital of France?

content/docs/getting-started/quickstart.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -55,7 +55,7 @@ go run main.go
5555

5656
To request a specific quantization, append it to the ID:
5757

58-
```
58+
```text
5959
google/gemma-3-4b/Q8_0
6060
```
6161

@@ -107,7 +107,7 @@ zerfoo run gemma-3-1b-q4
107107

108108
This starts an interactive chat session:
109109

110-
```
110+
```text
111111
Model loaded. Type your message (Ctrl-D to quit).
112112
113113
> What is the capital of France?

0 commit comments

Comments
 (0)