From b4c99f0d0a1feb46607dfb668ac130bf2469949a Mon Sep 17 00:00:00 2001 From: Aditya Shiva Sharma Date: Fri, 20 Feb 2026 14:25:18 +0530 Subject: [PATCH] Improve documentation and fix install script URL Change Author Email --- BLOG.md | 6 +++--- CONTRIBUTING.md | 2 +- README.md | 40 ++++++++++++++++++++-------------------- install.sh | 2 +- 4 files changed, 25 insertions(+), 25 deletions(-) diff --git a/BLOG.md b/BLOG.md index d5c2aac..9fa9dd4 100644 --- a/BLOG.md +++ b/BLOG.md @@ -24,8 +24,8 @@ The answer is PicoLLM — a pure-C inference engine with zero dependencies beyon ``` ┌─────────────────────────────────┐ - │ picolm.c (CLI + Gen Loop) │ - │ 227 lines │ + │ picolm.c (CLI + Gen Loop) │ + │ 227 lines │ └──────────┬──────────────────────┘ │ ┌────────────────┼────────────────┐ @@ -503,7 +503,7 @@ make model # Downloads TinyLlama 1.1B Q4_K_M Or use the one-line installer: ```bash -curl -sSL https://raw.githubusercontent.com/picolm/picolm/main/install.sh | bash +curl -sSL https://raw.githubusercontent.com/rightnow-ai/picolm/main/install.sh | bash ``` --- diff --git a/CONTRIBUTING.md b/CONTRIBUTING.md index 95f6db7..ee584b2 100644 --- a/CONTRIBUTING.md +++ b/CONTRIBUTING.md @@ -33,7 +33,7 @@ Thanks for your interest in PicoLLM! This project is intentionally small (~2,500 ### 1. Fork & clone ```bash -git clone https://github.com/picolm/picolm.git +git clone https://github.com/rightnow-ai/picolm.git cd picolm/picolm ``` diff --git a/README.md b/README.md index 8241d06..1d9ad7b 100644 --- a/README.md +++ b/README.md @@ -66,8 +66,8 @@ Telegram / Discord / CLI │ ▼ ┌──────────┐ stdin: prompt ┌───────────┐ - │ PicoClaw │ ──────────────────► │ picolm │ - │ (Go) │ ◄────────────────── │ (C) │ + │ PicoClaw │ ──────────────────► │ picolm │ + │ (Go) │ ◄────────────────── │ (C) │ └──────────┘ stdout: response │ + model │ │ └───────────┘ ▼ 45 MB RAM @@ -157,11 +157,11 @@ The model file (638MB) stays on disk. PicoLM **memory-maps** it and streams one ``` ┌──────────────────────────────────────────┐ What goes │ 45 MB Runtime RAM │ - in RAM │ ┌─────────┐ ┌──────────┐ ┌───────────┐ │ - │ │ Buffers │ │ FP16 KV │ │ Tokenizer │ │ - │ │ 1.2 MB │ │ Cache │ │ 4.5 MB │ │ - │ │ │ │ ~40 MB │ │ │ │ - │ └─────────┘ └──────────┘ └───────────┘ │ + in RAM │ ┌─────────┐ ┌──────────┐ ┌───────────┐ │ + │ │ Buffers │ │ FP16 KV │ │ Tokenizer │ │ + │ │ 1.2 MB │ │ Cache │ │ 4.5 MB │ │ + │ │ │ │ ~40 MB │ │ │ │ + │ └─────────┘ └──────────┘ └───────────┘ │ └──────────────────────────────────────────┘ ┌──────────────────────────────────────────┐ @@ -216,7 +216,7 @@ This will: ### Build from source ```bash -git clone https://github.com/picolm/picolm.git +git clone https://github.com/rightnow-ai/picolm.git cd picolm/picolm # Auto-detect CPU (enables SSE2/AVX on x86, NEON on ARM) @@ -373,8 +373,8 @@ Measured on TinyLlama 1.1B Q4_K_M (638 MB model): │ KV Cache I/O │ ┌──────────┴──────────┐ └───┬────────┬────┘ │ grammar.h/c │ │ │ │ JSON Constraint │ - ┌────────┘ └────────┐ │ Logit Masking │ - │ │ └─────────────────────┘ + ┌────────┘ └───────┐ │ Logit Masking │ + │ │ └─────────────────────┘ ┌─────┴──────┐ ┌───────┴────────┐ │ tensor.h/c │ │ tokenizer.h/c │ │ matmul │ │ BPE Encode │ @@ -410,20 +410,20 @@ Input Token ┌───────────────┐ ×22 layers │ RMSNorm │─────────────────────────────────────────┐ │ │ │ -│ Q = xb @ Wq │ Matrix-vector multiply (quantized) │ -│ K = xb @ Wk │ Store K,V in FP16 KV cache │ -│ V = xb @ Wv │ │ +│ Q = xb @ Wq │ Matrix-vector multiply (quantized) │ +│ K = xb @ Wk │ Store K,V in FP16 KV cache │ +│ V = xb @ Wv │ │ │ │ │ -│ RoPE(Q, K) │ Rotary position encoding (table lookup) │ +│ RoPE(Q, K) │ Rotary position encoding (table lookup)│ │ │ │ -│ Attention │ Flash attention with online softmax │ -│ (GQA 32→4) │ Grouped-query: 32 Q heads, 4 KV heads │ +│ Attention │ Flash attention with online softmax │ +│ (GQA 32→4) │ Grouped-query: 32 Q heads, 4 KV heads │ │ │ │ -│ x += Out@Wo │ Output projection + residual │ +│ x += Out@Wo │ Output projection + residual │ │ │ │ │ RMSNorm │ │ │ │ │ -│ SwiGLU FFN │ gate=SiLU(xb@Wg), up=xb@Wu │ +│ SwiGLU FFN │ gate=SiLU(xb@Wg), up=xb@Wu │ │ │ x += (gate*up) @ Wd │ └───────┬───────┘─────────────────────────────────────────┘ │ @@ -545,8 +545,8 @@ PicoLM/ │ ├── tokenizer.h/c ← BPE tokenizer (32 + ~200 lines) │ ├── sampler.h/c ← temperature + top-p sampling (19 + ~100 lines) │ ├── grammar.h/c ← JSON grammar constraints (64 + 175 lines) -│ ├── Makefile ← build targets for all platforms -│ └── build.bat ← Windows MSVC build script +│ ├── Makefile ← build targets for all platforms +│ └── build.bat ← Windows MSVC build script │ └── tinyllama-1.1b-chat-v1.0.Q4_K_M.gguf ← model file (638 MB, not in git) ``` diff --git a/install.sh b/install.sh index 1b630ea..366b5ce 100644 --- a/install.sh +++ b/install.sh @@ -3,7 +3,7 @@ # PicoLLM + PicoClaw installer for Raspberry Pi & Linux # # Usage: -# curl -sSL https://raw.githubusercontent.com/picolm/picolm/main/install.sh | bash +# curl -sSL https://raw.githubusercontent.com/rightnow-ai/picolm/main/install.sh | bash # # Or locally: # chmod +x install.sh && ./install.sh