diff --git a/pages/changelog.mdx b/pages/changelog.mdx
index ae0d119..42fd2b7 100644
--- a/pages/changelog.mdx
+++ b/pages/changelog.mdx
@@ -6,6 +6,29 @@ The format is based on [Keep a Changelog](https://keepachangelog.com/en/1.0.0/).
---
+## Unreleased
+
+### Added
+
+- **Prompt Node Text Connections** — Prompt nodes can now receive incoming text connections, enabling LLM-to-Prompt chaining ([#38](https://github.com/shrimbly/node-banana/pull/38))
+ - Connect LLM Generate output directly to Prompt input
+ - Prompt text area becomes read-only when receiving text from another node
+ - Enables automated prompt enhancement workflows
+
+- **LLM Node Enhancements** — Improved parameter controls and output handling ([#38](https://github.com/shrimbly/node-banana/pull/38))
+ - Collapsible Parameters section with temperature (0-2) and max tokens (256-16384) sliders
+ - Copy to clipboard button for generated text output
+ - Max tokens default setting (256-16384) in Project Setup → Node Defaults tab
+ - Visual feedback when text is copied (green checkmark)
+
+### Fixed
+
+- **Node Selection Bug** — Drag selections no longer incorrectly include distant nodes ([#38](https://github.com/shrimbly/node-banana/pull/38))
+ - Uses IQR-based statistical outlier detection to identify and deselect nodes outside selection area
+ - Resolves issue where nodes with undefined React Flow bounds were included in selections
+
+---
+
## 1.2.0 — 2026-01-20
### Added
diff --git a/pages/core-concepts.mdx b/pages/core-concepts.mdx
index 91a2aa7..2c6b6f7 100644
--- a/pages/core-concepts.mdx
+++ b/pages/core-concepts.mdx
@@ -85,10 +85,14 @@ Used for video content. Video data flows as URLs or base64-encoded data.
### Text Handles
Used for string content like prompts and generated text.
-- Found on: Prompt, LLM Generate, Generate Image (input), Generate Video (input)
+- Found on: Prompt (input and output), LLM Generate (input and output), Generate Image (input), Generate Video (input)
- Color: Green
- Accepts: Any text string
+
+ Prompt nodes can now both receive and send text. Connect LLM Generate output to Prompt input to chain text processing nodes together.
+
+
### Reference Handles
Used for organizational purposes, primarily with Split Grid.
diff --git a/pages/nodes.mdx b/pages/nodes.mdx
index 97b7e11..96e7cec 100644
--- a/pages/nodes.mdx
+++ b/pages/nodes.mdx
@@ -7,7 +7,7 @@ This page documents all available node types in Node Banana. Each node serves a
| Node | Purpose | Inputs | Outputs |
|------|---------|--------|---------|
| [Image Input](#image-input) | Load images | — | Image |
-| [Prompt](#prompt) | Text prompts | — | Text |
+| [Prompt](#prompt) | Text prompts | Text (optional) | Text |
| [Generate Image](#generate-image) | AI image generation | Image, Text | Image |
| [Generate Video](#generate-video) | AI video generation | Image, Text | Video |
| [LLM Generate](#llm-generate) | AI text generation | Text, Image | Text |
@@ -46,7 +46,10 @@ import { Callout } from 'nextra/components'
## Prompt
-The Prompt node provides text input for your workflow. Use it to write prompts for image or text generation.
+The Prompt node provides text input for your workflow. Use it to write prompts for image or text generation, or to receive text from other nodes like LLM Generate.
+
+### Inputs
+- **Text** (optional) — Incoming text from another node (e.g., LLM output)
### Outputs
- **Text** — The prompt text string
@@ -55,12 +58,28 @@ The Prompt node provides text input for your workflow. Use it to write prompts f
- Inline text editing
- Expand button for larger editor (modal)
- Full-screen editing mode for complex prompts
+- **Text input connection**: Receive text from LLM Generate or other text-producing nodes
+ - When connected, the text area becomes read-only
+ - Placeholder text indicates "Receiving text from connected node..."
+ - Enables automated prompt workflows and LLM chaining
### Usage
+
+**Manual entry:**
1. Add a Prompt node
2. Type your prompt in the text area
3. Click the expand icon for a larger editor
-4. Connect to Nano Banana or LLM Generate nodes
+4. Connect to Generate Image or LLM Generate nodes
+
+**Receiving text from other nodes:**
+1. Add a Prompt node
+2. Connect an LLM Generate (or other text output) to the Prompt's text input handle
+3. The Prompt automatically receives and displays the connected text
+4. Use the Prompt output to feed into Generate Image nodes or other text consumers
+
+
+ **LLM-to-Prompt Chaining**: Connect an LLM Generate output to a Prompt input to create automated prompt enhancement workflows. For example: `[Initial Prompt] → [LLM: "enhance this prompt"] → [Enhanced Prompt] → [Generate Image]`
+
### Writing Effective Prompts
@@ -218,8 +237,21 @@ The LLM Generate node creates text using large language models. Use it for promp
| Setting | Description |
|---------|-------------|
| **Model** | Select from Gemini or OpenAI models |
-| **Temperature** | Controls randomness (0-1) |
-| **Max Tokens** | Maximum output length |
+| **Temperature** | Controls randomness (0-2) — adjustable in collapsible Parameters section |
+| **Max Tokens** | Maximum output length (256-16384) — adjustable in collapsible Parameters section |
+
+### Parameters
+
+The LLM Generate node includes a collapsible **Parameters** section to configure generation settings without cluttering the node interface:
+
+- **Temperature slider**: 0 to 2.0 in 0.1 increments
+ - Lower values (0-0.5) produce more focused, deterministic output
+ - Higher values (1.5-2.0) produce more creative, varied output
+- **Max Tokens slider**: 256 to 16,384 in 256-token increments
+ - Controls the maximum length of generated text
+ - Higher values allow longer outputs but may increase costs
+
+Click "Parameters" to expand/collapse this section.
### Available Models
@@ -236,12 +268,20 @@ The LLM Generate node creates text using large language models. Use it for promp
OpenAI models require a separate `OPENAI_API_KEY` in your environment.
+### Features
+
+- **Copy to Clipboard**: Click the copy button on generated output to copy text to your clipboard
+ - Button shows a green checkmark when text is copied
+ - Useful for quickly transferring LLM output to other applications
+
### Usage
1. Add an LLM Generate node
2. Connect a Prompt node with your instructions
3. Optionally connect images for multimodal input
-4. Configure model and parameters
-5. Run to generate text
+4. Configure model in the dropdown
+5. (Optional) Expand Parameters section to adjust temperature and max tokens
+6. Run to generate text
+7. Use the copy button to copy output to clipboard
### Example: Prompt Enhancement