DreamStack - Symbolic Cognition Simulator
A Python script that implements symbolic cognition using the llama-cpp-python library with a stacked equation structure.
Overview
DreamStack simulates symbolic cognition using this stacked equation structure:
D₀ = A₀ · f₀(x) + B₀ · g₀(x)
D₁ = A₁ · f₁(g₀) + B₁ · g₁(g₀)
D₂ = A₂ · f₂(g₁) + B₂ · g₂(g₁)
... up to a user-defined recursion depth
Each level contains:
fₙ(x): the first output (real)gₙ(x): the next 3 alternate outputs (imaginary)- Recursively process each imaginary output as a new prompt
Prerequisites
- Python 3.8+ installed on your system
- GGUF model file (e.g.,
mistral-7b-instruct-v0.2.Q4_K_M.gguf)
Installation
Install dependencies:
pip install -r requirements.txtDownload a GGUF model:
- Download a GGUF format model (e.g., from Hugging Face)
- Place it in a
models/directory relative to the script - Default expected path:
./models/mistral-7b-instruct-v0.2.Q4_K_M.gguf
Usage
Run the script:
python dreamstack.pyFollow the prompts:
- Enter your initial prompt
- Specify recursion depth (default: 3)
View results:
- Results are displayed in the terminal
- JSON results are saved to
dreamstack_results.json
Configuration
You can modify the model configuration in the DreamStack class:
dreamstack = DreamStack(
model_path="./models/your-model.gguf",
n_ctx=2048, # Context window size
n_threads=4 # Number of threads
)
Output Format
The script generates a JSON structure like:
{
"D_0": {
"prompt": "user prompt",
"real": "f₀ output",
"imaginary": ["g₀₁", "g₀₂", "g₀₃"]
},
"D_1": {
"prompt": "Based on: g₀₁",
"real": "f₁ output",
"imaginary": ["g₁₁", "g₁₂", "g₁₃"]
},
...
}
Example
$ python dreamstack.py
=== DreamStack - Symbolic Cognition Simulator ===
Enter your prompt: What is the meaning of life?
Enter recursion depth (default: 3): 2
Loading model from ./models/mistral-7b-instruct-v0.2.Q4_K_M.gguf...
Model loaded successfully!
=== Starting DreamStack Processing ===
Initial prompt: What is the meaning of life?
Depth: 2
--- Generating D_0 ---
Prompt: What is the meaning of life?
Generating real output...
Real output: The meaning of life is to find purpose and fulfillment...
Generating imaginary output 1/3...
Imaginary output 1: From a philosophical perspective...
Generating imaginary output 2/3...
Imaginary output 2: In terms of biological evolution...
Generating imaginary output 3/3...
Imaginary output 3: From a spiritual viewpoint...
=== Processing D_1 ===
...
Requirements
llama-cpp-python>=0.2.0typing-extensions>=4.0.0
Notes
- The script requires a GGUF format model file
- Processing time depends on model size and recursion depth
- Results are automatically saved to JSON format
- Each layer generates 1 real and 3 imaginary outputs
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
🙋
Ask for provider support