DreamStack - Symbolic Cognition Simulator

A Python script that implements symbolic cognition using the llama-cpp-python library with a stacked equation structure.

Overview

DreamStack simulates symbolic cognition using this stacked equation structure:

D₀ = A₀ · f₀(x) + B₀ · g₀(x)
D₁ = A₁ · f₁(g₀) + B₁ · g₁(g₀)
D₂ = A₂ · f₂(g₁) + B₂ · g₂(g₁)
... up to a user-defined recursion depth

Each level contains:

  • fₙ(x): the first output (real)
  • gₙ(x): the next 3 alternate outputs (imaginary)
  • Recursively process each imaginary output as a new prompt

Prerequisites

  1. Python 3.8+ installed on your system
  2. GGUF model file (e.g., mistral-7b-instruct-v0.2.Q4_K_M.gguf)

Installation

  1. Install dependencies:

    pip install -r requirements.txt
    
  2. Download a GGUF model:

    • Download a GGUF format model (e.g., from Hugging Face)
    • Place it in a models/ directory relative to the script
    • Default expected path: ./models/mistral-7b-instruct-v0.2.Q4_K_M.gguf

Usage

  1. Run the script:

    python dreamstack.py
    
  2. Follow the prompts:

    • Enter your initial prompt
    • Specify recursion depth (default: 3)
  3. View results:

    • Results are displayed in the terminal
    • JSON results are saved to dreamstack_results.json

Configuration

You can modify the model configuration in the DreamStack class:

dreamstack = DreamStack(
    model_path="./models/your-model.gguf",
    n_ctx=2048,        # Context window size
    n_threads=4        # Number of threads
)

Output Format

The script generates a JSON structure like:

{
  "D_0": {
    "prompt": "user prompt",
    "real": "f₀ output",
    "imaginary": ["g₀₁", "g₀₂", "g₀₃"]
  },
  "D_1": {
    "prompt": "Based on: g₀₁",
    "real": "f₁ output",
    "imaginary": ["g₁₁", "g₁₂", "g₁₃"]
  },
  ...
}

Example

$ python dreamstack.py

=== DreamStack - Symbolic Cognition Simulator ===

Enter your prompt: What is the meaning of life?
Enter recursion depth (default: 3): 2

Loading model from ./models/mistral-7b-instruct-v0.2.Q4_K_M.gguf...
Model loaded successfully!

=== Starting DreamStack Processing ===
Initial prompt: What is the meaning of life?
Depth: 2

--- Generating D_0 ---
Prompt: What is the meaning of life?
Generating real output...
Real output: The meaning of life is to find purpose and fulfillment...
Generating imaginary output 1/3...
Imaginary output 1: From a philosophical perspective...
Generating imaginary output 2/3...
Imaginary output 2: In terms of biological evolution...
Generating imaginary output 3/3...
Imaginary output 3: From a spiritual viewpoint...

=== Processing D_1 ===
...

Requirements

  • llama-cpp-python>=0.2.0
  • typing-extensions>=4.0.0

Notes

  • The script requires a GGUF format model file
  • Processing time depends on model size and recursion depth
  • Results are automatically saved to JSON format
  • Each layer generates 1 real and 3 imaginary outputs
Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support