Metadata-Version: 2.4
Name: bielik
Version: 0.2.6
Summary: Bielik — Local-first AI assistant with HuggingFace and SpeakLeash models
Author: Tom Sapletta
License-Expression: Apache-2.0
Keywords: llm,chat,huggingface,bielik,polish,ai,assistant
Classifier: Development Status :: 4 - Beta
Classifier: Intended Audience :: Developers
Classifier: Intended Audience :: End Users/Desktop
Classifier: Programming Language :: Python :: 3
Classifier: Programming Language :: Python :: 3.9
Classifier: Programming Language :: Python :: 3.10
Classifier: Programming Language :: Python :: 3.11
Classifier: Topic :: Software Development :: Libraries
Classifier: Topic :: Communications :: Chat
Classifier: Topic :: Scientific/Engineering :: Artificial Intelligence
Classifier: Operating System :: POSIX :: Linux
Classifier: Operating System :: MacOS :: MacOS X
Classifier: Operating System :: Microsoft :: Windows
Classifier: Natural Language :: Polish
Classifier: Natural Language :: English
Requires-Python: >=3.9
Description-Content-Type: text/markdown
License-File: LICENSE
Requires-Dist: typer<1.0.0,>=0.9.0
Requires-Dist: rich<14.0.0,>=13.0.0
Requires-Dist: python-dotenv<2.0.0,>=1.0.0
Requires-Dist: pypdf<4.0.0,>=3.0.0
Requires-Dist: python-docx<1.0.0,>=0.8.11
Requires-Dist: beautifulsoup4<5.0.0,>=4.8.0
Requires-Dist: html2text<2025.0.0,>=2020.1.16
Requires-Dist: huggingface-hub<1.0.0,>=0.16.0
Requires-Dist: tokenizers<1.0.0,>=0.13.0
Requires-Dist: sentencepiece<1.0.0,>=0.1.99
Requires-Dist: protobuf<5.0.0,>=4.21.0
Provides-Extra: local
Requires-Dist: llama-cpp-python<1.0.0,>=0.2.0; extra == "local"
Requires-Dist: sentence-transformers<3.0.0,>=2.2.0; extra == "local"
Provides-Extra: gpu
Requires-Dist: bielik[local]; extra == "gpu"
Requires-Dist: torch<3.0.0,>=2.0.0; extra == "gpu"
Requires-Dist: transformers<5.0.0,>=4.30.0; extra == "gpu"
Requires-Dist: accelerate<1.0.0,>=0.20.0; extra == "gpu"
Requires-Dist: bitsandbytes<1.0.0,>=0.40.0; extra == "gpu"
Provides-Extra: ai
Requires-Dist: bielik[gpu]; extra == "ai"
Requires-Dist: transformers[torch]<5.0.0,>=4.30.0; extra == "ai"
Requires-Dist: datasets<3.0.0,>=2.12.0; extra == "ai"
Requires-Dist: evaluate<1.0.0,>=0.4.0; extra == "ai"
Provides-Extra: dev
Requires-Dist: bielik[ai]; extra == "dev"
Requires-Dist: pytest<8.0.0,>=7.0.0; extra == "dev"
Requires-Dist: pytest-cov<5.0.0,>=4.0.0; extra == "dev"
Requires-Dist: flake8<7.0.0,>=6.0.0; extra == "dev"
Requires-Dist: black<24.0.0,>=23.0.0; extra == "dev"
Requires-Dist: isort<6.0.0,>=5.12.0; extra == "dev"
Requires-Dist: mypy<2.0.0,>=1.0.0; extra == "dev"
Requires-Dist: build<1.0.0,>=0.10.0; extra == "dev"
Requires-Dist: twine<5.0.0,>=4.0.0; extra == "dev"
Requires-Dist: pre-commit<4.0.0,>=3.0.0; extra == "dev"
Provides-Extra: docs
Requires-Dist: mkdocs<2.0.0,>=1.4.0; extra == "docs"
Requires-Dist: mkdocs-material<10.0.0,>=9.0.0; extra == "docs"
Requires-Dist: mkdocstrings[python]<1.0.0,>=0.20.0; extra == "docs"
Provides-Extra: test
Requires-Dist: pytest<8.0.0,>=7.0.0; extra == "test"
Requires-Dist: pytest-cov<5.0.0,>=4.0.0; extra == "test"
Requires-Dist: pytest-mock<4.0.0,>=3.10.0; extra == "test"
Requires-Dist: pytest-asyncio<1.0.0,>=0.20.0; extra == "test"
Provides-Extra: completion
Requires-Dist: click-completion<1.0.0,>=0.5.0; extra == "completion"
Requires-Dist: shellingham<2.0.0,>=1.5.0; extra == "completion"
Provides-Extra: web
Requires-Dist: fastapi<1.0.0,>=0.95.0; extra == "web"
Requires-Dist: uvicorn[standard]<1.0.0,>=0.21.0; extra == "web"
Requires-Dist: jinja2<4.0.0,>=3.0.0; extra == "web"
Requires-Dist: python-multipart<1.0.0,>=0.0.6; extra == "web"
Provides-Extra: all
Requires-Dist: bielik[ai,completion,dev,docs,test,web]; extra == "all"
Dynamic: license-file

# 🦅 Bielik


**Bielik** is a groundbreaking Polish language model created by **[Speakleash](https://speakleash.org/)** - a foundation dedicated to the development of Polish artificial intelligence.
> 🇵🇱 **Bielik** is available on **[huggingface](https://huggingface.co/speakleash)**.
 
Now you can test it directly in the shell, using several features that allow you to build a multi-agent environment in your company 🚀 

- 🎯 **HuggingFace Integration** - Direct model downloads from HF Hub
- 💬 **Polish Language Optimized** - Built for Polish conversation and analysis  
- 🖼️ **Vision Capabilities** - Image analysis and visual question answering
- 📁 **Document Processing** - PDF, DOCX, web content analysis
- 🐳 **Docker Ready** - Containerized testing environments
- ⚡ **Lightweight** - Minimal (~50MB) or Full (~2GB) installation options

**Author:** Tom Sapletta  
**License:** Apache-2.0

[![PyPI](https://img.shields.io/pypi/v/bielik.svg)](https://pypi.org/project/bielik/)
[![Python](https://img.shields.io/pypi/pyversions/bielik.svg)](https://www.python.org/)
[![License](https://img.shields.io/badge/license-Apache%202.0-blue.svg)](LICENSE)
[![Build](https://img.shields.io/github/actions/workflow/status/tom-sapletta-com/bielik/python-app.yml?branch=main)](https://github.com/tom-sapletta-com/bielik/actions)
[![Downloads](https://img.shields.io/pypi/dm/bielik.svg)](https://pypi.org/project/bielik/)
[![Code style: flake8](https://img.shields.io/badge/code%20style-flake8-black.svg)](https://flake8.pycqa.org/)
[![Issues](https://img.shields.io/github/issues/tom-sapletta-com/bielik.svg)](https://github.com/tom-sapletta-com/bielik/issues)
[![Stars](https://img.shields.io/github/stars/tom-sapletta-com/bielik.svg)](https://github.com/tom-sapletta-com/bielik/stargazers)
[![Forks](https://img.shields.io/github/forks/tom-sapletta-com/bielik.svg)](https://github.com/tom-sapletta-com/bielik/network)


## 📋 **Navigation Menu**
- [🏗️ Architecture](#️-architecture)
- [⚡ Quick Start](#-quick-start)
- [🛠️ Installation](#️-installation)
- [🎯 Context Provider Commands](#-context-provider-commands)
- [💬 Usage Examples](#-usage-examples)
- [🐳 Docker Testing Framework](#-docker-testing-framework)
- [📚 Documentation](#-documentation)
- [🤝 Contributing](#-contributing)

---

## 🏗️ Architecture

```
┌─────────────────────────────────────────────────────────────────┐
│                       🦅 BIELIK SYSTEM                          │
├─────────────────────┬───────────────────────┬───────────────────┤
│   🖥️  CLI Shell     │  🌐 FastAPI Server   │  🐳 Docker Tests   │
│  ┌─────────────────┐│ ┌───────────────────┐ │ ┌───────────────┐ │
│  │ • Interactive   ││ │ • REST /chat      │ │ │ • Minimal     │ │
│  │ • Personalized  ││ │ • WebSocket /ws   │ │ │ • Full        │ │
│  │ • Multi-modal   ││ │ • Port 8000       │ │ │ • CI/CD       │ │
│  └─────────────────┘│ └───────────────────┘ │ └───────────────┘ │
└─────────────────────┴───────────────────────┴───────────────────┘
            │                       │                       │
            ▼                       ▼                       ▼
┌───────────────────────────────────────────────────────────────┐
│                🤗 HUGGINGFACE INTEGRATION                     │
│  ┌──────────────────────┐    ┌─────────────────────────────┐  │
│  │   Direct Downloads   │◄──►│     Local Model Execution   │  │
│  │ ┌─ HF Hub API        │    │ ┌─ Transformers Pipeline    │  │
│  │ └─ Model Management  │    │ └─ Vision Models (optional) │  │
│  └──────────────────────┘    └─────────────────────────────┘  │
└─────────────────────┬─────────────────────────────────────────┘
                      ▼
┌───────────────────────────────────────────────────────────────┐
│                 📚 POLISH LANGUAGE MODELS                     │
│  ┌─────────────────────────────────────────────────────────┐  │
│  │ 🤖 Speakleash/Bielik Models (HuggingFace Hub)           │  │
│  │ 🔗 Direct: HuggingFace → Local Storage → Execution      │  │
│  │ 🎯 Polish-optimized conversation and analysis           │  │
│  └─────────────────────────────────────────────────────────┘  │
└───────────────────────────────────────────────────────────────┘
```

![terminal](terminal.png)

## 🤖 About Bielik Model

**Bielik** is a groundbreaking Polish language model created by **[Speakleash](https://speakleash.org/)** - a foundation dedicated to the development of Polish artificial intelligence.

### 🔗 External Dependencies & Links:
- **[HuggingFace Hub](https://huggingface.co)** - Primary source for Polish language models
- **[Bielik Models on HuggingFace](https://huggingface.co/speakleash)** - Official model repository
- **[Speakleash Foundation](https://speakleash.org/)** - Creators of the Bielik models
- **[Polish AI Initiative](https://www.gov.pl/web/ai)** - Government support for Polish AI

### 🚀 How it Works:
1. **Bielik CLI** connects directly to **HuggingFace Hub**
2. **Models** are downloaded from **Speakleash** organization on HuggingFace
3. **Local execution** uses **Transformers** library (optional for vision)
4. **Chat interface** (CLI/Web) → **Local Models** → **Polish responses**
5. **Modular design** supports text-only or full vision capabilities

---

## 📌 Features

- **🎯 HuggingFace Integration** — Direct model downloads and management
  - **🔍 Model Discovery** — Browse available Polish language models
  - **📦 Smart Downloads** — Automatic model caching and versioning
  - **🚀 Auto-Switch** — Automatically switches to newly downloaded models
  - **🛠️ Interactive** — User-friendly model selection on first startup
- **🖥️ Enhanced CLI** `python -m bielik` — Personalized chat experience
  - **📋 Commands** — `:help`, `:models`, `:download`, `:delete`, `:switch`, `:settings`, `:name`
  - **⚙️ Personalization** — Custom user names, dynamic assistant names
  - **🤗 HF Management** — Direct HuggingFace model operations
  - **📁 Content Analysis** — Folder scanning, document processing
  - **🌍 Cross-platform** — Windows, macOS, Linux support
- **🖼️ Vision Capabilities** (Full version only)
  - **🔍 Image Analysis** — Automatic image captioning in Polish
  - **❓ Visual QA** — Ask questions about images
  - **🎨 Multi-modal** — Combined text and image understanding
  - **⚡ GPU Support** — Hardware acceleration for faster processing
- **🐍 Python API** — Programmatic access via `BielikClient` class
  - **💬 Chat methods** — `chat()`, `query()`, conversation management
  - **🔧 Model control** — Download, switch, and manage models
  - **🤗 HF Models** — Full HuggingFace integration
  - **📤 Export** — Conversation history in multiple formats
- **🌐 Web Server** (FastAPI on port 8000):  
  - **📡 REST** — `POST /chat` endpoint for JSON communication
  - **⚡ WebSocket** — `WS /ws` for real-time chat
  - **🖼️ Multi-modal** — Support for text and image inputs
- **🐳 Docker Support** — Complete containerized testing
  - **📦 Minimal Version** — Lightweight text-only container (~50MB)
  - **🎯 Full Version** — Complete vision-enabled container (~2GB)
  - **🔧 CI/CD** — Automated testing environments  

---

## ⚡ Performance Optimized Installation

Bielik now features **lazy loading** and **optimized startup** for faster performance. We recommend using **Conda** for the most reliable installation experience.

# 🚀 Installation Guide

## 🐍 **Conda Installation (Recommended)**

### Prerequisites
- Install [Miniconda](https://docs.conda.io/en/latest/miniconda.html) or [Anaconda](https://www.anaconda.com/download/)
- For GPU support: Install appropriate NVIDIA drivers and CUDA toolkit

### 1. Create and Activate Environment

```bash
# Create a new conda environment
conda create -n bielik python=3.11 -y
conda activate bielik
```

### 2. Install System Dependencies

```bash
conda install -c conda-forge -y \
    cmake \
    make \
    gcc_linux-64 \
    gxx_linux-64 \
    libgcc \
    libstdcxx-ng
```

### 3. Install PyTorch (CPU or GPU)

For CPU-only:
```bash
conda install -c pytorch -y \
    pytorch \
    torchvision \
    torchaudio \
    cpuonly
```

For NVIDIA GPU (CUDA 11.8):
```bash
conda install -c pytorch -y \
    pytorch \
    torchvision \
    torchaudio \
    pytorch-cuda=11.8 \
    -c nvidia
```

### 4. Install Bielik and Dependencies

```bash
# Install Hugging Face libraries
conda install -c huggingface -y \
    transformers \
    accelerate \
    sentencepiece \
    huggingface-hub \
    bitsandbytes

# Install llama-cpp-python via conda (recommended - avoids C++ compilation issues)
conda install -c conda-forge llama-cpp-python

# Install Bielik in development mode
git clone https://github.com/tom-sapletta-com/bielik.git
cd bielik
conda develop .
```

### 5. Verify Installation

```bash
python scripts/verify_installation.py
```

## ⚡ Optimizing Model Loading

Bielik now includes optimized model loading with a 10-second timeout and debug mode:

```bash
# Start Bielik with debug mode (shows detailed loading information)
BIELIK_DEBUG=1 bielik

# Set a custom timeout (in seconds)
BIELIK_LOAD_TIMEOUT=15 bielik
```

If a model takes longer than the timeout to load, the debugger will automatically activate and show detailed logs to help diagnose the issue.

## 🖥️ Platform-Specific Notes

### Windows
```bash
# Use Anaconda Prompt with Administrator privileges
# Install Visual Studio Build Tools with C++ workload
```

### macOS (Intel/Apple Silicon)
```bash
# For Apple Silicon with Metal acceleration:
conda install -c conda-forge llama-cpp-python

# Or use pip with Metal flags (if conda not available):
# CMAKE_ARGS="-DLLAMA_METAL=on" FORCE_CMAKE=1 \
# pip install llama-cpp-python --no-cache-dir
```

### Linux
```bash
# Install system dependencies
sudo apt-get update && sudo apt-get install -y \
    build-essential \
    libssl-dev \
    zlib1g-dev \
    libbz2-dev \
    libreadline-dev \
    libsqlite3-dev \
    wget \
    llvm \
    libncurses5-dev \
    libncursesw5-dev \
    xz-utils \
    tk-dev \
    libffi-dev \
    liblzma-dev \
    python-openssl \
    git
```

## 🛠️ Troubleshooting

### Model Loading Issues
If models are loading slowly or timing out:
1. Check your internet connection
2. Verify you have enough disk space (~10GB free recommended)
3. Enable debug mode: `BIELIK_DEBUG=1 bielik`
4. Try clearing the model cache: `rm -rf ~/.cache/huggingface/hub`

### GPU Acceleration
To enable GPU acceleration, ensure you have:
1. Latest NVIDIA drivers installed
2. CUDA toolkit matching your PyTorch version
3. Properly configured `LD_LIBRARY_PATH`

### Common Errors
- **CUDA out of memory**: Reduce batch size or use a smaller model
- **Missing libraries**: Install required system dependencies
- **Permission errors**: Run with appropriate permissions or use `--user` flag

## 🛠 **Manual Installation**

```bash
# 1. Clone repository
git clone https://github.com/tom-sapletta-com/bielik.git
cd bielik

# 2. Run universal installer
python install.py
```

## 🔧 Troubleshooting

### Common Issues

#### 1. Conda Environment Issues

**Problem**: Conda command not found  
**Solution**:
```bash
# Add Conda to PATH
export PATH="$HOME/miniconda3/bin:$PATH"
# Initialize Conda for your shell
eval "$(conda shell.bash hook)"
```

#### 2. CUDA/GPU Support

**Problem**: CUDA not detected  
**Solution**: Install CUDA toolkit and configure Conda:
```bash
conda install -c nvidia cuda-toolkit
conda install -c conda-forge cudatoolkit-dev
```

#### 3. Slow Model Loading

**Problem**: First load is slow  
**Solution**: This is normal for the first load. Subsequent loads will be faster thanks to lazy loading.

#### 4. Missing Dependencies

**Problem**: Missing system libraries  
**Solution**: Install required system packages:
```bash
# Ubuntu/Debian
sudo apt-get update && sudo apt-get install -y \
    build-essential \
    cmake \
    libopenblas-dev \
    liblapack-dev \
    libjpeg-dev

# CentOS/RHEL
sudo yum groupinstall -y "Development Tools"
sudo yum install -y cmake openblas-devel lapack-devel libjpeg-turbo-devel
```

### Verifying Your Installation

Run the verification script to check for common issues:

```bash
python scripts/verify_installation.py
```

### Getting Help

If you encounter any issues, please:
1. Check the [GitHub Issues](https://github.com/tom-sapletta-com/bielik/issues) for known problems
2. Run the verification script and include its output when reporting issues
3. Provide details about your system and the exact error message

## 🚀 Performance Tips

1. **First Run**: The first run will be slower as models are loaded and cached
2. **Subsequent Runs**: Enjoy faster startup times thanks to lazy loading
3. **Memory Usage**: Close other memory-intensive applications when running Bielik
4. **GPU Acceleration**: For best performance, use a system with CUDA-compatible GPU

### 🪟 **Windows Users**
```batch
REM Double-click install.bat or run:
install.bat

REM Launch Bielik:
run.bat
```

### 🐧 **Linux/macOS Users**
```bash
# Run installer:
./install.sh

# Launch Bielik:
./run.sh
```

### ⚙️ **Installation Options**
```bash
# Basic installation (recommended)
python install.py

# Skip AI models (fastest installation)
python install.py --skip-ai

# Use conda/mamba instead of pip
python install.py --conda

# Development installation
python install.py --dev

# Show all options
python install.py --help
```

### 🎯 **What the Universal Installer Does:**
- ✅ **Auto-detects** your operating system and Python version
- ✅ **Creates** isolated virtual environment 
- ✅ **Installs** all dependencies with smart fallback strategies
- ✅ **Attempts** multiple llama-cpp-python installation methods
- ✅ **Creates** platform-specific launcher scripts
- ✅ **Works** even if AI models fail to install (Context Provider Commands still work)
- ✅ **Provides** clear next steps and troubleshooting

### 🚀 **After Installation**
```bash
# Universal launcher (any platform)
python run.py

# Platform-specific launchers
./run.sh        # Linux/macOS
run.bat         # Windows

# Or direct activation
.venv/bin/python -m bielik.cli.main     # Linux/macOS
.venv\Scripts\python -m bielik.cli.main # Windows
```

---

## 📦 **Alternative: PyPI Installation** 

For users who prefer traditional pip installation:

### 🪶 **Minimal Version** (Text-only)

```bash
# Install minimal version (text-only)
pip install bielik

# Start CLI and download your first model
python -m bielik
```

**What's included:**
- ✅ Polish conversation and text analysis
- ✅ HuggingFace model downloads and management  
- ✅ Document processing (PDF, DOCX, TXT)
- ✅ Web content analysis
- ✅ Context Provider Commands (folder:, calc:, pdf:)
- ✅ Personalized CLI experience
- ❌ Local AI models (requires manual llama-cpp-python installation)

### 🎯 **Full Version** (With vision)

```bash
# Install full version (text + vision)
pip install bielik[vision]

# Start CLI with image analysis support
python -m bielik
```

**What's included:**
- ✅ **Everything from minimal version**
- ✅ **Image analysis and captioning**
- ✅ **Visual question answering**
- ✅ **GPU acceleration support**
- ✅ **Multi-modal document processing**

### 🔄 **Upgrade Options**

```bash
# Upgrade minimal → full
pip install bielik[vision]

# Or install specific optional features
pip install bielik[local]    # Local model execution
pip install bielik[gpu]      # GPU acceleration
pip install bielik[dev]      # Development tools
pip install bielik[vision]   # Vision capabilities
```

---

## 🚀 Quick Start Guide

### 🎯 **Instant Start** (No setup required!)

Bielik now works **without any external dependencies**. Just install and start chatting:

```bash
# Install minimal version
pip install bielik

# Start CLI and choose your first model
python -m bielik
```

**What happens on first run:**
- 🔍 **Model Selection** — Choose from available Polish models
- 📥 **Auto Download** — Selected model downloads from HuggingFace
- 🔄 **Auto Switch** — Automatically switches to the new model
- 💬 **Ready to Chat** — Start conversing in Polish immediately!

### 📱 **Choose Your Experience**

#### 🪶 **Minimal Setup** (Recommended)
Perfect for text conversations and document analysis:
```bash
# 1. Install
pip install bielik

# 2. Start and download your first model
python -m bielik
:download speakleash/bielik-4.5b-v3.0-instruct

# 3. Start chatting in Polish!
Cześć! Jak mogę Ci pomóc?
```

#### 🎯 **Full Setup** (For image analysis)
Includes vision capabilities for image analysis:
```bash
# 1. Install with vision support
pip install bielik[vision]

# 2. Start and download models
python -m bielik
:download speakleash/bielik-4.5b-v3.0-instruct

# 3. Analyze images
Przeanalizuj to zdjęcie: image.jpg
```

#### 🐳 **Docker Setup** (For testing)
Use Docker for isolated testing environments:
```bash
# Clone repository
git clone https://github.com/tom-sapletta-com/bielik.git
cd bielik

# Test minimal version
docker-compose --profile minimal up bielik-minimal

# Test full version  
docker-compose --profile full up bielik-full
```

---

## 🎯 Context Provider Commands

Bielik CLI features a standardized **Context Provider Command** system that allows you to generate structured data for AI analysis. All external commands use the format `name:` (with colon after the command name).

### 📋 **Available Commands**

| Command | Format | Description | Example |
|---------|--------|-------------|---------|
| **📁 Folder Analysis** | `folder: <path>` | Directory scanning and file analysis | `folder: ~/Documents` |
| **🧮 Calculator** | `calc: <expression>` | Mathematical calculations and evaluations | `calc: 2 + 3 * 4` |
| **📄 Document Reader** | `pdf: <file>` | Text extraction from PDF, DOCX, TXT files | `pdf: report.pdf` |

### 💡 **How Context Provider Commands Work**

1. **Generate Context**: Commands analyze input and create structured data
2. **AI Integration**: Data is formatted for AI consumption and analysis
3. **Independent Operation**: Commands work without AI models installed
4. **Project Integration**: Artifacts are automatically saved to your active project
4. **Standardized Format**: All external commands use `name:` format for consistency

### 🚀 **Usage Examples**

```bash
# Directory Analysis
bielik -p "folder: ."
bielik -p "folder: ~/Documents --recursive"

# Mathematical Calculations
bielik -p "calc: 2 + 3 * 4"
bielik -p "calc: sqrt(16) + sin(pi/2)"
bielik -p "calc: factorial(5)"

# Document Processing
bielik -p "pdf: document.pdf"
bielik -p "pdf: report.docx --metadata"
bielik -p "pdf: file.pdf --pages=1-5"
```

### ⚡ **Interactive Usage**

In interactive mode, Context Provider Commands load data for AI analysis:

```bash
python -m bielik

# Load directory context
> folder: ~/projects

✅ Context loaded! You can now ask questions about the provided data.

# Ask AI about the loaded context
> What files are in this directory?
> Which file is the largest?
> Are there any Python files?
```

---

## 🚀 **Project Management System**

Bielik features a comprehensive **session-based project management system** that organizes your analysis artifacts into beautiful, shareable HTML projects.

### 🎯 **Key Features**

- **🗂️ Multi-Project Sessions** - Work on multiple projects simultaneously
- **📄 HTML Artifact Generation** - Beautiful, interactive project representations
- **🔄 Automatic Artifact Collection** - Context Provider Commands auto-save to active project
- **🌐 Browser-Friendly Viewing** - Projects open in your default browser
- **📊 Rich Metadata** - Embedded project information and artifact tracking
- **✅ Validation System** - Comprehensive integrity checking

### 📁 **Project Management Commands**

| Command | Description | Example |
|---------|-------------|---------|
| `:project create <name>` | Create new project | `:project create "Data Analysis"` |
| `:project create <name> <desc> --tags <tags>` | Create with metadata | `:project create "ML Project" "Machine learning analysis" --tags ml,data` |
| `:project switch <id\|name>` | Switch to project | `:project switch data-analysis` |
| `:project list` | List all projects | `:project list` |
| `:project info [id]` | Show project details | `:project info` |
| `:project open [id]` | Open in browser | `:project open` |
| `:project validate [id]` | Validate project | `:project validate` |

### 🎨 **Project Workflow**

```bash
# 1. Create a new project
python -m bielik
> :project create "Website Analysis" "Analyzing project structure"

✅ Project Created Successfully
🎯 Project is now active. Use Context Provider Commands to add artifacts.

# 2. Add artifacts using Context Provider Commands
> folder: ~/my-website
🎯 Artifact Added to Project: Website Analysis

> calc: file_count * average_size / 1024
🎯 Artifact Added to Project: Website Analysis

> pdf: documentation.pdf  
🎯 Artifact Added to Project: Website Analysis

# 3. View your project
> :project open
🌐 Project Opened in Browser

# 4. Switch between projects
> :project list
📋 Bielik Projects (Current Session)

🎯 **ACTIVE** Website Analysis
   🆔 ID: abc123ef...
   📊 Artifacts: 3
   📅 Created: 2024-12-20 14:30

📁 Data Analysis
   🆔 ID: def456gh...
   📊 Artifacts: 5  
   📅 Created: 2024-12-20 12:15

> :project switch def456gh
✅ Switched to Project: Data Analysis
```

### 🏗️ **HTML Project Structure**

Each project generates a beautiful HTML file with:

- **📊 Interactive Dashboard** - Project overview with metadata
- **🎨 Beautiful Design** - Modern, responsive interface
- **📑 Artifact Viewer** - Organized display of all artifacts
- **🔍 Rich Metadata** - Embedded in HTML attributes for programmatic access
- **🌍 Offline Capability** - Works without internet connection

**Example Project Structure:**
```
./bielik_projects/
├── abc123ef-ghij-klmn-opqr-stuvwxyz0123/
│   ├── index.html          # Interactive project page
│   └── metadata.json       # Project metadata
└── def456gh-ijkl-mnop-qrst-uvwxyz012345/
    ├── index.html
    └── metadata.json
```

### 🔍 **Validation System**

Bielik includes comprehensive validators for:

- **🏠 HTML Artifacts** - Structure, metadata, and compliance checking
- **⚙️ Environment Files** - .env configuration validation  
- **📜 Command Scripts** - Code quality and compliance verification

```bash
# Validate current project
> :project validate
🔍 Project Validation Results
📊 Status: ✅ VALID (0 errors, 0 warnings)

# Validate specific project
> :project validate abc123ef
```

---

## 💻 Usage

### 🖥️ CLI Commands & Options

#### Starting Bielik

```bash
# Basic usage
python -m bielik                         # Start interactive chat
bielik                                   # Alternative (if in PATH)

# Advanced options  
python -m bielik --help                  # Show all options
```

#### Interactive Commands (inside chat session)

**📋 Model Management:**
```bash
:models                    # List available HuggingFace models
:download <model-name>     # Download model from HuggingFace
:switch <model-name>       # Switch to downloaded model
:delete <model-name>       # Delete model from local storage
```

**⚙️ Personalization:**
```bash
:name <your-name>          # Set your display name
:settings                  # Show current configuration
```

**🛠️ Utilities:**
```bash
:help                      # Show all commands
:clear                     # Clear conversation history
:exit                      # Quit (or Ctrl+C)
```

#### Usage Examples

**First Time Setup:**
```bash
$ python -m bielik
# Choose from recommended models:
# 1. speakleash/bielik-4.5b-v3.0-instruct (Recommended)
# 2. speakleash/bielik-7b-instruct-v0.1
# Enter choice: 1

🔄 Downloading speakleash/bielik-4.5b-v3.0-instruct...
✅ Model ready! Switching to bielik-4.5b-v3.0-instruct

👤 You: Cześć! Jak się masz?
🤖 bielik-4.5b: Cześć! Mam się dobrze, dziękuję...
```

**Everyday Usage:**
```bash
👤 You: :name Jan
✅ Display name set to: Jan

👤 Jan: Przeanalizuj folder ~/dokumenty
🤖 bielik-4.5b: [Analyzes folder structure and contents]

👤 Jan: Opisz to zdjęcie: vacation.jpg  # (Full version only)
🤖 bielik-4.5b: [Describes image in Polish]
```

### 🐍 Python API

Use Bielik programmatically in your Python applications:

```python
from bielik.client import BielikClient

# Create client (auto-downloads model if needed)
client = BielikClient()

# Send a Polish message
response = client.chat("Napisz krótki wiersz o Polsce")
print(response)

# Get model status
status = client.get_status()
print(f"Current model: {status['current_model']}")
print(f"Models available: {status['models_available']}")

# Export conversation
history = client.export_conversation(format="markdown")
```

**Quick functions:**
```python
from bielik.client import quick_chat

# One-off Polish query
response = quick_chat("Co to jest sztuczna inteligencja?")
print(response)
```

**BielikClient Options:**
- `model`: Specific HuggingFace model to use
- `auto_download`: Auto-download model if missing (default: True)

### 🌐 Web Server

```bash
# Start web server
uvicorn bielik.server:app --port 8000

# Or with Docker
docker-compose --profile minimal up bielik-minimal
```

**Endpoints:**
- `POST /chat` - JSON chat endpoint
- `WS /ws` - WebSocket real-time chat
- `GET /models` - List available models

**Example request:**
```json
{"messages": [{"role":"user","content":"Cześć! Jak się masz?"}]}
```

### 🐳 Docker Usage

**Quick Testing:**
```bash
# Test minimal version
docker run -it bielik:minimal

# Test full version with GPU
docker run --gpus all -it bielik:full
```

**With persistent storage:**
```bash
# Minimal with model persistence
docker run -it -v $(pwd)/models:/app/models bielik:minimal

# Full with models and images
docker run -it \
  -v $(pwd)/models:/app/models \
  -v $(pwd)/images:/app/images \
  bielik:full
```

**Development setup:**
```bash
# Clone and test
git clone https://github.com/tom-sapletta-com/bielik.git
cd bielik

# Run automated tests
docker-compose --profile test up test-runner

# Interactive development
docker-compose --profile minimal run --rm bielik-minimal bash
```

---

## 🔧 Environment Variables

**Core Settings:**
* `BIELIK_CLI_USERNAME` — Your display name in CLI (auto-detected from system)
* `BIELIK_CLI_CURRENT_MODEL` — Currently selected model
* `BIELIK_CLI_ASSISTANT_NAME` — Assistant display name (auto-set from model)
* `BIELIK_CLI_AUTO_SWITCH` — Auto-switch to newly downloaded models (default: true)

**Storage & Cache:**
* `BIELIK_MODELS_DIR` — Local model storage directory
* `BIELIK_DATA_DIR` — User data and settings directory
* `HF_HOME` — HuggingFace cache directory

**Docker Environment:**
* `BIELIK_MODE` — `minimal` or `full` (Docker only)
* `BIELIK_IMAGES_DIR` — Images directory for analysis (full version)

---

## 🛠️ Troubleshooting

### Installation Issues

**Problem:** Minimal version works but vision features don't
```bash
# Upgrade to full version
pip install bielik[vision]

# Verify vision packages
python -c "import PIL, transformers; print('Vision packages OK')"
```

**Problem:** Model download fails or times out
```bash
# Check HuggingFace connectivity
python -c "from huggingface_hub import HfApi; print('HF connection OK')"

# Check available disk space (models are 2-8GB)
df -h

# Manual download with timeout
:download speakleash/bielik-4.5b-v3.0-instruct
```

**Problem:** "No models available" on first startup
```bash
# Download a model manually
python -m bielik
:models
:download speakleash/bielik-4.5b-v3.0-instruct
```

### Runtime Issues

**Problem:** Model responses are slow or use too much memory
```bash
# Use smaller model
:switch speakleash/bielik-4.5b-v3.0-instruct  # instead of 7b

# Check system resources
htop      # Linux/macOS
taskmgr   # Windows

# Enable GPU acceleration (full version)
pip install bielik[gpu]
```

**Problem:** Image analysis not working
```bash
# Check if vision packages installed
python -c "from bielik.image_analyzer import ImageAnalyzer; ia = ImageAnalyzer(); print(f'Available: {ia.is_available()}')"

# Install vision support
pip install bielik[vision]
```

**Problem:** CLI settings not persisting
```bash
# Check .env file creation
ls -la ~/.bielik/ || ls -la ./

# Reset settings
:name YourName
:settings
```

### Docker Issues

**Problem:** Docker containers fail to start
```bash
# Build images manually
docker build -f docker/Dockerfile.minimal -t bielik:minimal .
docker build -f docker/Dockerfile.full -t bielik:full .

# Check container logs
docker logs bielik-minimal
```

**Problem:** Models not persisting between Docker runs
```bash
# Use volume mounts
docker run -v $(pwd)/models:/app/models bielik:minimal

# Or use Docker Compose
docker-compose --profile minimal up
```

### Getting Help

- **GitHub Issues:** [Report bugs and feature requests](https://github.com/tom-sapletta-com/bielik/issues)
- **Command Help:** `python -m bielik --help` or `:help` in CLI
- **Test Environment:** Use Docker for isolated testing
- **Check Status:** Use `:settings` command for current configuration

---

## 🛠️ **Creating Custom Context Provider Commands**

Bielik CLI supports **extensible Context Provider Commands** that allow you to create custom `name:` commands that generate structured context data for AI analysis.

### 📋 **Quick Setup Guide**

**1. Create command directory:**
```bash
mkdir -p commands/mycommand
cd commands/mycommand
```

**2. Create your command files:**
- `main.py` - Your command implementation
- `config.json` - Command configuration and metadata

### 🐍 **Python Code Examples**

#### **Basic Context Provider Command Template**

```python
# commands/mycommand/main.py
from typing import List, Dict, Any
from bielik.cli.command_api import ContextProviderCommand

class MyCommand(ContextProviderCommand):
    def __init__(self):
        super().__init__(
            name="mycommand",
            description="Custom context provider that analyzes something",
            usage="mycommand: <path_or_input>"
        )
    
    def provide_context(self, args: List[str], context: Dict[str, Any]) -> Dict[str, Any]:
        """
        Generate context data from the provided arguments.
        
        Args:
            args: List of arguments after 'mycommand:'
            context: Additional context data
            
        Returns:
            Dict containing structured context data
        """
        if not args:
            return {
                "error": "Please provide input after 'mycommand:'",
                "usage": self.get_usage()
            }
        
        input_data = " ".join(args)
        
        # Your custom logic here
        analysis_result = self._analyze_input(input_data)
        
        return {
            "command": "mycommand",
            "input": input_data,
            "analysis": analysis_result,
            "timestamp": context.get("timestamp", ""),
            "context_type": "custom_analysis"
        }
    
    def _analyze_input(self, input_data: str) -> Dict[str, Any]:
        """Custom analysis logic - implement your functionality here."""
        return {
            "input_length": len(input_data),
            "word_count": len(input_data.split()),
            "type": "text_analysis",
            "summary": f"Analyzed text with {len(input_data)} characters"
        }

# Required: Command instance for the registry
command = MyCommand()
```

#### **Advanced Example: Git Repository Analyzer**

```python
# commands/git/main.py
import os
import subprocess
from typing import List, Dict, Any
from bielik.cli.command_api import ContextProviderCommand

class GitCommand(ContextProviderCommand):
    def __init__(self):
        super().__init__(
            name="git",
            description="Analyzes Git repository information and recent commits",
            usage="git: [path_to_repo]"
        )
    
    def provide_context(self, args: List[str], context: Dict[str, Any]) -> Dict[str, Any]:
        repo_path = args[0] if args else "."
        
        if not self._is_git_repo(repo_path):
            return {
                "error": f"Not a git repository: {repo_path}",
                "suggestion": "Run this command in a git repository directory"
            }
        
        return {
            "command": "git",
            "repository_path": os.path.abspath(repo_path),
            "branch_info": self._get_branch_info(repo_path),
            "recent_commits": self._get_recent_commits(repo_path),
            "status": self._get_status(repo_path),
            "remote_info": self._get_remote_info(repo_path),
            "context_type": "git_analysis"
        }
    
    def _is_git_repo(self, path: str) -> bool:
        """Check if directory is a git repository."""
        try:
            result = subprocess.run(
                ["git", "rev-parse", "--git-dir"],
                cwd=path, capture_output=True, text=True
            )
            return result.returncode == 0
        except:
            return False
    
    def _get_branch_info(self, path: str) -> Dict[str, str]:
        """Get current branch information."""
        try:
            branch_result = subprocess.run(
                ["git", "branch", "--show-current"],
                cwd=path, capture_output=True, text=True
            )
            return {
                "current_branch": branch_result.stdout.strip(),
                "status": "active"
            }
        except:
            return {"error": "Could not determine branch info"}
    
    def _get_recent_commits(self, path: str, limit: int = 5) -> List[Dict[str, str]]:
        """Get recent commit information."""
        try:
            result = subprocess.run([
                "git", "log", "--oneline", f"-{limit}",
                "--pretty=format:%h|%an|%ar|%s"
            ], cwd=path, capture_output=True, text=True)
            
            commits = []
            for line in result.stdout.strip().split('\n'):
                if '|' in line:
                    hash_id, author, date, message = line.split('|', 3)
                    commits.append({
                        "hash": hash_id,
                        "author": author,
                        "date": date,
                        "message": message
                    })
            return commits
        except:
            return [{"error": "Could not fetch commit history"}]
    
    def _get_status(self, path: str) -> Dict[str, Any]:
        """Get git status information."""
        try:
            result = subprocess.run(
                ["git", "status", "--porcelain"],
                cwd=path, capture_output=True, text=True
            )
            
            modified_files = []
            for line in result.stdout.strip().split('\n'):
                if line.strip():
                    status, filename = line[:2], line[3:]
                    modified_files.append({
                        "file": filename,
                        "status": status.strip()
                    })
            
            return {
                "modified_files": modified_files,
                "clean": len(modified_files) == 0
            }
        except:
            return {"error": "Could not get status"}
    
    def _get_remote_info(self, path: str) -> Dict[str, str]:
        """Get remote repository information."""
        try:
            result = subprocess.run(
                ["git", "remote", "get-url", "origin"],
                cwd=path, capture_output=True, text=True
            )
            return {
                "origin_url": result.stdout.strip(),
                "has_remote": True
            }
        except:
            return {"has_remote": False}

# Required: Command instance for the registry
command = GitCommand()
```

### 📄 **Configuration File (config.json)**

```json
{
  "name": "mycommand",
  "description": "Custom context provider for specialized analysis",
  "version": "1.0.0",
  "author": "Your Name",
  "category": "analysis",
  "usage_examples": [
    {
      "command": "mycommand: sample input text",
      "description": "Analyzes the provided text input"
    },
    {
      "command": "mycommand: /path/to/file",
      "description": "Analyzes content from a file path"
    }
  ],
  "requirements": [
    "python>=3.8"
  ],
  "optional_dependencies": [
    "numpy",
    "requests"
  ],
  "tags": ["analysis", "custom", "text"]
}
```

### 🚀 **Using Your Custom Command**

After creating your command, restart Bielik CLI to load it:

```bash
# Restart Bielik CLI
./run.py

# Your command will now be available:
mycommand: analyze this text

# Check if it's loaded:
:help
```

### 🔍 **Context Provider Command Features**

**Built-in capabilities:**
- ✅ **Automatic loading** from `commands/` directory
- ✅ **Error handling** and validation
- ✅ **Help generation** from docstrings and config
- ✅ **Context integration** with AI prompts
- ✅ **Argument parsing** from `name: args` format
- ✅ **Multiple command support** in single session

**What makes a good Context Provider:**
- 📊 **Structured output** - Return organized, meaningful data
- 🎯 **Specific purpose** - Focus on one type of analysis
- ⚡ **Fast execution** - Avoid long-running operations
- 📝 **Clear documentation** - Good descriptions and examples
- 🛡️ **Error handling** - Graceful failure with helpful messages

### 🌟 **Real-world Examples**

**Available Context Providers:**
```bash
folder: .              # Analyze directory structure and files
calc: 2+3*4           # Mathematical calculations  
pdf: document.pdf     # PDF document analysis
git: /path/to/repo    # Git repository information (example above)
```

**Example workflow:**
```bash
👤 You: folder: ~/projects
✅ Context loaded! Directory analysis complete.

👤 You: What are the largest Python files in this project?
🤖 AI: [Uses folder context to identify and analyze Python files]

👤 You: git: .
✅ Context loaded! Git repository analysis complete.

👤 You: Summarize recent commits and current status
🤖 AI: [Uses git context to provide commit summaries and status]
```

---

## 📝 Development

```bash
git clone https://github.com/tom-sapletta-com/bielik.git
cd bielik
python -m venv .venv
source .venv/bin/activate
pip install -e .
pip install -e .[local]
pip install -e .[dev]
pip install -e .[gpu]
pip install -e .[vision]
```


# 🐳 Docker Testing Framework

Bielik CLI includes a comprehensive Docker testing framework for multiplatform installation verification across all major Linux distributions.

## Available Docker Tests

```bash
# Complete multiplatform test suite (all distributions)
make docker-test

# Individual distribution tests
make docker-test-ubuntu    # Ubuntu 22.04
make docker-test-debian    # Debian 12
make docker-test-alpine    # Alpine Linux 3.19
make docker-test-centos    # CentOS Stream 9
make docker-test-arch      # Arch Linux
make docker-test-oneliner  # One-liner installation simulation

# Docker management
make docker-build          # Build all test images
make docker-clean          # Clean Docker artifacts

# Complete test suite (Python + Docker)
make test-all              # Run both unit tests and Docker tests
```

## What Docker Tests Verify

✅ **Installation Success**: Bielik CLI installs without errors  
✅ **Context Provider Commands**: `folder:`, `calc:`, `pdf:` work correctly  
✅ **Cross-platform Compatibility**: Works on Ubuntu, Debian, Alpine, CentOS, Arch  
✅ **One-liner Installation**: `curl | bash` installation process  
✅ **Dependency Management**: Python packages install correctly  
✅ **Virtual Environment**: `.venv` setup and activation  

## Docker Test Architecture

```
docker/
├── test-multiplatform.yml     # Docker Compose configuration
├── Dockerfile.test-ubuntu     # Ubuntu 22.04 test environment
├── Dockerfile.test-debian     # Debian 12 test environment  
├── Dockerfile.test-alpine     # Alpine Linux 3.19 test environment
├── Dockerfile.test-centos     # CentOS Stream 9 test environment
├── Dockerfile.test-arch       # Arch Linux test environment
└── Dockerfile.test-oneliner   # One-liner installation test
```

Each test environment:
1. **Sets up clean Linux distribution**
2. **Installs system dependencies** (Python, git, build tools)
3. **Runs `python3 install.py --skip-ai`** to install Bielik with Context Provider Commands
4. **Verifies installation** with `python3 run.py --info`
5. **Tests Context Provider Commands** (folder analysis, calculations)

## Running Docker Tests in CI/CD

```yaml
# GitHub Actions example
- name: Run multiplatform Docker tests
  run: make docker-test

- name: Run specific distribution test
  run: make docker-test-ubuntu
```

# 📂 Package Structure

```
bielik/
├── bielik/
│   ├── __init__.py          # Package initialization
│   ├── cli.py               # CLI entry point (wrapper)
│   ├── client.py            # Client entry point (wrapper)
│   ├── server.py            # FastAPI web server
│   ├── config.py            # Configuration management
│   ├── hf_models.py         # Hugging Face model management
│   ├── content_processor.py # Content processing utilities
│   ├── cli/                 # Modular CLI components
│   │   ├── __init__.py
│   │   ├── main.py          # Main CLI entry and argument parsing
│   │   ├── commands.py      # Command processing and execution
│   │   ├── command_api.py   # Context Provider Commands API
│   │   ├── models.py        # HF model management CLI
│   │   ├── setup.py         # Interactive setup manager
│   │   └── send_chat.py     # Chat communication handling
│   └── client/              # Modular client components
│       ├── __init__.py      # Client package exports
│       ├── core.py          # Core BielikClient class
│       ├── model_manager.py # HF model operations for client
│       └── utils.py         # Client utility functions
├── commands/                # Context Provider Commands
│   ├── folder/             # Directory analysis command
│   ├── calc/               # Advanced calculator command
│   └── pdf/                # Document processing command
├── docker/                 # Docker testing framework
│   ├── test-multiplatform.yml
│   └── Dockerfile.test-*   # Test environments
├── tests/
│   ├── __init__.py
│   ├── test_cli.py          # CLI unit tests (updated for local HF models)
│   └── test_server.py       # Server unit tests
├── install.py              # Universal Python installer
├── run.py                  # Universal launcher
├── quick-install.sh        # Unix one-liner installer
├── quick-install.bat       # Windows one-liner installer
├── pyproject.toml          # Modern Python packaging
├── setup.cfg               # Package configuration
├── MANIFEST.in             # Package manifest
├── LICENSE                 # Apache 2.0 license
├── README.md               # This comprehensive documentation
├── Makefile                # Development automation with Docker tests
├── todo.md                 # Project specifications
└── .github/workflows/      # CI/CD automation
    └── python-publish.yml
```

## Author

Tom Sapletta — DevOps Engineer & Systems Architect

    💻 15+ years in DevOps, Software Development, and Systems Architecture
    🏢 Founder & CEO at Telemonit (Portigen - edge computing power solutions)
    🌍 Based in Germany | Open to remote collaboration
    📚 Passionate about edge computing, hypermodularization, and automated SDLC


## 📜 License

[Apache License 2.0](LICENSE)
