diff --git a/yuy-chat/README.md b/yuy-chat/README.md
deleted file mode 100644
index 696ecd9..0000000
--- a/yuy-chat/README.md
+++ /dev/null
@@ -1,180 +0,0 @@
-# yuy-chat
-
-
-
-```
-$$\ $$\
-\$$\ $$ |
- \$$\ $$ /$$\ $$\ $$\ $$\
- \$$$$ / $$ | $$ |$$ | $$ |
- \$$ / $$ | $$ |$$ | $$ |
- $$ | $$ | $$ |$$ | $$ |
- $$ | \$$$$$$ |\$$$$$$$ |
- \__| \______/ \____$$ |
- $$\ $$ |
- \$$$$$$ |
- \______/
-```
-
-**Beautiful TUI chat interface for local AI models**
-
-[](https://www.rust-lang.org)
-[](https://opensource.org/licenses/MIT)
-
-
-
----
-
-## 🌟 Features
-
-- ✨ **Beautiful TUI** - Gorgeous terminal interface powered by ratatui
-- 🔍 **Auto-discovery** - Automatically finds `.gguf` and `.llamafile` models
-- 🎨 **Presets** - Creative, Balanced, and Precise modes
-- 💾 **Save conversations** - Keep your chat history
-- 🌐 **HuggingFace API** - Use models from HuggingFace (optional)
-- ⚡ **Fast & Lightweight** - ~5MB binary, minimal dependencies
-- 🚀 **Streaming responses** - See words appear as they're generated
-- 🎯 **Zero configuration** - Just run and chat
-
-## 📦 Installation
-
-### From source:
-
-```bash
-git clone https://github.com/YuuKi-OS/yuy-chat
-cd yuy-chat
-cargo build --release
-```
-
-### Install globally:
-
-```bash
-cargo install --path .
-```
-
-## 🚀 Quick Start
-
-```bash
-# Run yuy-chat
-yuy-chat
-
-# It will auto-scan ~/.yuuki/models/ for .gguf and .llamafile files
-# Select a model and start chatting!
-```
-
-## 📁 Supported Model Formats
-
-- ✅ **GGUF** (`.gguf`) - Runs with llama.cpp
-- ✅ **Llamafile** (`.llamafile`) - Self-contained executables
-
-## 🎮 Controls
-
-### Model Selector
-- `↑/↓` or `j/k` - Navigate models
-- `Enter` - Select model
-- `R` - Refresh model list
-- `Q` - Quit
-
-### Chat
-- `Type` - Write your message
-- `Enter` - Send message
-- `Shift+Enter` - New line
-- `Ctrl+Enter` - Send (always)
-- `Ctrl+C` - Open menu
-- `Ctrl+L` - Clear chat
-- `Ctrl+S` - Save conversation
-- `↑/↓` - Scroll chat (when input is empty)
-
-### Menu
-- `1` - Change model
-- `2` - Change preset
-- `3` - Save conversation
-- `4` - Load conversation
-- `5` - Clear chat
-- `6` - Settings
-- `Q` - Back to chat
-
-## ⚙️ Configuration
-
-Config file location: `~/.config/yuy-chat/config.toml`
-
-```toml
-models_dir = "/home/user/.yuuki/models"
-hf_token = "hf_xxxxxxxxxxxxx" # Optional
-default_preset = "Balanced"
-save_history = true
-theme = "Dark"
-```
-
-## 🎯 Presets
-
-- **Creative** (temp: 0.8, top_p: 0.9) - More random and creative
-- **Balanced** (temp: 0.6, top_p: 0.7) - Good middle ground
-- **Precise** (temp: 0.3, top_p: 0.5) - More focused and deterministic
-
-## 🌐 HuggingFace Integration
-
-Add your HuggingFace token in settings to use models via API:
-
-1. Press `Ctrl+C` → `6` (Settings)
-2. Edit `HuggingFace Token`
-3. Paste your token from https://huggingface.co/settings/tokens
-4. Save and refresh models
-
-## 📚 Directory Structure
-
-```
-~/.config/yuy-chat/
-├── config.toml # Configuration
-└── conversations/ # Saved chats
- ├── conversation-20240206-143022.json
- └── conversation-20240206-150133.json
-```
-
-## 🔧 Requirements
-
-- **Rust 1.70+** (for building)
-- **llama.cpp** (for .gguf models) - Install with: `yuy runtime install llama-cpp`
-- **chmod +x** (for .llamafile models)
-
-## 🤝 Integration with yuy
-
-yuy-chat is designed to work alongside [yuy](https://github.com/YuuKi-OS/yuy):
-
-```bash
-# Download models with yuy
-yuy download Yuuki-best
-
-# Chat with yuy-chat
-yuy-chat
-```
-
-## 🐛 Troubleshooting
-
-**No models found?**
-- Make sure you have models in `~/.yuuki/models/`
-- Or specify custom directory: `yuy-chat --models-dir /path/to/models`
-
-**llama.cpp not found?**
-- Install with: `yuy runtime install llama-cpp`
-- Or: `brew install llama.cpp` (macOS)
-- Or: `pkg install llama-cpp` (Termux)
-
-**Streaming not working?**
-- Ensure llama.cpp is installed and in PATH
-- Check model file permissions
-
-## 📝 License
-
-MIT License - see [LICENSE](LICENSE) file
-
-## 🌸 Credits
-
-Made with love by the Yuuki team
-
-- TUI Framework: [ratatui](https://github.com/ratatui-org/ratatui)
-- Inference: [llama.cpp](https://github.com/ggerganov/llama.cpp)
-
----
-
-**For model management, see [yuy](https://github.com/YuuKi-OS/yuy)**