Skip to content

EndoTheDev/OllaMoMa-v2

Repository files navigation

OllaMoMa v2

A modern web interface for interacting with Ollama models, built with Nuxt 4 and Nuxt UI v4.

⚠️ Work in Progress: This project is under active development.

OllaMoMa v2 Preview

🌟 Features

  • 🎨 Interactive Chat Interface
    • Real-time streaming responses
    • Multiple model support
    • Clear conversation history
  • 🤖 Model Management
    • View and manage Ollama models
    • Search and sort functionality
    • Detailed model information
  • 🎨 Modern UI with Nuxt UI v4
    • Dark/Light mode with theme customization
    • Responsive design for all devices
    • Customizable UI elements
  • ⚡ Performance Focused
    • Fast and responsive interface
    • Efficient model handling
    • Type-safe development

🛠️ Tech Stack

📋 Prerequisites

Before you begin, ensure you have:

  • Node.js (v18+)
  • Ollama installed and running locally
  • Git for version control
  • Docker (optional, for containerized deployment)

🚀 Quick Start

Local Development

  1. Clone and Install

     # Clone the repository
     git clone https://github.com/EndoTheDev/OllaMoMa-v2.git
     cd OllaMoMa-v2
     # Install dependencies
     npm install  # or pnpm install / yarn install / bun install
  2. Start Development Server

     npm run dev  # or pnpm dev / yarn dev / bun run dev
  3. Configure Ollama

    • Default settings: host: 127.0.0.1, port: 11434
    • Adjust in application settings if needed

Docker Deployment

  1. Using Docker

     # Clone the repository
     git clone https://github.com/EndoTheDev/OllaMoMa-v2.git
     cd OllaMoMa-v2
     # Build the image
     docker build -t ollamoma-v2 .
     # Run the container
     docker run -d -p 3000:3000 --name ollamoma ollamoma-v2
  2. Using Docker Compose

     # Clone the repository
     git clone https://github.com/EndoTheDev/OllaMoMa-v2.git
     cd OllaMoMa-v2
     # Start the application
     docker compose up -d
     # Stop the application
     docker compose down
  3. Important Docker Notes

    • Ensure Ollama is running on your host machine
    • For Windows/macOS: Use host.docker.internal instead of localhost
    • For Linux: Use your host machine's IP address

📚 Documentation

Detailed documentation is available in the docs directory:

🏗️ Project Structure

.
├── app/                    # Application source (Nuxt 4 structure)
│   ├── assets/            # Static assets
│   ├── components/        # Vue components
│   │   ├── app/           # App components
│   │   │   ├── chat/      # Chat components
│   │   │   ├── modelfile/ # Modelfile components
│   │   │   ├── models/    # Models components
│   │   │   └── settings/  # Settings components
│   │   ├── base/          # Base components
│   │   └── ui/            # UI components
│   ├── composables/       # Vue composables
│   ├── pages/             # File-based routing
│   ├── stores/            # Pinia stores
│   ├── types/             # TypeScript type definitions
│   ├── app.vue            # Root component
│   └── app.config.ts      # App configuration
├── docs/                  # Documentation
├── public/                # Public static files
├── server/                # Server API routes
├── test/                  # Test files
└── nuxt.config.ts         # Nuxt configuration

📝 Roadmap

Current Progress

  • ✅ Core Features
    • ✅ Chat interface with streaming
    • ✅ Model management
    • ✅ Theme customization
    • ✅ Settings management

Upcoming Features

  • 🔄 LangChain integration
  • 📊 Knowledge graph visualization
  • 🎨 Enhanced theme customization
    • Color picker
    • Border radius settings
  • 🗃️ Database/Files management

🤝 Contributing

We welcome contributions! Please see our Development Guide for details on:

  • Setting up your development environment
  • Our coding standards
  • The contribution workflow
  • Testing requirements

📄 License

This project is licensed under the MIT License.


Built with ❤️ by EndoTheDev using Nuxt 4 and Nuxt UI v4

About

A modern web interface for Ollama, built with Nuxt 3 and Vue. Features a clean UI with dark/light modes, model management (copy/rename/delete), customizable themes, and a responsive layout. Designed for easy interaction with your local Ollama instance.

Topics

Resources

Stars

Watchers

Forks

Contributors

Languages