Template for a new AI Cloud project.
Click on Use this template to start your own project!
This project is a generative ai template. It contains the following features: LLMs, information extraction, chat, rag & evaluation. It uses LLMs(local or cloud), NiceGUI (frontend) & FastAPI (backend) & Promptfoo as an evaluation and redteam framework for your AI system.
| Test LLM |
|---|
![]() |
Engineering tools:
- Use UV to manage packages in a workspace (
frontendandbackend). - pre-commit hooks: use
ruffto ensure the code quality &detect-secretsto scan the secrets in the code. - Logging using loguru (with colors)
- Pytest for unit tests
- Dockerized project (Dockerfile & docker-compose).
- NiceGUI (frontend) & FastAPI (backend)
- Make commands to handle everything for you: install, run, test
AI tools:
- LLM running locally with Ollama or in the cloud with any LLM provider (LiteLLM)
- Information extraction and Question answering from documents
- Chat to test the AI system
- Efficient async code using asyncio.
- AI Evaluation framework: using Promptfoo, Ragas & more...
CI/CD & Maintenance tools:
- CI/CD pipelines:
.github/workflowsfor GitHub (Testing the AI system, local models with Ollama and the dockerized app) - Local CI/CD pipelines: GitHub Actions using
github act - GitHub Actions for deploying to GitHub Pages with mkdocs gh-deploy
- Dependabot
.github/dependabot.ymlfor automatic dependency and security updates
Documentation tools:
- Wiki creation and setup of documentation website using Mkdocs
- GitHub Pages deployment using mkdocs gh-deploy plugin
Upcoming features:
- add RAG again
- optimize caching in CI/CD
- Pull requests templates
- Additional MLOps templates: https://github.com/fmind/mlops-python-package
- Add MLFlow
- add Langfuse
This project is a monorepo containing two main packages:
frontend: A NiceGUI application.backend: A FastAPI application that serves the AI models and business logic.
The project uses uv as a package manager and is configured as a workspace, so dependencies for both packages can be installed with a single command.
frontend/: The NiceGUI frontend application.pyproject.toml: Frontend-specific dependencies.src/: Source code for the frontend.
backend/: The FastAPI backend application.pyproject.toml: Backend-specific dependencies.src/: Source code for the backend.
.github/: GitHub-specific files.workflows/: GitHub Actions CI/CD pipelines.dependabot.yml: Dependabot configuration for dependency updates.
assets/: Static assets like images.scripts/: Utility scripts.tests/: Pytest unit and integration tests..env.example: Example environment variables. Create a.envfrom this.Dockerfile: To build the application container.docker-compose.yml: To run services likefrontend,backend, andollama..gitlab-ci.yml: GitLab CI configuration file.Makefile: Shortcuts for common commands likeinstall,run,test.pyproject.toml: Definesuvworkspace and shared dependencies.uv.lock: Lock file foruvpackage manager..pre-commit-config.yaml: Configuration for pre-commit hooks.mkdocs.yml: Configuration for the documentation site.README.md,CONTRIBUTING.md,CODE_OF_CONDUCT.md,LICENSE: Project documentation.
- Ubuntu 22.04 or MacOS
- git clone the repository
- UV & Python 3.12 (will be installed by the Makefile)
- Create a
.envfile (take a look at the.env.examplefile)
This project uses a Makefile to simplify the installation and execution process.
-
For CPU-based environment (or MacOS) To install all dependencies for both
frontendandbackendfor a CPU environment, run:make install-dev
-
For NVIDIA GPU (CUDA) environment If you have an NVIDIA GPU and want to use CUDA for acceleration, run:
make install-dev-cuda
This will install the CUDA-enabled version of PyTorch.
The project can be fully containerized using Docker. This is the recommended way to run the application as it handles all services and networks.
- The
docker-compose.ymlanddocker-compose-cuda.ymlfiles define the services. - To build the main docker image:
make docker-build
- To run the entire application stack (frontend, backend, database, Ollama) using Docker Compose:
make run-app
Once installed (either locally or via Docker), you can run the services.
-
Run Everything: The
make run-appcommand is the easiest way to start all services, including the frontend, backend, database, and Ollama. -
Run Services Individually:
- Run Frontend:
make run-frontend - Run Backend:
make run-backend
- Run Frontend:
You can then access:
- Frontend (NiceGUI): http://localhost:8080 (or the configured port)
- Backend (FastAPI): http://localhost:8000 (or the configured port). Docs http://localhost:8000/docs
- Local model (Ollama):
- Install Ollama:
make install-ollama- Ensure Ollama is running (make run-ollamacan help). - Set your.envfile to point to the local Ollama endpoint (copy and paste from the.env.examplefile). - Download a model:make download-ollama-model- Test the connection:make test-ollama- Test the connection:make test-inference-llm - Cloud model (OpenAI, Anthropic, etc.):
- Update your
.envfile with the correct API keys and model names, following the LiteLLM naming convention. - Test the connection:
make test-inference-llm
- Update your
Check the CONTRIBUTING.md file for more information.
Check the CONTRIBUTING.md file for more information.

