# better-ai-launcher Better AI Launcher Image for RunPod and local development. ## RunPod Better App Manager Welcome to the RunPod Better App Manager! This image allows you to easily manage and run various AI applications on your RunPod instance. ### Features - Easy installation of pre-configured AI applications. - Start, stop, and monitor running applications. - View application logs in real-time. - Force kill applications if needed. - Download Manager for **HuggingFace** and **CivitAI** with `token` support for privat and gated models. - Shared Models Management for **Downloading and Sharing all models of all types to all installed AI applications**! ### Supported Applications - Better Comfy UI - Better Forge - Better A1111 - more Apps coming soon (AI Trainers as `Kohya` and `ai-toolkit` are planned) ### Getting Started - Access the Better App Manager interface through your RunPod instance URL. - Install the desired application by clicking the **Install** button. - Once installed, use the **Start** button to launch the application. - Access the running application using the **Open App** button. ### Troubleshooting If you encounter any issues: - Check the application logs for error messages. - Try stopping and restarting the application. - Use the `Force Kill` option if an application becomes unresponsive. - Refer to the RunPod documentation or contact support for further assistance. For more detailed information and guides, please visit the RunPod Documentation. Part of the `madiator-docker-runpod` familiy of **RunPod Docker Containers for RunPod** ## Github https://github.com/kodxana/madiator-docker-runpod
found under the directory `official-templates/better-ai-launcher` ## Build Options To build with default options, run `docker buildx bake`, to build a specific target, run `docker buildx bake `. ### Build Vars (based on bake selection) BASE_IMAGE=`$BASE_IMAGE`, e.g.
BASE_IMAGE=madiator2011/better-base:cuda12.4 ## Ports (System) - 22/tcp (SSH) - 7222/http (App-Manager) - 7777/http (VSCode-Server) - 8181/http (File-Browser) ## Ports (Apps) - 3000/http (ComfyUI) - 6006/http (Tensorboard [needed by kohya_ss]) - 7860/http (Kohya-ss) with FLUX.1 support - 7862/http (Forge) aka Stable-Diffiusion-WebUI-Forge - 7863/http (A1111) aka Stable-Diffiusion-WebUI ## ENV Vars (System) These ENV vars go into the docker container to support local debugging:
see also explanantion in `".vscode/tasks.json"` or `"docker-compose.debug.yml"` - LOCAL_DEBUG=True change app to localhost Urls and local Websockets (unsecured) for local debugging.
**TODO**: need to also setup a `bind workspace` in `".vscode/tasks.json"` or `"docker-compose.debug.yml"` if you **NOT** want need this behaviour, then set `LOCAL_DEBUG=False` [default],
which is the same as NOT setting this ENV var at all. - FLASK_ENV=development changed from "`production`" [default].
only needed when `LOCAL_DEBUG=True`, otherwise this ENV var can be obmitted. - GEVENT_SUPPORT=True gevent monkey-patching is being used, enable gevent support in the debugger.
only needed when `LOCAL_DEBUG=True`, otherwise this ENV var can be obmitted. - FLASK_DEBUG=0 "1" allows debugging in Chrome, but then the VSCode debugger will not works.
"0" is the [default], which is the same as NOT setting this ENV var at all. ### APP specific Vars - DISABLE_PULLBACK_MODELS=False The default is, that app model files, which are found locally (in only one app), get automatically `pulled-back` into the `"/workspace/shared_models"` folder.
From there they will be re-linked back not only to their own `pulled-back` model-type folder, but also will be linked back into all other corresponding app model-type folders.
So the `pulled-back` model is automatically shared to all installed apps. If you **NOT** want this behaviour, then set `DISABLE_PULLBACK_MODELS=True`,
otherwise set `DISABLE_PULLBACK_MODELS=False` [default], which is the same as NOT setting this ENV var at all. ## ENV Vars (User and Secret Tokens) **TODO: rename the file `"env.txt"` to `".env"` and adjust the ENV vars for your personal settings** - PUBLIC_KEY=ssh-ed25519 xxx...xxx usermail@domain.com your `PUBLIC ssh-key`
**Note**: make sure to use the **full line content** from your `"*.pub"` key file! - HF_TOKEN=hf_xxx...xxx Your `HuggingFace` token.

Can be a `READ` scoped token for downloading your `private` models, or `gated models` as e.g. `Flux.1 Dev` or METAs `Llama LLM models`.
The HF_TOKEN need to be a `READ/WRITE` scoped token, if you plan also to **UPLOAD** models to `HuggingFace` later, when we have Trainer Apps like `Kohya` or `ai-toolkit`. - CIVITAI_API_TOKEN=xxx...xxx Your `CivitAI` API token.

**Note**: CivitAI currently only provides a `FULL` user token, acting as `you`, so be careful with how to setup this token and with whom you share it! **SECURITY TIP:** These three, user-specific and **security sensitive environment vars**, should be stored as RUNPOD **`SECRETS`** and be referenced directly in your POD Template in the format `{{ RUNPOD_SECRET_MYENVVAR }}`. From https://docs.runpod.io/pods/templates/secrets You can reference your Secret directly in the Environment Variables section of your Pod template. To reference your Secret, reference it's key appended to the `RUNPOD_SECRET_` prefix. That mean, for this template/image, you should use these formats to pass the above ENV vars into the docker container: - `{{ RUNPOD_SECRET_PUBLIC_KEY}}` - `{{ RUNPOD_SECRET_HF_TOKEN }}` - `{{ RUNPOD_SECRET_CIVITAI_API_TOKEN }}` (c) 2024 RunPod Better App Manager. Created by Madiator2011.