App Consolidation, app_configs and app_utils, Dockerfile

This commit is contained in:
lutzapps 2024-11-09 04:19:56 +07:00
parent 8aa61f672c
commit 6a71890ae4
7 changed files with 683 additions and 440 deletions

View file

@ -7,22 +7,75 @@ FROM ${BASE_IMAGE:-madiator2011/better-base:cuda12.4} AS base
ARG BASE_IMAGE
ENV BASE_IMAGE=$BASE_IMAGE
# lutzapps - replaced by above bake build-args
#FROM madiator2011/better-base:cuda12.4 AS base
# lutzapps - prepare for local developement and debugging
# needed to change the ORDER of "apt-get commands" and move the "update-alternatives" for python3
# AFTER the "apt-get remove -y python3.10" cmd, OTHERWISE the symlink to python3
# is broken in the image and the VSCode debugger could not exec "python3" as CMD overwrite
# also fixed a boring "Blinker" blocking error
# Install Python 3.11, set it as default, and remove Python 3.10
RUN apt-get update && \
### ---> needed Tools for Installer
# removed: 2x git nginx ffmpeg (as they are already installed with the base image)
# added: pigz (for parallel execution of TAR files); zip (for easier folder compression)
apt-get install -y python3.11 python3.11-venv python3.11-dev python3.11-distutils \
aria2 pigz zip pv rsync zstd libtcmalloc-minimal4 bc && \
apt-get remove -y python3.10 python3.10-minimal libpython3.10-minimal libpython3.10-stdlib && \
apt-get install -y aria2 pigz zip pv rsync zstd libtcmalloc-minimal4 bc \
# add Python3.11 as system Python version, serving the Python Flask App
python3.11 python3.11-venv python3.11-dev python3.11-distutils && \
# not remove Python3.10, as we need it for "official" app support (e.g. for kohya_ss VENV)
###apt-get remove -y python3.10 python3.10-minimal libpython3.10-minimal libpython3.10-stdlib && \
#
# setup an "alias" for "python" be symlinked to Python3.11
# (which is the default anyway after this installation here of Python 3.11)
update-alternatives --install /usr/bin/python python /usr/bin/python3.11 1 && \
#
# setup the "python3" alias for Python3.11, as this is what the debugger needs (not work with 3.10)
update-alternatives --install /usr/bin/python3 python3 /usr/bin/python3.11 1 && \
#
# VENV will have their own "preferred/supported/recommended" Python version, as e.g.
# the "kohya_ss" app, only supports Python up to version 3.10 (but not 3.11)
#
# until here we have a broken "Blinker" installation from some base images before,
# and if we try to "update" "Blinker" for Python3.10 or via e.g. "pip install --upgrade blinker",
# or "pip install blinker==x.y.z.z", this breaks, as it was installed in an APT bundle,
# which can not be safely upgraded (= Uninstall/Reinstall)! It breaks as follows:
# we get a blinker 1.4 uninstall error chained by trying to uninstall "distutils":
# 8.568 Found existing installation: blinker 1.4
# 8.782 error: uninstall-distutils-installed-package
# 8.782
# 8.782 × Cannot uninstall blinker 1.4
# 8.782 ╰─> It is a distutils installed project and thus we cannot accurately determine which files belong to it which would lead to only a partial uninstall.
# that not only blocks building the docker image, but later also breaks the "kohya_ss" app during setup,
# which try to install Blinker and Python310-venv from "setup-runpod.sh":
# # Install tk and python3.10-venv
# echo "Installing tk and python3.10-venv..."
# apt update -y && apt install -y python3-tk python3.10-venv
#
# Python 3.10 needs to run as Kohya's "official" requirement, and is used in the VENV
#
# first uninstall the APT bundle package for "Blinker"
apt-get remove -y python3-blinker && \
# then re-install the APT unbundled package of "Blinker back",
# together with Python3.10 venv, which we need to setup the kohya_ss VENV
apt-get install -y python3-tk python3.10-venv && \
# this re-captures back the "python3" alias for Python3.10, but not the "python" alias (stays for Python3.11)
# the "Python3.11" and "Python3.10" cmds work too
# global PIP is 3.11, VENV pip is 3.10
#
# ---> CUDA 12.4 Toolkit (is already in the 12.4 base-image)
wget https://developer.download.nvidia.com/compute/cuda/repos/ubuntu2204/x86_64/cuda-keyring_1.1-1_all.deb && \
dpkg -i cuda-keyring_1.1-1_all.deb && \
### need to refresh the repository metatdata, after downloading this NVIDIA downloaded package list!!!
apt-get update && \
apt-get -y install cuda-toolkit-12-4 && \
#
# ---> get the latest cuDNN 9.x version supporting CUDA 12.x
# remove the current dev package which depends on libcudnn9-cuda-12 and breaks the upgrade otherwise
apt-get remove -y --allow-change-held-packages libcudnn9-cuda-12 libcudnn9-dev-cuda-12 && \
apt-get -y --allow-change-held-packages install cudnn-cuda-12 && \
#
# clean-up resources and caches
apt-get autoremove -y && \
apt-get clean && \
rm -rf /var/lib/apt/lists/*

View file

@ -28,8 +28,7 @@ from utils.shared_models import (
SHARED_MODELS_DIR, SHARED_MODEL_FOLDERS, SHARED_MODEL_FOLDERS_FILE, ensure_shared_models_folders,
APP_INSTALL_DIRS, APP_INSTALL_DIRS_FILE, init_app_install_dirs, # APP_INSTALL_DIRS dict/file/function
MAP_APPS, sync_with_app_configs_install_dirs, # internal MAP_APPS dict and sync function
SHARED_MODEL_APP_MAP, SHARED_MODEL_APP_MAP_FILE, init_shared_model_app_map, # SHARED_MODEL_APP_MAP dict/file/function
write_dict_to_jsonfile, read_dict_from_jsonfile, PrettyDICT # JSON helper functions
SHARED_MODEL_APP_MAP, SHARED_MODEL_APP_MAP_FILE, init_shared_model_app_map # SHARED_MODEL_APP_MAP dict/file/function
)
# the "update_model_symlinks()" function replaces the app.py function with the same same
# and redirects to same function name "update_model_symlinks()" in the new "utils.shared_models" module
@ -340,92 +339,12 @@ def remove_existing_app_config(app_name):
return jsonify({'status': 'success', 'message': f'App {app_name} removed successfully'})
return jsonify({'status': 'error', 'message': f'App {app_name} not found'})
# unused function
def obsolate_update_model_symlinks():
# lutzapps - CHANGE #3 - use the new "shared_models" module for app model sharing
# remove this whole now unused function
return "replaced by utils.shared_models.update_model_symlinks()"
# modified function
def setup_shared_models():
# lutzapps - CHANGE #4 - use the new "shared_models" module for app model sharing
jsonResult = update_model_symlinks()
return SHARED_MODELS_DIR # shared_models_dir is now owned and managed by the "shared_models" utils module
# remove below unused code
shared_models_dir = '/workspace/shared_models'
model_types = ['Stable-diffusion', 'VAE', 'Lora', 'ESRGAN']
# Create shared models directory if it doesn't exist
os.makedirs(shared_models_dir, exist_ok=True)
for model_type in model_types:
shared_model_path = os.path.join(shared_models_dir, model_type)
# Create shared model type directory if it doesn't exist
os.makedirs(shared_model_path, exist_ok=True)
# Create a README file in the shared models directory
readme_path = os.path.join(shared_models_dir, 'README.txt')
if not os.path.exists(readme_path):
with open(readme_path, 'w') as f:
f.write("Upload your models to the appropriate folders:\n\n")
f.write("- Stable-diffusion: for Stable Diffusion models\n")
f.write("- VAE: for VAE models\n")
f.write("- Lora: for LoRA models\n")
f.write("- ESRGAN: for ESRGAN upscaling models\n\n")
f.write("These models will be automatically linked to all supported apps.")
print(f"Shared models directory created at {shared_models_dir}")
print("Shared models setup completed.")
return shared_models_dir
# unused function
def obsolate_update_model_symlinks():
# lutzapps - CHANGE #5 - use the new "shared_models" module for app model sharing
# remove this whole now unused function
return "replaced by utils.shared_models.update_model_symlinks()"
shared_models_dir = '/workspace/shared_models'
apps = {
'stable-diffusion-webui': '/workspace/stable-diffusion-webui/models',
'stable-diffusion-webui-forge': '/workspace/stable-diffusion-webui-forge/models',
'ComfyUI': '/workspace/ComfyUI/models'
}
model_types = ['Stable-diffusion', 'VAE', 'Lora', 'ESRGAN']
for model_type in model_types:
shared_model_path = os.path.join(shared_models_dir, model_type)
if not os.path.exists(shared_model_path):
continue
for app, app_models_dir in apps.items():
if app == 'ComfyUI':
if model_type == 'Stable-diffusion':
app_model_path = os.path.join(app_models_dir, 'checkpoints')
elif model_type == 'Lora':
app_model_path = os.path.join(app_models_dir, 'loras')
elif model_type == 'ESRGAN':
app_model_path = os.path.join(app_models_dir, 'upscale_models')
else:
app_model_path = os.path.join(app_models_dir, model_type.lower())
else:
app_model_path = os.path.join(app_models_dir, model_type)
# Create the app model directory if it doesn't exist
os.makedirs(app_model_path, exist_ok=True)
# Create symlinks for each file in the shared model directory
for filename in os.listdir(shared_model_path):
src = os.path.join(shared_model_path, filename)
dst = os.path.join(app_model_path, filename)
if os.path.isfile(src) and not os.path.exists(dst):
os.symlink(src, dst)
print("Model symlinks updated.")
def update_symlinks_periodically():
while True:
@ -436,57 +355,6 @@ def start_symlink_update_thread():
thread = threading.Thread(target=update_symlinks_periodically, daemon=True)
thread.start()
# unused function
def obsolate_recreate_symlinks():
# lutzapps - CHANGE #6 - use the new "shared_models" module for app model sharing
# remove this whole now unused function
return "replaced by utils.shared_models.update_model_symlinks()"
shared_models_dir = '/workspace/shared_models'
apps = {
'stable-diffusion-webui': '/workspace/stable-diffusion-webui/models',
'stable-diffusion-webui-forge': '/workspace/stable-diffusion-webui-forge/models',
'ComfyUI': '/workspace/ComfyUI/models'
}
model_types = ['Stable-diffusion', 'VAE', 'Lora', 'ESRGAN']
for model_type in model_types:
shared_model_path = os.path.join(shared_models_dir, model_type)
if not os.path.exists(shared_model_path):
continue
for app, app_models_dir in apps.items():
if app == 'ComfyUI':
if model_type == 'Stable-diffusion':
app_model_path = os.path.join(app_models_dir, 'checkpoints')
elif model_type == 'Lora':
app_model_path = os.path.join(app_models_dir, 'loras')
elif model_type == 'ESRGAN':
app_model_path = os.path.join(app_models_dir, 'upscale_models')
else:
app_model_path = os.path.join(app_models_dir, model_type.lower())
else:
app_model_path = os.path.join(app_models_dir, model_type)
# Remove existing symlinks
if os.path.islink(app_model_path):
os.unlink(app_model_path)
elif os.path.isdir(app_model_path):
shutil.rmtree(app_model_path)
# Create the app model directory if it doesn't exist
os.makedirs(app_model_path, exist_ok=True)
# Create symlinks for each file in the shared model directory
for filename in os.listdir(shared_model_path):
src = os.path.join(shared_model_path, filename)
dst = os.path.join(app_model_path, filename)
if os.path.isfile(src) and not os.path.exists(dst):
os.symlink(src, dst)
return "Symlinks recreated successfully."
# modified function
@app.route('/recreate_symlinks', methods=['POST'])
def recreate_symlinks_route():
@ -494,13 +362,6 @@ def recreate_symlinks_route():
jsonResult = update_model_symlinks()
return jsonResult
# remove below unused code
try:
message = recreate_symlinks()
return jsonify({'status': 'success', 'message': message})
except Exception as e:
return jsonify({'status': 'error', 'message': str(e)})
# modified function
@app.route('/create_shared_folders', methods=['POST'])
@ -508,39 +369,6 @@ def create_shared_folders():
# lutzapps - CHANGE #8 - use the new "shared_models" module for app model sharing
jsonResult = ensure_shared_models_folders()
return jsonResult
# remove below unused code
try:
shared_models_dir = '/workspace/shared_models'
model_types = ['Stable-diffusion', 'Lora', 'embeddings', 'VAE', 'hypernetworks', 'aesthetic_embeddings', 'controlnet', 'ESRGAN']
# Create shared models directory if it doesn't exist
os.makedirs(shared_models_dir, exist_ok=True)
for model_type in model_types:
shared_model_path = os.path.join(shared_models_dir, model_type)
# Create shared model type directory if it doesn't exist
os.makedirs(shared_model_path, exist_ok=True)
# Create a README file in the shared models directory
readme_path = os.path.join(shared_models_dir, 'README.txt')
if not os.path.exists(readme_path):
with open(readme_path, 'w') as f:
f.write("Upload your models to the appropriate folders:\n\n")
f.write("- Stable-diffusion: for Stable Diffusion checkpoints\n")
f.write("- Lora: for LoRA models\n")
f.write("- embeddings: for Textual Inversion embeddings\n")
f.write("- VAE: for VAE models\n")
f.write("- hypernetworks: for Hypernetwork models\n")
f.write("- aesthetic_embeddings: for Aesthetic Gradient embeddings\n")
f.write("- controlnet: for ControlNet models\n")
f.write("- ESRGAN: for ESRGAN upscaling models\n\n")
f.write("These models will be automatically linked to all supported apps.")
return jsonify({'status': 'success', 'message': 'Shared model folders created successfully.'})
except Exception as e:
return jsonify({'status': 'error', 'message': str(e)})
def save_civitai_token(token):
with open(CIVITAI_TOKEN_FILE, 'w') as f:
@ -644,7 +472,7 @@ def get_model_types_route():
'desc': model_type_description
}
i = i + 1
i += 1
return model_types_dict

View file

@ -2418,7 +2418,7 @@
<div class="download-info">
<span class="download-speed"></span>
<span class="download-eta"></span>
</div>4
</div><!-- 4 lutzapps - remove a typo "4" -->
<div class="install-stage"></div>
</div>
</div>
@ -2838,6 +2838,7 @@
function initializeUI() {
updateStatus();
// TODO: need a way this 2 functions not pollute the logs every second or 5 seconds
setInterval(updateStatus, 5000);
setInterval(updateLogs, 1000);
@ -3281,15 +3282,6 @@
});
}
// lutzapps - obsolete function (can be deleted)
function copyToClipboard(text) {
navigator.clipboard.writeText(text).then(() => {
alert('URL copied to clipboard!');
}, (err) => {
console.error('Could not copy text: ', err);
});
}
// Call this function when the Models tab is opened
document.querySelector('.navbar-tabs a[onclick="openTab(event, \'models-tab\')"]').addEventListener('click', function() {
//alert("querySelector");

View file

@ -3,71 +3,251 @@ import xml.etree.ElementTree as ET
import requests
import json
def fetch_app_info():
url = "https://better.s3.madiator.com/"
response = requests.get(url)
root = ET.fromstring(response.content)
app_info = {}
for content in root.findall('{http://s3.amazonaws.com/doc/2006-03-01/}Contents'):
key = content.find('{http://s3.amazonaws.com/doc/2006-03-01/}Key').text
size = int(content.find('{http://s3.amazonaws.com/doc/2006-03-01/}Size').text)
app_name = key.split('/')[0]
# lutzapps - fix "bug" in key element of the S3 XML document
# all other three apps have a "key" element like "bcomfy/bcomfy.tar.gz" or "bforge/bforge.tar.gz",
# with their "app_name" prefix + "/" + tar_filename
# only kohya is missing this "app_name" prefix and has a key element of only its tar_filename "bkohya.tar.gz"
# this results in the app_name "bkohya.tar.gz", instead of only "bkohya"
# TODO for madiator - move the "bkohya.tar.gz" into a subfolder "bkohya" in your S3 bucket
app_name = app_name.replace(".tar.gz", "") # cut any extension postfixes resulting from the wrong key.split() command
if app_name in ['ba1111', 'bcomfy', 'bforge', 'bkohya']: # lutzapps - added new kohya app
app_info[app_name] = {
'download_url': f"https://better.s3.madiator.com/{key}",
'size': size
}
return app_info
# this is the replacement for the XML manifest, and defines all app_configs in full detail
APP_CONFIGS_MANIFEST_URL = "https://better.s3.madiator.com/app_configs.json"
# if this JSON can not be downloaded, the below code defaults apply
# this app_configs dict can also be generated from code when at least one of following
# 2 ENV vars are found with following values:
# 1. LOCAL_DEBUG = 'True' # this ENV var should not be passed when in the RUNPOD environment, as it disabled the CF proxy Urls of the App-Manager
# and this ENV var also controls some other aspects of the app.
#
# 2. APP_CONFIGS_FILE = 'True' # only exists for this one purpose, to generate the below Dict as file
# "/workspace/_app_configs.json", which then can be uploaded to the above defined APP_CONFIGS_MANIFEST_URL
# NOTE:
app_configs = {
'bcomfy': {
'id': 'bcomfy',
'name': 'Better Comfy UI',
'command': 'cd /workspace/bcomfy && . ./bin/activate && cd /workspace/ComfyUI && python main.py --listen --port 3000 --enable-cors-header',
'venv_path': '/workspace/bcomfy',
'app_path': '/workspace/ComfyUI',
'port': 3000,
'download_url': 'https://better.s3.madiator.com/bcomfy/bcomfy.tar.gz', # (2024-11-08 18:50:00Z - lutzapps)
#'venv_uncompressed_size': 6452737952, # uncompressed size of the tar-file (in bytes) - lutzapps new version
'venv_uncompressed_size': 6155295493, # uncompressed size of the tar-file (in bytes) - original version
#'archive_size': 3389131462 # tar filesize (in bytes) - lutzapps new version
'archive_size': 3179595118, # tar filesize (in bytes) - original version
#'sha256_hash': '18e7d71b75656924f98d5b7fa583aa7c81425f666a703ef85f7dd0acf8f60886', # lutzapps new version
'sha256_hash': '7fd60808a120a1dd05287c2a9b3d38b3bdece84f085abc156e0a2ee8e6254b84', # original version
'repo_url': 'https://github.com/comfyanonymous/ComfyUI.git',
'branch_name': '', # empty branch_name means default = 'master'
'commit': '', # or commit hash (NYI)
'recursive': False,
'refresh': False,
'custom_nodes': [ # following custom_nodes will be git cloned and installed with "pip install -r requirements.txt" (in Testing)
{
'name': 'ComfyUI-Manager (ltdrdata)', # this node is installed in the VENV
'path': 'ComfyUI-Manager',
'repo_url': 'https://github.com/ltdrdata/ComfyUI-Manager.git'
},
{
'name': 'ComfyUI-Essentials (cubic)', # this node is installed in the VENV
'path': 'ComfyUI_essentials',
'repo_url': 'https://github.com/cubiq/ComfyUI_essentials'
}
### planned custom nodes - To Be Discussed
# {
# 'name': 'rgthree comfy',
# 'path': 'rgthree-comfy',
# 'repo_url': 'https://github.com/rgthree/rgthree-comfy'
# },
# {
# 'name': 'was node suite comfyui',
# 'path': 'was-node-suite-comfyui',
# 'repo_url': 'https://github.com/WASasquatch/was-node-suite-comfyui'
# },
# {
# 'name': 'comfyui controlnet aux',
# 'path': 'comfyui_controlnet_aux',
# 'repo_url': 'https://github.com/Fannovel16/comfyui_controlnet_aux'
# },
# {
# 'name': 'x-flux-comfyui (XLabs-AI)',
# 'path': 'x-flux-comfyui',
# 'repo_url': 'https://github.com/XLabs-AI/x-flux-comfyui'
# },
# {
# 'name': 'ComfyUI-GGUF (city96)',
# 'path': 'ComfyUI-GGUF',
# 'repo_url': 'https://github.com/city96/ComfyUI-GGUF'
# },
# {
# 'name': 'ComfyUI-Florence2 (kijai)',
# 'path': 'ComfyUI-Florence2F',
# 'repo_url': 'https://github.com/kijai/ComfyUI-Florence2'
# },
# {
# 'name': 'ComfyUI-KJNodes (kijai)',
# 'path': 'ComfyUI-KJNodes',
# 'repo_url': 'https://github.com/kijai/ComfyUI-KJNodes'
# },
# {
# 'name': 'ComfyUI_UltimateSDUpscale (ssitu)',
# 'path': 'ComfyUI_UltimateSDUpscale',
# 'repo_url': 'https://github.com/ssitu/ComfyUI_UltimateSDUpscale'
# },
# {
# 'name': 'ControlAltAI Nodes (gseth)',
# 'path': 'ControlAltAI-Nodes',
# 'repo_url': 'https://github.com/gseth/ControlAltAI-Nodes'
# },
# {
# 'name': 'ComfyUI Easy-Use (yolain)',
# 'path': 'ComfyUI-Easy-Use',
# 'repo_url': 'https://github.com/yolain/ComfyUI-Easy-Use'
# },
# {
# 'name': 'ComfyUI Impact-Pack (tdrdata)',
# 'path': 'ComfyUI-Impact-Pack',
# 'repo_url': 'https://github.com/ltdrdata/ComfyUI-Impact-Pack'
# }
]
},
'bforge': {
'id': 'bforge', # app_name
'name': 'Better Forge',
'command': 'cd /workspace/bforge && . ./bin/activate && cd /workspace/stable-diffusion-webui-forge && ./webui.sh -f --listen --enable-insecure-extension-access --api --port 7862',
'venv_path': '/workspace/bforge',
'app_path': '/workspace/stable-diffusion-webui-forge',
'port': 7862,
'download_url': 'https://better.s3.madiator.com/bforge/bforge.tar.gz',
'venv_uncompressed_size': 7689838771, # uncompressed size of the tar-file (in bytes),
'archive_size': 3691004078, # tar filesize (in bytes)
'sha256_hash': 'e87dae2324a065944c8d36d6ac4310af6d2ba6394f858ff04a34c51aa5f70bfb',
'repo_url': 'https://github.com/lllyasviel/stable-diffusion-webui-forge.git',
'branch_name': '', # empty branch_name means default = 'master'
'commit': '', # or commit hash (NYI)
'clone_recursive': False,
'refresh': False
},
'ba1111': {
'id': 'ba1111', # app_name
'name': 'Better A1111',
'command': 'cd /workspace/ba1111 && . ./bin/activate && cd /workspace/stable-diffusion-webui && ./webui.sh -f --listen --enable-insecure-extension-access --api --port 7863',
'venv_path': '/workspace/ba1111',
'app_path': '/workspace/stable-diffusion-webui',
'port': 7863,
'download_url': 'https://better.s3.madiator.com/ba1111/ba1111.tar.gz',
'venv_uncompressed_size': 6794367826, # uncompressed size of the tar-file (in bytes),
'archive_size': 3383946179, # tar filesize (in bytes)
'sha256_hash': '1d70276bc93f5f992a2e722e76a469bf6a581488fa1723d6d40739f3d418ada9',
'repo_url': 'https://github.com/AUTOMATIC1111/stable-diffusion-webui.git',
'branch_name': '', # empty branch_name means default = 'master'
'commit': '', # or commit hash (NYI)
'clone_recursive': False,
'refresh': False
},
'bkohya': {
'id': 'bkohya', # app_name
'name': 'Better Kohya',
'command': 'cd /workspace/bkohya && . ./bin/activate && cd /workspace/kohya_ss && ./gui.sh --listen --port 7860',
'command': 'cd /workspace/bkohya && . ./bin/activate && cd /workspace/kohya_ss && python ./kohya_gui.py --headless --share --server_port 7860', # TODO!! check ./kohya_gui.py
### for Gradio supported reverse proxy:
# --share -> Share the gradio UI
# --root_path ROOT_PATH -> root_path` for Gradio to enable reverse proxy support. e.g. /kohya_ss
# --listen LISTEN -> IP to listen on for connections to Gradio
# usage: kohya_gui.py [-h] [--config CONFIG] [--debug] [--listen LISTEN]
# [--username USERNAME] [--password PASSWORD]
# [--server_port SERVER_PORT] [--inbrowser] [--share]
# [--headless] [--language LANGUAGE] [--use-ipex]
# [--use-rocm] [--do_not_use_shell] [--do_not_share]
# [--requirements REQUIREMENTS] [--root_path ROOT_PATH]
# [--noverify]
#
# options:
# -h, --help show this help message and exit
# --config CONFIG Path to the toml config file for interface defaults
# --debug Debug on
# --listen LISTEN IP to listen on for connections to Gradio
# --username USERNAME Username for authentication
# --password PASSWORD Password for authentication
# --server_port SERVER_PORT
# Port to run the server listener on
# --inbrowser Open in browser
# --share Share the gradio UI
# --headless Is the server headless
# --language LANGUAGE Set custom language
# --use-ipex Use IPEX environment
# --use-rocm Use ROCm environment
# --do_not_use_shell Enforce not to use shell=True when running external
# commands
# --do_not_share Do not share the gradio UI
# --requirements REQUIREMENTS
# requirements file to use for validation
# --root_path ROOT_PATH
# `root_path` for Gradio to enable reverse proxy
# support. e.g. /kohya_ss
# --noverify Disable requirements verification
'venv_path': '/workspace/bkohya',
'app_path': '/workspace/kohya_ss',
'port': 7860,
'download_url': 'https://better.s3.madiator.com/kohya.tar.gz', # (2024-11-08 13:13:00Z) - lutzapps
'venv_uncompressed_size': 12128345264, # uncompressed size of the tar-file (in bytes)
'archive_size': 6314758227, # tar filesize (in bytes)
'sha256_hash': '9a0c0ed5925109e82973d55e28f4914fff6728cfb7f7f028a62e2ec1a9e4f60a',
'repo_url': 'https://github.com/bmaltais/kohya_ss.git',
'branch_name': 'sd3-flux.1', # make sure we use Kohya with FLUX support branch
# this branch also uses a 'sd-scripts' HEAD branch of 'SD3', which gets automatically checked-out too
'commit': '', # or commit hash (NYI)
'clone_recursive': True, # is recursive clone
'refresh': False
}
}
def update_app_configs():
# lutzapps - not used anymore TODO: remove later
""" def fetch_app_info():
manifest_url = "https://better.s3.madiator.com/"
download_base_url = "https://better.s3.madiator.com/" # could be different base as the manifest file
app_info = {}
try: # be graceful when the server is not reachable, be it S3 or anything else
response = requests.get(manifest_url)
root = ET.fromstring(response.content)
for content in root.findall('{http://s3.amazonaws.com/doc/2006-03-01/}Contents'):
app_name_and_url = content.find('{http://s3.amazonaws.com/doc/2006-03-01/}Key').text
app_name = app_name_and_url.split('/')[0] # e.g. "bkohya/bkohya.tar.gz" -> "bkohya"
download_url = os.path.join(download_base_url, app_name_and_url)
if not (app_name in ['ba1111', 'bcomfy', 'bforge', 'bkohya']):
continue # skip unsupported app
# load code defaults
archive_size = app_configs[app_name]["archive_size"]
venv_uncompressed_size = app_configs[app_name]["venv_uncompressed_size"]
sha256_hash = app_configs[app_name]["sha256_hash"]
try: # try to find overwrites from code defaults
archive_size = int(content.find('archive_size').text)
venv_uncompressed_size = int(content.find('{http://s3.amazonaws.com/doc/2006-03-01/}venv_uncompressed_size').text)
sha256_hash = int(content.find('{http://s3.amazonaws.com/doc/2006-03-01/}sha256_hash').text)
except: # swallow any exception, mainly from not being defined (yet) in the XML manifest
print(f"App '{app_name}' Metadata could not be found in manifest '{manifest_url}', using code defaults!")
app_info[app_name] = {
'download_url': download_url,
'archive_size': archive_size,
'venv_uncompressed_size': venv_uncompressed_size, # TODO: provide in XML manifest
'sha256_hash': sha256_hash # TODO: provide in XML manifest
}
except requests.RequestException as e: # server not reachable, return empty dict
print(f"Manifest Url '{manifest_url}' not reachable, using code defaults!")
return app_info
"""
# lutzapps - not used anymore TODO: remove later
""" def update_app_configs():
app_info = fetch_app_info()
for app_name, info in app_info.items():
if app_name in app_configs:
app_configs[app_name].update(info)
app_configs[app_name].update(info) """
def get_app_configs():
def get_app_configs() -> dict:
return app_configs
def add_app_config(app_name, config):
@ -78,7 +258,8 @@ def remove_app_config(app_name):
del app_configs[app_name]
# Update app_configs when this module is imported
update_app_configs()
# lutzapps - not used anymore TODO: remove later
#update_app_configs()
### lutzapps section
@ -108,7 +289,7 @@ def write_dict_to_jsonfile(dict:dict, json_filepath:str, overwrite:bool=False) -
return True, "" # success
# helper function called by init_app_install_dirs(), init_shared_model_app_map(), init_shared_models_folders() and init_debug_settings()
def read_dict_from_jsonfile(json_filepath:str) -> dict:
def read_dict_from_jsonfile(json_filepath:str) -> tuple [dict, str]:
# Read JSON file from 'json_filepath' and return it as 'dict'
try:
@ -135,9 +316,9 @@ def pretty_dict(dict:dict) -> str:
return dict_string
# helper function for "init_app_install_dirs(), "init_shared_model_app_map()", "init_shared_models_folders()" and "inir_debug_settings()"
def init_global_dict_from_file(dict:dict, dict_filepath:str, dict_description:str, SHARED_MODELS_DIR:str="") -> bool:
# load or initialize the 'dict' for 'dict_description' from 'dict_filepath'
# helper function for "init_app_install_dirs(), "init_shared_model_app_map()", "init_shared_models_folders()" and "inir_DEBUG_SETTINGS()"
def load_global_dict_from_file(dict:dict, dict_filepath:str, dict_description:str, SHARED_MODELS_DIR:str="", write_file:bool=True) -> tuple[bool, dict]:
# returns the 'dict' for 'dict_description' from 'dict_filepath'
try:
if not SHARED_MODELS_DIR == "" and not os.path.exists(SHARED_MODELS_DIR):
@ -170,51 +351,40 @@ def init_global_dict_from_file(dict:dict, dict_filepath:str, dict_description:st
print(f"\nUsing {'external' if dict_filepath_found else 'default'} '{dict_description}':\n{pretty_dict(dict)}")
except Exception as e:
error_msg = f"ERROR in shared_models:init_global_dict_from_file() - initializing dict Map File '{dict_filepath}'\nException: {str(e)}"
print(error_msg)
print(f"ERROR in shared_models:load_global_dict_from_file() - initializing dict Map File '{dict_filepath}'\nException: {str(e)}")
return False, error_msg
return False, {}
return True, "" # success
return True, dict # success
DEBUG_SETTINGS_FILE = "/workspace/_debug_settings.json"
DEBUG_SETTINGS = {
# these setting will be READ:
"manifests": { # uncompressed sizes of the tar-files
"bcomfy": {
"venv_uncompressed_size": 6155283197,
"sha256_hash": ""
},
"ba1111": {
"venv_uncompressed_size": 6794355530,
"sha256_hash": ""
},
"bforge": {
"venv_uncompressed_size": 7689838771,
"sha256_hash": ""
},
"bkohya": {
"venv_uncompressed_size": 12192767148,
"sha256_hash": ""
}
},
"installer_codeversion": "2", # can be "1" (original) or "2" (fast)
"delete_tarfile_after_download": "1", # can be set to "0" to test only local unpack time and github setup
"use_bkohya_tar_folder_fix": "1", # the fix unpacks to "/workspace" and not to "/workspace/bkohya"
"use_bkohya_local_venv_symlink": "1", # when active, creates a folder symlink "venv" in "/workspace/kohya_ss" -> "/workspace/bkohya" VENV
"APP_CONFIGS_MANIFEST_URL": "", # this setting, when not blank, overwrites the global APP_CONFIGS_MANIFEST_URL
"installer_codeversion": "v2", # can be "v1" (original) or "v2" (fast)
"delete_tar_file_after_download": True, # can be set to True to test only local unpack time and github setup
"create_bkohya_to_local_venv_symlink": True, # when True, creates a folder symlink "venv" in "/workspace/kohya_ss" -> "/workspace/bkohya" VENV
"skip_to_github_stage": False, # when True, skip download and decompression stage and go directly to GH repo installation
# these settings will be WRITTEN:
"used_local_tar": "0", # works together with the above TAR local caching
"app_name": "",
"tar_filename": "",
"download_url": "",
"total_duration_download": "0",
"total_duration_unpack": "0",
"total_duration": "0"
"app_name": "", # last app_name the code run on
"used_local_tarfile": True, # works together with the above TAR local caching
"tar_filename": "", # last local tar_filename used
"download_url": "", # last used tar download_url
"total_duration_download": "00:00:00", # timespan-str "hh:mm:ss"
"total_duration_unpack": "00:00:00", # timespan-str "hh:mm:ss"
"total_duration": "00:00:00" # timespan-str "hh:mm:ss"
}
def init_debug_settings():
global DEBUG_SETTINGS
init_global_dict_from_file(DEBUG_SETTINGS, DEBUG_SETTINGS_FILE, "DEBUG_SETTINGS")
local_debug = os.environ.get('LOCAL_DEBUG', 'False') # support local browsing for development/debugging
generate_debug_settings_file = os.environ.get('DEBUG_SETTINGS_FILE', 'False') # generate the DEBUG_SETTINGS_FILE, if not exist already
write_file_if_not_exist = local_debug == 'True' or generate_debug_settings_file == 'True'
success, dict = load_global_dict_from_file(DEBUG_SETTINGS, DEBUG_SETTINGS_FILE, "DEBUG_SETTINGS", write_file=write_file_if_not_exist)
if success:
DEBUG_SETTINGS = dict
# read from DEBUG_SETTINGS
# installer_codeversion = DEBUG_SETTINGS['installer_codeversion'] # read from DEBUG_SETTINGS
@ -233,14 +403,60 @@ def write_debug_setting(setting_name:str, setting_value:str):
# lutzapps - init some settings from DEBUG_SETTINGS_FILE
init_debug_settings()
# lutzapps - add kohya_ss support and required local VENV
APP_CONFIGS_FILE = APP_CONFIGS_MANIFEST_URL # default is the online manifest url defined as "master"
# can be overwritten with DEBUG_SETTINGS['APP_CONFIGS_MANIFEST_URL'], e.g. point to "/workspace/_app_configs.json"
# # which is the file, that is generated when the ENV var LOCAL_DEBUG='True' or the ENV var APP_CONFIGS_FILE='True'
# NOTE: an existing serialized dict in the "/workspace" folder will never be overwritten agin from the code defaults,
# and "wins" against the code-defaults. So even changes in the source-code for this dicts will NOT be used,
# when a local file exists. The idea here is that it is possible to overwrite code-defaults.
# BUT as long as the APP_CONFIGS_MANIFEST_URL not gets overwritten, the global "app_configs" dict will be always loaded
# from the central S3 server, or whatever is defined.
# the only way to overwrite this url, is via the DEBUG_SETTINGS_FILE "/workspace/_debug_settings.json"
# the default source-code setting for DEBUG_SETTINGS['APP_CONFIGS_MANIFEST_URL']: "" (is an empty string),
# which still makes the default APP_CONFIGS_MANIFEST_URL the central master.
# only when this setting is not empty, it can win against the central url, but also only when the Url is valid (locally or remote)
# should there be an invalid Url (central or local), or any other problem, then the code-defaults will be used.
#
# The DEBUG_SETTINGS_FILE is a dict which helps during debugging, testing of APP Installations,
# and generating ENV TAR files.
# Is will also NOT be generated as external FILE, as long the same 2 ENV vars, which control the APP_CONFIGS_FILE generation are set:
# LOCAL_DEBUG='True' or APP_CONFIGS_FILE='True'
#
# SUMMARY: The DEBUG_SETTINGS and APP_CONFIGS (aka app_configs in code) will never be written to the /workspace,
# when the IMAGE is used normally.
def init_app_configs():
global APP_CONFIGS_FILE
global app_configs
# check for overwrite of APP_CONFIGS_MANIFEST_URL
if not DEBUG_SETTINGS['APP_CONFIGS_MANIFEST_URL'] == "":
APP_CONFIGS_FILE = DEBUG_SETTINGS['APP_CONFIGS_MANIFEST_URL']
local_debug = os.environ.get('LOCAL_DEBUG', 'False') # support local browsing for development/debugging
generate_app_configs_file = os.environ.get('APP_CONFIGS_FILE', 'False') # generate the APP_CONFIGS_FILE, if not exist already
write_file_if_not_exists = local_debug == 'True' or generate_app_configs_file == 'True'
success, dict = load_global_dict_from_file(app_configs, APP_CONFIGS_FILE, "APP_CONFIGS", write_file=write_file_if_not_exists)
if success:
app_configs = dict # overwrite code-defaults (from local or external settings)
#else app_configs = <code defaults already initialized>
return
init_app_configs() # load from JSON file (local or remote) with code-defaults otherwise
# lutzapps - add kohya_ss support and handle the required local "venv" within the "kohya_ss" app folder
def ensure_kohya_local_venv_is_symlinked() -> tuple[bool, str]:
### create a folder symlink for kohya's "local" 'venv' dir
# as kohya_ss' "setup.sh" assumes a "local" VENV under "/workspace/kohya_ss/venv",
# we will create a folder symlink "/workspace/kohya_ss/venv" -> "/workspace/bkohya"
# to our global VENV and rename the original "venv" folder to "venv(BAK)"
# to our global VENV and rename the original "venv" folder to "venv(BAK)", if any exists,
# will we not the case normally.
if not DEBUG_SETTINGS['use_bkohya_local_venv_symlink'] == "1":
return True, "" # not fix the local KOHYA_SS VENV
if not DEBUG_SETTINGS['create_bkohya_to_local_venv_symlink']:
return True, "" # not fix the local KOHYA_SS VENV requirement
import shutil
@ -251,13 +467,19 @@ def ensure_kohya_local_venv_is_symlinked() -> tuple[bool, str]:
bapp_app_path = app_configs[bapp_name]["app_path"] # '/workspace/kohya_ss'
bapp_app_path_venv = f"{bapp_app_path}/venv" # '/workspace/kohya_ss/venv'
name = app_configs[bapp_name]["name"]
if not os.path.exists(bapp_app_path): # kohya is not installed
return True, "" # no need to fix the local KOHYA VENV
return True, f"{name} is not installed." # no need to fix the local KOHYA VENV
# kohya installed and has a local "venv" folder
if os.path.exists(bapp_app_path_venv) and os.path.isdir(bapp_app_path_venv):
# check the src-folder of 'bkohya' downloaded VENV exists
if not os.path.exists(bapp_venv_path): # src_path to bkohya downloaded venv does NOT exists
return True, f"{name} VENV is not installed." # no need to fix the local KOHYA VENV, as the global KOHYA VENV does not exist
# check if this local VENV is a folderlink to target our bkohya global VENV to venv_path
# kohya_ss is installed
if os.path.isdir(bapp_app_path_venv): # and has a local "venv" folder
# check if this local VENV is a folderlink to target the bkohya global VENV to venv_path
if os.path.islink(bapp_app_path_venv):
success_message = f"kohya_ss local venv folder '{bapp_app_path_venv}' is already symlinked"
@ -273,17 +495,16 @@ def ensure_kohya_local_venv_is_symlinked() -> tuple[bool, str]:
i += 1
suffix = str(i)
bak_venv_path += suffix # free target bame for "rename"
bak_venv_path += suffix # free target name for "rename"(move) operation of the folder
shutil.move(bapp_app_path_venv, bak_venv_path) # move=rename
print(f"local venv folder '{bapp_app_path_venv}' detected and renamed to '{bak_venv_path}'")
# now the path to the local "venv" is free, if it was already created it is now renamed
### create a folder symlink for kohya's "local" venv dir
# check the src-folder to kohya downloaded venv exists
if os.path.exists(bapp_venv_path): # src_path to bkohya downloaded venv exists
# create a folder symlink for kohya local venv dir
os.symlink(bapp_venv_path, bapp_app_path_venv, target_is_directory=True)
success_message = f"created a symlink for kohya_ss local venv folder: '{bapp_venv_path}' -> '{bapp_app_path_venv}'"
success_message = f"created a symlink for kohya_ss local venv folder: '{bapp_app_path_venv}' -> '{bapp_venv_path}'"
print(success_message)
return True, success_message
@ -296,3 +517,41 @@ def ensure_kohya_local_venv_is_symlinked() -> tuple[bool, str]:
# lutzapps - add kohya_ss venv support
ensure_kohya_local_venv_is_symlinked()
# some verification steps of the VENV setup of the "kohya_ss" app:
# even if it "looks" like the "venv" is in a local sub-folder of the "kohya_ss" dir,
# this location is only "aliased/symlinked" there from the globally downloaded
# tarfile "bkohya.tar.gz" which was expanded spearately into the folder "/workspace/bkohya".
# So the VENV can be redownloaded separately from the github app at "/workspace/kohya_ss"
# root@9452ad7f4cd6:/workspace/kohya_ss# python --version
# Python 3.11.10
# root@fe889cc68f5a:/workspace/kohya_ss# pip --version
# pip 24.3.1 from /usr/local/lib/python3.11/dist-packages/pip (python 3.11)
#
# root@9452ad7f4cd6:/workspace/kohya_ss# python3 --version
# Python 3.11.10
# root@fe889cc68f5a:/workspace/kohya_ss# pip3 --version
# pip 24.3.1 from /usr/local/lib/python3.11/dist-packages/pip (python 3.11)
#
# root@9452ad7f4cd6:/workspace/kohya_ss# ls venv -la
# lrwxr-xr-x 1 root root 17 Nov 8 00:06 venv -> /workspace/bkohya
#
# root@9452ad7f4cd6:/workspace/kohya_ss# source venv/bin/activate
#
# (bkohya) root@9452ad7f4cd6:/workspace/kohya_ss# ls venv/bin/python* -la
# lrwxr-xr-x 1 root root 10 Nov 8 00:48 venv/bin/python -> python3.10
# lrwxr-xr-x 1 root root 10 Nov 8 00:48 venv/bin/python3 -> python3.10
# lrwxr-xr-x 1 root root 19 Nov 8 00:48 venv/bin/python3.10 -> /usr/bin/python3.10
#
# (bkohya) root@9452ad7f4cd6:/workspace/kohya_ss# python --version
# Python 3.10.12
# (bkohya) root@fe889cc68f5a:/workspace/kohya_ss# pip --version
# pip 22.0.2 from /workspace/venv/lib/python3.10/site-packages/pip (python 3.10)
#
# (bkohya) root@9452ad7f4cd6:/workspace/kohya_ss# python3 --version
# Python 3.10.12
# (bkohya) root@fe889cc68f5a:/workspace/kohya_ss# pip3 --version
# pip 22.0.2 from /workspace/venv/lib/python3.10/site-packages/pip (python 3.10)
#
# (bkohya) root@9452ad7f4cd6:/workspace/kohya_ss# deactivate
# root@9452ad7f4cd6:/workspace/kohya_ss#

View file

@ -13,6 +13,8 @@ import xml.etree.ElementTree as ET
import time
import datetime
import shutil
from utils.app_configs import (DEBUG_SETTINGS, pretty_dict, init_app_configs, init_debug_settings, write_debug_setting, ensure_kohya_local_venv_is_symlinked)
from utils.model_utils import (get_sha256_hash_from_file)
INSTALL_STATUS_FILE = '/tmp/install_status.json'
@ -180,41 +182,45 @@ import time
# yield (out_line.rstrip(), err_line.rstrip())
# this ist the v2 ("fast") version for "download_and_unpack_venv()" - can be (de-)/activated in DEBUG_SETTINGS dict
def download_and_unpack_venv_fastversion(app_name, app_configs, send_websocket_message) -> tuple[bool, str]:
# this is the v2 ("fast") version for "download_and_unpack_venv()" - can be (de-)/activated in DEBUG_SETTINGS dict
def download_and_unpack_venv_v2(app_name:str, app_configs:dict, send_websocket_message) -> tuple[bool, str]:
# load the latest configured DEBUG_SETTINGS from the stored setting of the DEBUG_SETTINGS_FILE
init_debug_settings() # reload latest DEBUG_SETTINGS
# as this could overwrite the APP_CONFIGS_MANIFEST_URL, we reload the app_configs global dict
# from whatever Url is now defined
init_app_configs() # reload lastest app_configs dict
app_config = app_configs.get(app_name)
if not app_config:
return False, f"App '{app_name}' not found in configurations."
venv_path = app_config['venv_path']
app_path = app_config['app_path']
download_url = app_config['download_url']
total_size = app_config['size']
archive_size = app_config['archive_size']
tar_filename = os.path.basename(download_url)
workspace_dir = '/workspace'
downloaded_file = os.path.join(workspace_dir, tar_filename)
from utils.app_configs import (DEBUG_SETTINGS, pretty_dict, init_debug_settings, write_debug_setting, ensure_kohya_local_venv_is_symlinked)
# load the latest configured DEBUG_SETTINGS from the stored setting of the DEBUG_SETTINGS_FILE
init_debug_settings()
# show currently using DEBUG_SETTINGS
print(f"\nCurrently using 'DEBUG_SETTINGS':\n{pretty_dict(DEBUG_SETTINGS)}")
write_debug_setting('tar_filename', tar_filename)
write_debug_setting('download_url', download_url)
try:
if DEBUG_SETTINGS['skip_to_github_stage']:
success, message = clone_application(app_config,send_websocket_message)
return success, message
save_install_status(app_name, 'in_progress', 0, 'Downloading')
send_websocket_message('install_log', {'app_name': app_name, 'log': f'Downloading {total_size / (1024 * 1024):.2f} MB ...'})
send_websocket_message('install_log', {'app_name': app_name, 'log': f'Downloading {archive_size / (1024 * 1024):.2f} MB ...'})
start_time_download = time.time()
# debug with existing local cached TAR file
if os.path.exists(downloaded_file):
write_debug_setting('used_local_tar', "1") # indicate using cached TAR file
send_websocket_message('install_log', {'app_name': app_name, 'log': f"Used cached local tarfile '{downloaded_file}'"})
write_debug_setting('used_local_tarfile', True) # indicate using cached TAR file
send_websocket_message('used_local_tarfile', {'app_name': app_name, 'log': f"Used cached local tarfile '{downloaded_file}'"})
else:
write_debug_setting('used_local_tar', "0") # indicate no cached TAR file found
write_debug_setting('used_local_tarfile', False) # indicate no cached TAR file found
try: ### download with ARIA2C
@ -275,8 +281,8 @@ def download_and_unpack_venv_fastversion(app_name, app_configs, send_websocket_m
gid = match.group(1) # e.g., "cd57da"
downloaded_size_value = match.group(2) # e.g., "2.1"
downloaded_size_unit = match.group(3) # e.g., "GiB"
total_size_value = match.group(4) # e.g., "4.0"
total_size_unit = match.group(5) # e.g., "GiB"
total_size_value = match.group(4) # e.g., "4.0" (this could replace the 'archive_size' from the manifest)
total_size_unit = match.group(5) # e.g., "GiB" (with calculation to bytes, but not sure if its rounded)
percentage = int(match.group(6)) # e.g., "53"
connection_count = int(match.group(7)) # e.g., "16"
download_rate_value = match.group(8) # e.g., "1.9"
@ -296,8 +302,8 @@ def download_and_unpack_venv_fastversion(app_name, app_configs, send_websocket_m
### original code
#speed = downloaded_size / elapsed_time # bytes/sec
#percentage = (downloaded_size / total_size) * 100
#eta = (total_size - downloaded_size) / speed if speed > 0 else 0 # sec
#percentage = (downloaded_size / archive_size) * 100
#eta = (archive_size - downloaded_size) / speed if speed > 0 else 0 # sec
send_websocket_message('install_progress', {
'app_name': app_name,
@ -329,7 +335,7 @@ def download_and_unpack_venv_fastversion(app_name, app_configs, send_websocket_m
os.remove(f"{tar_filename}.aria2")
except Exception as e:
error_msg = f"ERROR in download_and_unpack_venv_fastversion():download with ARIA2C\ncmdline: '{cmd_line}'\nException: {str(e)}"
error_msg = f"ERROR in download_and_unpack_venv_v2():download with ARIA2C\ncmdline: '{cmd_line}'\nException: {str(e)}"
print(error_msg)
error_message = f"Downloading VENV failed: {download_process.stderr.read() if download_process.stderr else 'Unknown error'}"
@ -356,8 +362,8 @@ def download_and_unpack_venv_fastversion(app_name, app_configs, send_websocket_m
# if elapsed_time > 0:
# speed = downloaded_size / elapsed_time
# percentage = (downloaded_size / total_size) * 100
# eta = (total_size - downloaded_size) / speed if speed > 0 else 0
# percentage = (downloaded_size / archive_size) * 100
# eta = (archive_size - downloaded_size) / speed if speed > 0 else 0
# send_websocket_message('install_progress', {
# 'app_name': app_name,
@ -375,29 +381,72 @@ def download_and_unpack_venv_fastversion(app_name, app_configs, send_websocket_m
return False, error_message
send_websocket_message('install_log', {'app_name': app_name, 'log': 'Download completed. Starting unpacking...'})
send_websocket_message('install_progress', {'app_name': app_name, 'percentage': 100, 'stage': 'Download Complete'})
send_websocket_message('install_log', {'app_name': app_name, 'log': 'Download completed. Starting Verification ...'})
# we use a 99% progress and indicate 1% for Verification against the files SHA256 hash
send_websocket_message('install_progress', {'app_name': app_name, 'percentage': 99, 'stage': 'Downloading'})
total_duration_download = f"{datetime.timedelta(seconds=int(time.time() - start_time_download))}"
write_debug_setting('total_duration_download', total_duration_download)
print(f"download did run {total_duration_download} for app '{app_name}'")
### VERIFY stage
#
# Create TAR from the VENV current directory:
# IMPORTANT: cd INTO the folder you want to compress, as we use "." for source folder,
# to avoid having the foldername in the TAR file !!!
# PV piping is "nice-to-have" and is only used for showing "Progress Values" during compressing
#
# cd /workspace/bkohya
# #tar -czf | pv > /workspace/bkohya.tar.gz . (not the smallest TAR)#
# tar -cvf - . | gzip -9 - | pv > /workspace/bkohya.tar.gz
#
# afterwards create the SHA256 hash from this TAR with
# shasum -a 256 bkohya.tar.gz
#
# also report the uncompressed size from the current VENV directory,
# we need that as the 100% base for the progress indicators when uncompressing the TAR
# verify the downloaded TAR file against its SHA256 hash value from the manifest
download_sha256_hash = app_config["sha256_hash"].lower() # get the sha256_hash from the app manifest
file_verified = False
print(f"getting SHA256 Hash for '{downloaded_file}'")
successfull_HashGeneration, file_sha256_hash = get_sha256_hash_from_file(downloaded_file)
if successfull_HashGeneration and file_sha256_hash.lower() == download_sha256_hash.lower():
file_verified = True
message = f"Downloaded file '{os.path.basename(downloaded_file)}' was successfully (SHA256) verified."
print(message)
else:
if successfull_HashGeneration: # the generated SHA256 file hash did not match against the metadata hash
error_message = f"The downloaded file '{os.path.basename(downloaded_file)}' has DIFFERENT \nSHA256: {file_sha256_hash} as in the manifest\nFile is possibly corrupted and was DELETED!"
print(error_message)
os.remove(downloaded_file) # delete corrupted, downloaded file
else: # NOT successful, the hash contains the Exception
error_msg = file_sha256_hash
error_message = f"Exception occured while generating the SHA256 hash for '{downloaded_file}':\n{error_msg}"
print(error_message)
if not file_verified:
send_websocket_message('install_complete', {'app_name': app_name, 'status': 'error', 'message': error_message})
save_install_status(app_name, 'failed', 0, 'Failed')
return False, error_message
send_websocket_message('install_log', {'app_name': app_name, 'log': 'Verification completed. Starting unpacking ...'})
send_websocket_message('install_progress', {'app_name': app_name, 'percentage': 100, 'stage': 'Download Complete'})
### Decompression Stage (Unpacking the downloaded VENV)
start_time_unpack = time.time()
# lutzapps - fix TAR packaging bug (compressed from the workspace root instead of bkohya VENV folder)
# e.g. "bkohya/bin/activate", together with venv_path ("/workspace/bkohya") ends up as "/workspace/bkohya/bkohya/bin/activate"
# TODO: need to repackage Kohya VENV correctly and then remove this fix!!!
if app_name == "bkohya" and DEBUG_SETTINGS['use_bkohya_tar_folder_fix'] == "1":
venv_path = "/workspace" # extracts then correctly to '/workspace/bkohya', instead of '/workspace/bkohya/bkohya'
# Create TAR from the VENV current directory:
# cd ~/Projects/Docker/madiator/workspace/bkohya
# [tar -czf | pv > ~/Projects/Docker/madiator/workspace/bkohya.tar.gz . (not the smallest TAR)]
# tar -cvf - . | gzip -9 - | pv > ~/Projects/Docker/madiator/workspace/bkohya.tar.gz
# Ensure the venv directory exists
os.makedirs(f"{venv_path}/", exist_ok=True) # append trailing "/" to make sure the last sub-folder is created
@ -419,10 +468,7 @@ def download_and_unpack_venv_fastversion(app_name, app_configs, send_websocket_m
# 'bforge': 7689838771
# 'bkohya': 12192767148
uncompressed_size_bytes = DEBUG_SETTINGS["manifests"][app_name]["venv_uncompressed_size"]
#sha256_hash = DEBUG_SETTINGS["manifests"][app_name]["sha256_hash"]
# TODO: create with 'shasum -a 256 xxx.tar.gz'
uncompressed_size_bytes = app_config["venv_uncompressed_size"]
### NOTE: as it turns out GZIP has problems with files bigger than 2 or 4 GB due to internal field bit restrictions
@ -601,7 +647,7 @@ def download_and_unpack_venv_fastversion(app_name, app_configs, send_websocket_m
# else any other line in stdout (which we not process)
except Exception as e:
error_msg = f"ERROR in download_and_unpack_venv_fastversion():\ncmdline: '{cmd_line}'\nException: {str(e)}"
error_msg = f"ERROR in download_and_unpack_venv_v2():\ncmdline: '{cmd_line}'\nException: {str(e)}"
print(error_msg)
decompression_process.wait() # let the process finish
@ -621,8 +667,11 @@ def download_and_unpack_venv_fastversion(app_name, app_configs, send_websocket_m
send_websocket_message('install_progress', {'app_name': app_name, 'percentage': 100, 'stage': 'Unpacking Complete'})
print(f"'DEBUG_SETTINGS' after this run:\n{pretty_dict(DEBUG_SETTINGS)}")
### installing the App from GITHUB
# Clone the repository if it doesn't exist
success, message = clone_application(app_name)
print(f"'DEBUG_SETTINGS' after this run:\n{pretty_dict(DEBUG_SETTINGS)}")
### original "v1" code (very slow code because of STATISTICS glory
@ -647,66 +696,29 @@ def download_and_unpack_venv_fastversion(app_name, app_configs, send_websocket_m
# process.wait()
# rc = process.returncode
### installing the App from GITHUB
# Clone the repository if it doesn't exist
if not os.path.exists(app_path):
send_websocket_message('install_log', {'app_name': app_name, 'log': 'Cloning repository...'})
repo_url = ''
if app_name == 'bcomfy':
repo_url = 'https://github.com/comfyanonymous/ComfyUI.git'
elif app_name == 'bforge':
repo_url = 'https://github.com/lllyasviel/stable-diffusion-webui-forge.git'
elif app_name == 'ba1111':
repo_url = 'https://github.com/AUTOMATIC1111/stable-diffusion-webui.git'
elif app_name == 'bkohya': # lutzapps - added new Kohya app
repo_url = 'https://github.com/bmaltais/kohya_ss.git'
try: # add a repo assignment for Kohya
repo = git.Repo.clone_from(repo_url, app_path, progress=lambda op_code, cur_count, max_count, message: send_websocket_message('install_log', {
'app_name': app_name,
'log': f"Cloning: {cur_count}/{max_count} {message}"
}))
send_websocket_message('install_log', {'app_name': app_name, 'log': 'Repository cloned successfully.'})
# lutzapps - make sure we use Kohya with FLUX support
if app_name == 'bkohya':
branch_name = "sd3-flux.1" # this branch also uses a "sd-scripts" branch "SD3" automatically
repo.git.checkout(branch_name)
# Clone ComfyUI-Manager for Better ComfyUI
if app_name == 'bcomfy':
custom_nodes_path = os.path.join(app_path, 'custom_nodes')
os.makedirs(custom_nodes_path, exist_ok=True)
comfyui_manager_path = os.path.join(custom_nodes_path, 'ComfyUI-Manager')
if not os.path.exists(comfyui_manager_path):
send_websocket_message('install_log', {'app_name': app_name, 'log': 'Cloning ComfyUI-Manager...'})
git.Repo.clone_from('https://github.com/ltdrdata/ComfyUI-Manager.git', comfyui_manager_path)
send_websocket_message('install_log', {'app_name': app_name, 'log': 'ComfyUI-Manager cloned successfully.'})
except git.exc.GitCommandError as e:
send_websocket_message('install_log', {'app_name': app_name, 'log': f'Error cloning repository: {str(e)}'})
return False, f"Error cloning repository: {str(e)}"
if app_name == 'bkohya': # create a folder link for kohya_ss local "venv"
ensure_kohya_local_venv_is_symlinked()
success, error_message = clone_application(app_name, send_websocket_message)
# Clean up the downloaded file
send_websocket_message('install_log', {'app_name': app_name, 'log': 'Cleaning up...'})
# lutzapps - debug with local TAR
# do NOT delete the Kohya venv
if DEBUG_SETTINGS["delete_tarfile_after_download"] == "1": # this is the default, but can be overwritten
if DEBUG_SETTINGS["delete_tar_file_after_download"]: # this is the default, but can be overwritten
os.remove(downloaded_file)
send_websocket_message('install_log', {'app_name': app_name, 'log': 'Installation complete. Refresh page to start app'})
if success:
save_install_status(app_name, 'completed', 100, 'Completed')
send_websocket_message('install_complete', {'app_name': app_name, 'status': 'success', 'message': "Virtual environment installed successfully."})
return True, "Virtual environment installed successfully."
else:
return False, error_message
except requests.RequestException as e:
error_message = f"Download failed: {str(e)}"
error_message = f"Download/Decompression failed: {str(e)}"
send_websocket_message('install_complete', {'app_name': app_name, 'status': 'error', 'message': error_message})
save_install_status(app_name, 'failed', 0, 'Failed')
return False, error_message
@ -716,8 +728,143 @@ def download_and_unpack_venv_fastversion(app_name, app_configs, send_websocket_m
send_websocket_message('install_complete', {'app_name': app_name, 'status': 'error', 'message': error_message})
return False, error_message
### installing the App from GITHUB
# Clone the repository if it doesn't exist
def clone_application(app_config:dict, send_websocket_message) -> tuple[bool , str]:
try:
app_name = app_config['id']
app_path = app_config['app_path']
def download_and_unpack_venv(app_name, app_configs, send_websocket_message):
if not os.path.exists(app_path): # only install new apps
repo_url = app_config['repo_url']
branch_name = app_config['branch_name']
if branch_name == "": # use the default branch
branch_name = "master"
clone_recursive = app_config['clone_recursive']
send_websocket_message('install_log', {'app_name': app_name, 'log': f"Cloning repository '{repo_url}' branch '{branch_name}' recursive={clone_recursive} ..."})
repo = git.Repo.clone_from(repo_url, app_path, # first 2 params are fix, then use named params
#branch=branch_name, # if we provide a branch here, we ONLY get this branch downloaded
# we want ALL branches, so we can easy checkout different versions from kohya_ss late, without re-downloading
recursive=clone_recursive, # include cloning submodules recursively (if needed as with Kohya)
progress=lambda op_code, cur_count, max_count, message: send_websocket_message('install_log', {
'app_name': app_name,
'log': f"Cloning: {cur_count}/{max_count} {message}"
}))
send_websocket_message('install_log', {'app_name': app_name, 'log': 'Repository cloned successfully.'})
# lutzapps - make sure we use Kohya with FLUX support
if not branch_name == "master":
repo.git.checkout(branch_name) # checkout the "sd3-flux.1" branch, but could later switch back to "master" easy
# the setup can be easy verified with git, here e.g. for the "kohya_ss" app:
# root@fe889cc68f5a:~# cd /workspace/kohya_ss
# root@fe889cc68f5a:/workspace/kohya_ss# git branch
# master
# * sd3-flux.1
# root@fe889cc68f5a:/workspace/kohya_ss# cd sd-scripts
# root@fe889cc68f5a:/workspace/kohya_ss/sd-scripts# git branch
# * (HEAD detached at b8896aa)
# main
#
# in the case of kohya_ss we need to fix a bug in the 'setup.sh' file,
# where they forgot to adapt the branch name from "master" to "sd3-flux.1"
# in the "#variables" section for refreshing kohya via git with 'setup.sh'
if app_name == 'bkohya':
success, message = update_kohya_setup_sh(app_path) # patch the 'setup.sh' file
print(message) # shows, if the patch was needed, and apllied successfully
else: # refresh app
if app_path['refresh']: # app wants auto-refreshes
# TODO: implement app refreshes via git pull or, in the case of 'kohya_ss' via "setup.sh"
message = f"Refreshing of app '{app_name}' is NYI"
print(message)
# Clone ComfyUI-Manager and other defined custom_nodes for Better ComfyUI
if app_name == 'bcomfy':
# install all defined custom nodes
custom_nodes_path = os.path.join(app_path, 'custom_nodes')
os.makedirs(f"{custom_nodes_path}/", exist_ok=True) # append a trailing slash to be sure last dir is created
for custom_node in app_config['custom_nodes']:
name = custom_node['name']
path = custom_node['path']
repo_url = custom_node['repo_url']
custom_node_path = os.path.join(custom_nodes_path, path)
if not os.path.exists(custom_node_path): # only install new custom nodes
send_websocket_message('install_log', {'app_name': app_name, 'log': f"Cloning '{name}' ..."})
git.Repo.clone_from(repo_url, custom_node_path)
send_websocket_message('install_log', {'app_name': app_name, 'log': f"'{name}' cloned successfully."})
# install requirements
venv_path = app_config['venv_path']
#app_path = app_config['app_path'] # already defined
try:
# Activate the virtual environment and run the commands
activate_venv = f"source {venv_path}/bin/activate"
change_dir_command = f"cd {custom_node_path}"
pip_install_command = "pip install -r requirements.txt"
full_command = f"{activate_venv} && {change_dir_command} && {pip_install_command}"
# TODO: rewrite this without shell
process = subprocess.Popen(full_command, shell=True, stdout=subprocess.PIPE, stderr=subprocess.STDOUT, executable='/bin/bash')
output, _ = process.communicate()
if process.returncode == 0:
return True, f"Custom node requirements were successfully installed. Output: {output.decode('utf-8')}"
else:
return False, f"Error in custom node requirements installation. Output: {output.decode('utf-8')}"
except Exception as e:
return False, f"Error installing custom node requirements: {str(e)}"
except git.exc.GitCommandError as e:
send_websocket_message('install_log', {'app_name': app_name, 'log': f'Error cloning repository: {str(e)}'})
return False, f"Error cloning repository: {str(e)}"
except Exception as e:
send_websocket_message('install_log', {'app_name': app_name, 'log': f'Error cloning repository: {str(e)}'})
return False, f"Error cloning repository: {str(e)}"
if app_name == 'bkohya': # create a folder link for kohya_ss local "venv"
success, message = ensure_kohya_local_venv_is_symlinked()
if not success: # symlink not created, but still success=True and only a warning, can be fixed manually
message = f"{app_config['name']} was cloned and patched successfully, but the symlink to the local venv returned following problem:\n{message}"
else:
message = f"'{app_name}' was cloned successfully."
return True, message
def update_kohya_setup_sh(app_path:str) -> tuple[bool, str]:
try:
# patch 'setup.sh' within the kohya_ss main folder for BRANCH="sd3-flux.1"
setup_sh_path = os.path.join(app_path, 'setup.sh')
if not os.path.exists(setup_sh_path):
return False, f"file '{setup_sh_path}' was not found"
with open(setup_sh_path, 'r') as file:
content = file.read()
# Use regex to search & replace wrong branch variable in the file
patched_content = re.sub(r'BRANCH="master"', 'BRANCH="sd3-flux.1"', content)
if patched_content == content:
message = f"'{setup_sh_path}' already fine, patch not needed."
else:
with open(setup_sh_path, 'w') as file:
file.write(patched_content)
message = f"'{setup_sh_path}' needed patch, successfully patched."
return True, message
except Exception as e:
return False, str(e)
def download_and_unpack_venv_v1(app_name, app_configs, send_websocket_message):
app_config = app_configs.get(app_name)
if not app_config:
return False, f"App '{app_name}' not found in configurations."
@ -725,14 +872,14 @@ def download_and_unpack_venv(app_name, app_configs, send_websocket_message):
venv_path = app_config['venv_path']
app_path = app_config['app_path']
download_url = app_config['download_url']
total_size = app_config['size']
archive_size = app_config['size']
tar_filename = os.path.basename(download_url)
workspace_dir = '/workspace'
downloaded_file = os.path.join(workspace_dir, tar_filename)
try:
save_install_status(app_name, 'in_progress', 0, 'Downloading')
send_websocket_message('install_log', {'app_name': app_name, 'log': f'Starting download of {total_size / (1024 * 1024):.2f} MB...'})
send_websocket_message('install_log', {'app_name': app_name, 'log': f'Starting download of {archive_size / (1024 * 1024):.2f} MB...'})
# lutzapps - debug with existing local TAR
if not os.path.exists(downloaded_file):
@ -753,8 +900,8 @@ def download_and_unpack_venv(app_name, app_configs, send_websocket_message):
if elapsed_time > 0:
speed = downloaded_size / elapsed_time
percentage = (downloaded_size / total_size) * 100
eta = (total_size - downloaded_size) / speed if speed > 0 else 0
percentage = (downloaded_size / archive_size) * 100
eta = (archive_size - downloaded_size) / speed if speed > 0 else 0
send_websocket_message('install_progress', {
'app_name': app_name,
@ -869,31 +1016,34 @@ def download_and_unpack_venv(app_name, app_configs, send_websocket_message):
send_websocket_message('install_complete', {'app_name': app_name, 'status': 'error', 'message': error_message})
return False, error_message
### this is the function wgich switches between v0 and v1 debug setting for comparison
def download_and_unpack_venv(app_name, app_configs, send_websocket_message) -> tuple[bool, str]:
from app_configs import DEBUG_SETTINGS, write_debug_setting
### this is the function which switches between v0 and v1 debug setting for comparison
def download_and_unpack_venv(app_name:str, app_configs:dict, send_websocket_message) -> tuple[bool, str]:
from utils.app_configs import DEBUG_SETTINGS, write_debug_setting
installer_codeversion = DEBUG_SETTINGS['installer_codeversion'] # read from DEBUG_SETTINGS
print(f"download_and_unpack_venv v{installer_codeversion} STARTING for '{app_name}'")
print(f"download_and_unpack_venv_{installer_codeversion} STARTING for '{app_name}'")
import time
start_time = time.time()
if installer_codeversion == "1":
download_and_unpack_venv(app_name, app_configs, send_websocket_message)
elif installer_codeversion == "2":
download_and_unpack_venv_fastversion(app_name, app_configs, send_websocket_message)
if installer_codeversion == "v1":
success, message = download_and_unpack_venv_v1(app_name, app_configs, send_websocket_message)
elif installer_codeversion == "v2":
success, message = download_and_unpack_venv_v2(app_name, app_configs, send_websocket_message)
else:
print(f"unknown 'installer_codeversion' v{installer_codeversion} found, nothing run for app '{app_name}'")
error_msg = f"unknown 'installer_codeversion' {installer_codeversion} found, nothing run for app '{app_name}'"
print(error_msg)
success = False
message = error_msg
total_duration = f"{datetime.timedelta(seconds=int(time.time() - start_time))}"
write_debug_setting('app_name', app_name)
write_debug_setting('total_duration', total_duration)
print(f"download_and_unpack_venv v{installer_codeversion} did run {total_duration} for app '{app_name}'")
print(f"download_and_unpack_venv_v{installer_codeversion} did run {total_duration} for app '{app_name}'")
return success, message
def fix_custom_nodes(app_name, app_configs):
if app_name != 'bcomfy':
@ -921,54 +1071,9 @@ def fix_custom_nodes(app_name, app_configs):
return False, f"Error fixing custom nodes: {str(e)}"
# Replace the existing install_app function with this updated version
def install_app(app_name, app_configs, send_websocket_message):
def install_app(app_name:str, app_configs:dict, send_websocket_message) -> tuple[bool, str]:
if app_name in app_configs:
#return download_and_unpack_venv(app_name, app_configs, send_websocket_message)
return download_and_unpack_venv_fastversion(app_name, app_configs, send_websocket_message)
success, message = download_and_unpack_venv(app_name, app_configs, send_websocket_message)
return success, message
else:
return False, f"Unknown app: {app_name}"
# unused function
def onsolate_update_model_symlinks():
# lutzapps - CHANGE #7 - use the new "shared_models" module for app model sharing
# remove this whole now unused function
return "replaced by utils.shared_models.update_model_symlinks()"
shared_models_dir = '/workspace/shared_models'
apps = {
'stable-diffusion-webui': '/workspace/stable-diffusion-webui/models',
'stable-diffusion-webui-forge': '/workspace/stable-diffusion-webui-forge/models',
'ComfyUI': '/workspace/ComfyUI/models'
}
model_types = ['Stable-diffusion', 'VAE', 'Lora', 'ESRGAN']
for model_type in model_types:
shared_model_path = os.path.join(shared_models_dir, model_type)
if not os.path.exists(shared_model_path):
continue
for app, app_models_dir in apps.items():
if app == 'ComfyUI':
if model_type == 'Stable-diffusion':
app_model_path = os.path.join(app_models_dir, 'checkpoints')
elif model_type == 'Lora':
app_model_path = os.path.join(app_models_dir, 'loras')
elif model_type == 'ESRGAN':
app_model_path = os.path.join(app_models_dir, 'upscale_models')
else:
app_model_path = os.path.join(app_models_dir, model_type.lower())
else:
app_model_path = os.path.join(app_models_dir, model_type)
# Create the app model directory if it doesn't exist
os.makedirs(app_model_path, exist_ok=True)
# Create symlinks for each file in the shared model directory
for filename in os.listdir(shared_model_path):
src = os.path.join(shared_model_path, filename)
dst = os.path.join(app_model_path, filename)
if os.path.isfile(src) and not os.path.exists(dst):
os.symlink(src, dst)
print("Model symlinks updated.")

View file

@ -71,7 +71,7 @@ def check_huggingface_url(url):
return True, repo_id, filename, folder_name, branch_name
def download_model(url, model_name, model_type, civitai_token=None, hf_token=None, version_id=None, file_index=None):
def download_model(url, model_name, model_type, civitai_token=None, hf_token=None, version_id=None, file_index=None) -> tuple[bool, str]:
ensure_shared_folder_exists()
is_civitai, is_civitai_api, model_id, _ = check_civitai_url(url)
is_huggingface, repo_id, hf_filename, hf_folder_name, hf_branch_name = check_huggingface_url(url) # TODO: double call
@ -95,7 +95,7 @@ def download_model(url, model_name, model_type, civitai_token=None, hf_token=Non
return success, message
# lutzapps - added SHA256 checks for already existing ident and downloaded HuggingFace model
def download_civitai_model(url, model_name, model_type, civitai_token, version_id=None, file_index=None):
def download_civitai_model(url, model_name, model_type, civitai_token, version_id=None, file_index=None) -> tuple[bool, str]:
try:
is_civitai, is_civitai_api, model_id, url_version_id = check_civitai_url(url)
@ -186,7 +186,7 @@ def get_sha256_hash_from_file(file_path:str) -> tuple[bool, str]:
for byte_block in iter(lambda: f.read(4096), b""):
sha256_hash.update(byte_block)
return True, sha256_hash.hexdigest().upper()
return True, sha256_hash.hexdigest().lower()
except Exception as e:
return False, str(e)
@ -247,7 +247,7 @@ def get_modelfile_hash_and_ident_existing_modelfile_exists(model_name:str, model
raise NotImplementedError("Copying a non-LFS file is not implemented.")
lfs = repo_file.lfs # BlobLfsInfo class instance
download_sha256_hash = lfs.sha256.upper()
download_sha256_hash = lfs.sha256.lower()
print(f"Metadata from RepoFile LFS '{repo_file.rfilename}'")
print(f"SHA256: {download_sha256_hash}")
@ -283,8 +283,8 @@ def get_modelfile_hash_and_ident_existing_modelfile_exists(model_name:str, model
# if NOT successful, the hash contains the Exception
print(f"SHA256 hash generated from local file: '{model_path}'\n{model_sha256_hash}")
if successfull_HashGeneration and model_sha256_hash == download_sha256_hash:
message = f"Existing and ident model aleady found for '{os.path.basename(model_path)}'"
if successfull_HashGeneration and model_sha256_hash.lower() == download_sha256_hash.lower():
message = f"Existing and ident model already found for '{os.path.basename(model_path)}'"
print(message)
send_websocket_message('model_download_progress', {
@ -315,7 +315,7 @@ def get_modelfile_hash_and_ident_existing_modelfile_exists(model_name:str, model
# lutzapps - added SHA256 checks for already existing ident and downloaded HuggingFace model
def download_huggingface_model(url, model_name, model_type, repo_id, hf_filename, hf_folder_name, hf_branch_name, hf_token=None):
def download_huggingface_model(url, model_name, model_type, repo_id, hf_filename, hf_folder_name, hf_branch_name, hf_token=None) -> tuple[bool, str]:
try:
from huggingface_hub import hf_hub_download
@ -372,7 +372,7 @@ def download_huggingface_model(url, model_name, model_type, repo_id, hf_filename
# lutzapps - added SHA256 check for downloaded CivitAI model
def download_file(url, download_sha256_hash, file_path, headers=None):
def download_file(url, download_sha256_hash, file_path, headers=None) -> tuple[bool, str]:
try:
response = requests.get(url, stream=True, headers=headers)
response.raise_for_status()
@ -428,7 +428,7 @@ def check_downloaded_modelfile(model_path:str, download_sha256_hash:str, platfor
})
successfull_HashGeneration, model_sha256_hash = get_sha256_hash_from_file(model_path)
if successfull_HashGeneration and model_sha256_hash == download_sha256_hash:
if successfull_HashGeneration and model_sha256_hash.lower() == download_sha256_hash.lower():
send_websocket_message('model_download_progress', {
'percentage': 100,
'stage': 'Complete',

View file

@ -6,7 +6,7 @@ import time
from flask import jsonify
from utils.websocket_utils import (send_websocket_message, active_websockets)
from utils.app_configs import (get_app_configs, init_global_dict_from_file, pretty_dict)
from utils.app_configs import (get_app_configs, load_global_dict_from_file, pretty_dict)
### shared_models-v0.9.2 by lutzapps, Nov 5th 2024 ###
@ -189,7 +189,9 @@ SHARED_MODEL_FOLDERS = {
# helper function called by "inline"-main() and ensure_shared_models_folders()
def init_shared_models_folders(send_SocketMessage:bool=True):
global SHARED_MODEL_FOLDERS
init_global_dict_from_file(SHARED_MODEL_FOLDERS, SHARED_MODEL_FOLDERS_FILE, "SHARED_MODEL_FOLDERS", SHARED_MODELS_DIR)
success, dict = load_global_dict_from_file(SHARED_MODEL_FOLDERS, SHARED_MODEL_FOLDERS_FILE, "SHARED_MODEL_FOLDERS", SHARED_MODELS_DIR)
if success:
SHARED_MODEL_FOLDERS = dict
if os.path.exists(SHARED_MODEL_FOLDERS_FILE) and send_SocketMessage:
send_websocket_message('extend_ui_helper', {
@ -341,7 +343,9 @@ def sync_with_app_configs_install_dirs():
# NOTE: this APP_INSTALL_DIRS_FILE is temporary synced with the app_configs dict
def init_app_install_dirs():
global APP_INSTALL_DIRS
init_global_dict_from_file(APP_INSTALL_DIRS, APP_INSTALL_DIRS_FILE, "APP_INSTALL_DIRS", SHARED_MODELS_DIR)
success, dict = load_global_dict_from_file(APP_INSTALL_DIRS, APP_INSTALL_DIRS_FILE, "APP_INSTALL_DIRS", SHARED_MODELS_DIR)
if success:
APP_INSTALL_DIRS = dict
return
@ -496,7 +500,9 @@ SHARED_MODEL_APP_MAP = {
# which does a default mapping from app code or (if exists) from external JSON 'SHARED_MODEL_APP_MAP_FILE' file
def init_shared_model_app_map():
global SHARED_MODEL_APP_MAP
init_global_dict_from_file(SHARED_MODEL_APP_MAP, SHARED_MODEL_APP_MAP_FILE, "SHARED_MODEL_APP_MAP", SHARED_MODELS_DIR)
success, dict = load_global_dict_from_file(SHARED_MODEL_APP_MAP, SHARED_MODEL_APP_MAP_FILE, "SHARED_MODEL_APP_MAP", SHARED_MODELS_DIR)
if success:
SHARED_MODEL_APP_MAP = dict
return