mirror of
https://github.com/kodxana/madiator-docker-runpod.git
synced 2024-11-10 05:25:18 +01:00
revorked ComFYui TEMPLATE
This commit is contained in:
parent
3832394c7e
commit
de61e8452e
3 changed files with 108 additions and 40 deletions
|
@ -1,13 +1,8 @@
|
|||
# Stage 1: Base Image
|
||||
FROM madiator2011/better-pytorch:cuda12.1 as base
|
||||
FROM madiator2011/better-base:cuda12.1 as base
|
||||
|
||||
ARG PYTHON_VERSION1
|
||||
ARG TORCH
|
||||
|
||||
# Install PyTorch if specified
|
||||
RUN if [ -n "${TORCH}" ]; then \
|
||||
pip install --upgrade --no-cache-dir ${TORCH}; \
|
||||
fi
|
||||
ARG PYTHON_VERSION1=3.10
|
||||
ARG TORCH=torch==2.3.0+cu121
|
||||
|
||||
# Stage 2: ComfyUI Installation
|
||||
FROM base as comfyui-install
|
||||
|
@ -15,46 +10,74 @@ FROM base as comfyui-install
|
|||
# Create and activate virtual environment for ComfyUI installation
|
||||
RUN python -m venv /venv
|
||||
ENV PATH="/venv/bin:$PATH"
|
||||
ENV VIRTUAL_ENV="/venv"
|
||||
|
||||
# Install PyTorch within the virtual environment
|
||||
RUN if [ -n "${TORCH}" ]; then \
|
||||
/venv/bin/pip install --upgrade --no-cache-dir ${TORCH}; \
|
||||
fi
|
||||
|
||||
RUN git clone https://github.com/comfyanonymous/ComfyUI.git && \
|
||||
cd /ComfyUI && \
|
||||
pip install -r requirements.txt && \
|
||||
/venv/bin/pip install -r requirements.txt && \
|
||||
git clone https://github.com/ltdrdata/ComfyUI-Manager.git custom_nodes/ComfyUI-Manager && \
|
||||
cd custom_nodes/ComfyUI-Manager && \
|
||||
pip install -r requirements.txt && \
|
||||
pip install -U xformers --index-url https://download.pytorch.org/whl/cu121
|
||||
/venv/bin/pip install -r requirements.txt && \
|
||||
/venv/bin/pip install -U xformers --index-url https://download.pytorch.org/whl/cu121 && \
|
||||
/venv/bin/pip install -U accelerate wheel
|
||||
|
||||
# Stage 3: Model Setup and Final Image
|
||||
FROM comfyui-install as final
|
||||
# Stage 3: Model Setup
|
||||
FROM comfyui-install as model-setup
|
||||
|
||||
# Create model and cache directories
|
||||
RUN mkdir -p /root/.cache/huggingface && mkdir -p /comfy-models
|
||||
|
||||
# Download models directly
|
||||
RUN wget -q --show-progress https://huggingface.co/runwayml/stable-diffusion-v1-5/resolve/main/v1-5-pruned-emaonly.safetensors -O /comfy-models/v1-5-pruned-emaonly.safetensors && \
|
||||
wget -q --show-progress https://huggingface.co/stabilityai/stable-diffusion-2-1/resolve/main/v2-1_768-ema-pruned.safetensors -O /comfy-models/v2-1_768-ema-pruned.safetensors && \
|
||||
wget -q --show-progress https://huggingface.co/stabilityai/stable-diffusion-xl-base-1.0/resolve/main/sd_xl_base_1.0.safetensors -O /comfy-models/sd_xl_base_1.0.safetensors && \
|
||||
wget -q --show-progress https://huggingface.co/stabilityai/stable-diffusion-xl-refiner-1.0/resolve/main/sd_xl_refiner_1.0.safetensors -O /comfy-models/sd_xl_refiner_1.0.safetensors
|
||||
ARG INCLUDE_MODELS=false
|
||||
|
||||
# Download each model in a separate layer
|
||||
RUN if [ "${INCLUDE_MODELS}" = "true" ]; then \
|
||||
wget -q --show-progress https://huggingface.co/runwayml/stable-diffusion-v1-5/resolve/main/v1-5-pruned-emaonly.safetensors -O /comfy-models/v1-5-pruned-emaonly.safetensors; \
|
||||
fi
|
||||
|
||||
RUN if [ "${INCLUDE_MODELS}" = "true" ]; then \
|
||||
wget -q --show-progress https://huggingface.co/stabilityai/stable-diffusion-2-1/resolve/main/v2-1_768-ema-pruned.safetensors -O /comfy-models/v2-1_768-ema-pruned.safetensors; \
|
||||
fi
|
||||
|
||||
RUN if [ "${INCLUDE_MODELS}" = "true" ]; then \
|
||||
wget -q --show-progress https://huggingface.co/stabilityai/stable-diffusion-xl-base-1.0/resolve/main/sd_xl_base_1.0.safetensors -O /comfy-models/sd_xl_base_1.0.safetensors; \
|
||||
fi
|
||||
|
||||
RUN if [ "${INCLUDE_MODELS}" = "true" ]; then \
|
||||
wget -q --show-progress https://huggingface.co/stabilityai/stable-diffusion-xl-refiner-1.0/resolve/main/sd_xl_refiner_1.0.safetensors -O /comfy-models/sd_xl_refiner_1.0.safetensors; \
|
||||
fi
|
||||
|
||||
# Verify models were downloaded
|
||||
RUN ls -lh /comfy-models
|
||||
RUN if [ "${INCLUDE_MODELS}" = "true" ]; then \
|
||||
ls -lh /comfy-models; \
|
||||
fi
|
||||
|
||||
# NGINX Proxy
|
||||
COPY --from=proxy nginx.conf /etc/nginx/nginx.conf
|
||||
COPY --from=proxy readme.html /usr/share/nginx/html/readme.html
|
||||
# Stage 4: Final Image
|
||||
FROM comfyui-install as final
|
||||
|
||||
# Copy the README.md
|
||||
COPY README.md /usr/share/nginx/html/README.md
|
||||
# Copy models if they were included
|
||||
COPY --from=model-setup /comfy-models /comfy-models
|
||||
|
||||
# Set environment variables for runtime
|
||||
ENV PATH="/workspace/venv/bin:$PATH"
|
||||
ENV VIRTUAL_ENV="/workspace/venv"
|
||||
|
||||
# Start Scripts
|
||||
COPY pre_start.sh /pre_start.sh
|
||||
# Copy the README.md
|
||||
COPY README.md /usr/share/nginx/html/README.md
|
||||
|
||||
# Start Scripts
|
||||
# NGINX Proxy
|
||||
COPY --from=proxy nginx.conf /etc/nginx/nginx.conf
|
||||
COPY --from=proxy readme.html /usr/share/nginx/html/readme.html
|
||||
|
||||
# Copy all necessary scripts
|
||||
COPY pre_start.sh /pre_start.sh
|
||||
COPY --from=scripts start.sh /
|
||||
RUN chmod +x /start.sh
|
||||
RUN chmod +x /pre_start.sh
|
||||
|
||||
# CMD
|
||||
CMD [ "/start.sh" ]
|
|
@ -1,18 +1,59 @@
|
|||
group "default" {
|
||||
targets = ["py310-cuda121"]
|
||||
targets = ["full-version", "light-version", "light-experimental"]
|
||||
}
|
||||
|
||||
target "py310-cuda121" {
|
||||
target "base" {
|
||||
dockerfile = "Dockerfile"
|
||||
args = {
|
||||
BASE_IMAGE = "madiator2011/better-base:cuda12.1",
|
||||
TORCH = "torch==2.3.0+cu121 -f https://download.pytorch.org/whl/torch_stable.html",
|
||||
PYTHON_VERSION1 = "3.10"
|
||||
}
|
||||
contexts = {
|
||||
scripts = "../../container-template"
|
||||
proxy = "../../container-template/proxy"
|
||||
logo = "../../container-template"
|
||||
}
|
||||
}
|
||||
|
||||
target "full-version" {
|
||||
inherits = ["base"]
|
||||
args = {
|
||||
INCLUDE_MODELS = "true"
|
||||
}
|
||||
contexts = {
|
||||
scripts = "../../container-template"
|
||||
proxy = "../../container-template/proxy"
|
||||
logo = "../../container-template"
|
||||
}
|
||||
tags = ["madiator2011/better-comfyui:full"]
|
||||
}
|
||||
|
||||
target "light-version" {
|
||||
inherits = ["base"]
|
||||
args = {
|
||||
INCLUDE_MODELS = "false"
|
||||
}
|
||||
contexts = {
|
||||
scripts = "../../container-template"
|
||||
proxy = "../../container-template/proxy"
|
||||
logo = "../../container-template"
|
||||
}
|
||||
tags = ["madiator2011/better-comfyui:light"]
|
||||
}
|
||||
|
||||
target "light-experimental" {
|
||||
dockerfile = "Dockerfile"
|
||||
args = {
|
||||
BASE_IMAGE = "nvidia/cuda:12.1.0-cudnn8-devel-ubuntu22.04",
|
||||
BASE_IMAGE = "madiator2011/better-base:light-cuda",
|
||||
TORCH = "torch==2.3.0+cu121 -f https://download.pytorch.org/whl/torch_stable.html",
|
||||
PYTHON_VERSION = "3.10"
|
||||
PYTHON_VERSION1 = "3.10",
|
||||
INCLUDE_MODELS = "false"
|
||||
}
|
||||
tags = ["madiator2011/better-comfyui:cuda12.1"]
|
||||
contexts = {
|
||||
scripts = "../../container-template"
|
||||
proxy = "../../container-template/proxy"
|
||||
logo = "../../container-template"
|
||||
}
|
||||
tags = ["madiator2011/better-comfyui:light-experimental"]
|
||||
}
|
|
@ -10,31 +10,35 @@ print_feedback() {
|
|||
echo -e "${GREEN}[ComfyUI Startup]:${NC} $1"
|
||||
}
|
||||
|
||||
# Function to run rsync with progress bar
|
||||
# Function to run rsync with progress bar and optimizations
|
||||
rsync_with_progress() {
|
||||
stdbuf -i0 -o0 -e0 rsync -au --info=progress2 "$@" | stdbuf -i0 -o0 -e0 tr '\r' '\n' | stdbuf -i0 -o0 -e0 grep -oP '\d+%|\d+.\d+[mMgG]' | tqdm --bar-format='{l_bar}{bar}' --total=100 --unit='%' > /dev/null
|
||||
rsync -aHvx --info=progress2 "$@"
|
||||
}
|
||||
|
||||
# Check for required commands
|
||||
for cmd in stdbuf rsync tr grep tqdm; do
|
||||
for cmd in rsync; do
|
||||
if ! command -v $cmd &> /dev/null; then
|
||||
echo "$cmd could not be found, please install it."
|
||||
exit 1
|
||||
fi
|
||||
done
|
||||
|
||||
LOG_FILE="/workspace/comfyui.log"
|
||||
|
||||
print_feedback "Starting ComfyUI setup..."
|
||||
|
||||
print_feedback "Syncing virtual environment..."
|
||||
rsync_with_progress /venv/ /workspace/venv/
|
||||
|
||||
print_feedback "Activating virtual environment..."
|
||||
export VIRTUAL_ENV="/workspace/venv"
|
||||
export PATH="$VIRTUAL_ENV/bin:$PATH"
|
||||
source /workspace/venv/bin/activate
|
||||
|
||||
export PYTHONUNBUFFERED=1
|
||||
|
||||
print_feedback "Syncing ComfyUI files..."
|
||||
rsync_with_progress --remove-source-files /ComfyUI/ /workspace/ComfyUI/
|
||||
rsync_with_progress /ComfyUI/ /workspace/ComfyUI/
|
||||
|
||||
print_feedback "Creating symbolic links for model checkpoints..."
|
||||
ln -sf /comfy-models/* /workspace/ComfyUI/models/checkpoints/
|
||||
|
@ -47,7 +51,7 @@ print_feedback "ComfyUI will be available at http://0.0.0.0:3000"
|
|||
|
||||
# Check if CUSTOM_ARGS is set and not empty
|
||||
if [ -n "$CUSTOM_ARGS" ]; then
|
||||
exec /workspace/venv/bin/python main.py --listen --port 3000 $CUSTOM_ARGS
|
||||
exec /workspace/venv/bin/python main.py --listen --port 3000 $CUSTOM_ARGS 2>&1 | tee -a $LOG_FILE
|
||||
else
|
||||
exec /workspace/venv/bin/python main.py --listen --port 3000
|
||||
exec /workspace/venv/bin/python main.py --listen --port 3000 2>&1 | tee -a $LOG_FILE
|
||||
fi
|
Loading…
Reference in a new issue