ai-suite-rocm/Dockerfile

70 lines
2.6 KiB
Docker
Raw Normal View History

2023-11-10 21:10:06 +01:00
FROM rocm/dev-ubuntu-22.04:5.7-complete
ENV DEBIAN_FRONTEND=noninteractive \
PYTHONUNBUFFERED=1 \
PYTHONIOENCODING=UTF-8
WORKDIR /tmp
# Install python 3.10 and requirements
RUN apt-get update -y && apt-get install -y software-properties-common && \
add-apt-repository ppa:deadsnakes/ppa
RUN apt-get update &&\
apt-get install -y \
wget \
git \
python3.10 \
python3-dev \
python-is-python3 \
python3.10-venv \
rsync
RUN curl -sS https://bootstrap.pypa.io/get-pip.py | python3.10
RUN python3.10 -m pip install --upgrade pip wheel setuptools
# Install PyTorch
RUN python3.10 -m pip install torch torchvision --index-url https://download.pytorch.org/whl/nightly/rocm5.7
# Create ai folder for saving projects
RUN mkdir /ai/
# Install StableDiffusion in /ai/
WORKDIR /ai/stablediffusion-webui
RUN git clone -b dev https://github.com/AUTOMATIC1111/stable-diffusion-webui /ai/git/stablediffusion-webui
RUN cp -R /ai/git/stablediffusion-webui/* /ai/stablediffusion-webui/
# Create VENV for StableDiffusion with inherit pytorch
RUN python3.10 -m venv /ai/venv/stablediffusion/ --system-site-packages
RUN /ai/venv/stablediffusion/bin/python launch.py --skip-torch-cuda-test --exit
RUN /ai/venv/stablediffusion/bin/python -m pip install opencv-python-headless tomesd protobuf --upgrade
# RUN /ai/venv/stablediffusion/bin/python -m pip install -r requirements.txt # should not be needed
# Install Kobold AI in /ai/
WORKDIR /ai/koboldai
RUN git clone https://github.com/YellowRoseCx/koboldcpp-rocm.git -b main --depth 1 /ai/git/koboldai
RUN cp -R /ai/git/koboldai/* /ai/koboldai/
# Create VENV for KoboldAI with inherit pytorch
# RUN python3.10 -m venv /ai/venv/koboldai/ --system-site-packages # not needed actually
# Install python requirements
# RUN /ai/venv/koboldai/bin/python -m pip install -r requirements.txt # should not be needed
# Build KoboldAI for ROCM
RUN make LLAMA_HIPBLAS=1 -j4
# Install LlamaCPP in /ai/
WORKDIR /ai/llamacpp
RUN git clone https://github.com/ggerganov/llama.cpp.git -b master /ai/git/llamacpp
RUN cp -R /ai/git/llamacpp/* /ai/llamacpp/
# Install requirements for LlamaCPP build
RUN apt-get install -y hipblas hipblaslt hipsparse hipcub hip-runtime-amd rocthrust rocthrust-dev rocrand
# Build LlamaCPP for ROCM
RUN make LLAMA_HIPBLAS=1 -j4
# Set safe directory for extensions and stuff
RUN git config --global --add safe.directory "*"
# Go back to ai folder when done
WORKDIR /ai
# Setup PyTorch ENVs for AMD GPUs
ENV PYTORCH_HIP_ALLOC_CONF=garbage_collection_threshold:0.8,max_split_size_mb:128
# ENV HIP_VISIBLE_DEVICES="0"
# ENV AMDGPU_TARGETS="gfx1030"
# ENV HSA_OVERRIDE_GFX_VERSION='10.3.0'