diff --git a/docs/index.md b/docs/index.md index c4e7ac264c..6a4b3d841c 100644 --- a/docs/index.md +++ b/docs/index.md @@ -116,7 +116,7 @@ images in full-precision mode: `invoke.py` with the `--precision=float32` flag: ```bash - (invokeai) ~/InvokeAI$ python scripts/invoke.py --full_precision + (invokeai) ~/InvokeAI$ python scripts/invoke.py --precision=float32 ``` ## :octicons-package-dependencies-24: Installation diff --git a/docs/installation/010_INSTALL_AUTOMATED.md b/docs/installation/010_INSTALL_AUTOMATED.md index f500095524..b24c5f8ea9 100644 --- a/docs/installation/010_INSTALL_AUTOMATED.md +++ b/docs/installation/010_INSTALL_AUTOMATED.md @@ -118,9 +118,7 @@ experimental versions later. Terminal, including InvokeAI. This package is provided directly by Apple. To install, open a terminal window and run `xcode-select --install`. You will get a macOS system popup guiding you through the install. If you already have them installed, you will instead see some - output in the Terminal advising you that the tools are already installed. - More information can be found at [FreeCode - Camp](https://www.freecodecamp.org/news/install-xcode-command-line-tools/) + output in the Terminal advising you that the tools are already installed. More information can be found at [FreeCode Camp](https://www.freecodecamp.org/news/install-xcode-command-line-tools/) 3. The InvokeAI installer is distributed as a ZIP files. Go to the [latest release](https://github.com/invoke-ai/InvokeAI/releases/latest), diff --git a/docs/installation/030_INSTALL_CUDA_AND_ROCM.md b/docs/installation/030_INSTALL_CUDA_AND_ROCM.md index db138a4785..54dded0fde 100644 --- a/docs/installation/030_INSTALL_CUDA_AND_ROCM.md +++ b/docs/installation/030_INSTALL_CUDA_AND_ROCM.md @@ -8,27 +8,111 @@ title: NVIDIA Cuda / AMD ROCm </figure> +In order for InvokeAI to run at full speed, you will need a graphics +card with a supported GPU. InvokeAI supports NVidia cards via the CUDA +driver on Windows and Linux, and AMD cards via the ROCm driver on Linux. + ## :simple-nvidia: CUDA -### Container Runtime +### Linux and Windows Install -Fancy Site: https://developer.nvidia.com/nvidia-container-runtime +If you have used your system for other graphics-intensive tasks, such +as gaming, you may very well already have the CUDA drivers +installed. To confirm, open up a command-line window and type: -GitHub Source: https://github.com/NVIDIA/nvidia-container-runtime/ -(where the Fancy Site also links ot xD) +``` +nvidia-smi +``` -Maybe the most simple way to get InvokeAI running with NVIDIA CUDA will be the official -NVIDIA Container Runtime (not confirmed, but got told by a friend) that the Runtime even -works when you do not have the actual Drivers installed, since it is mounting the GPU -into the Container +If this command produces a status report on the GPU(s) installed on +your system, CUDA is installed and you have no more work to do. If +instead you get "command not found", or similar, then the driver will +need to be installed. + +We strongly recommend that you install the CUDA Toolkit package +directly from NVIDIA. **Do not try to install Ubuntu's +nvidia-cuda-toolkit package. It is out of date and will cause +conflicts among the NVIDIA driver and binaries.** + +Go to [CUDA Toolkit 11.7 +Downloads](https://developer.nvidia.com/cuda-11-7-0-download-archive), +and use the target selection wizard to choose your operating system, +hardware platform, and preferred installation method (e.g. "local" +versus "network"). + +This will provide you with a downloadable install file or, depending +on your choices, a recipe for downloading and running a install shell +script. Be sure to read and follow the full installation instructions. + +After an install that seems successful, you can confirm by again +running `nvidia-smi` from the command line. + +### Linux Install with a Runtime Container + +On Linux systems, an alternative to installing the driver directly on +your system is to run an NVIDIA software container that has the driver +already in place. This is recommended if you are already familiar with +containerization technologies such as Docker. + +For downloads and instructions, visit the [NVIDIA CUDA Container +Runtime Site](https://developer.nvidia.com/nvidia-container-runtime) ## :simple-amd: ROCm -### ROCm-docker +### Linux Install -GitHub Source: https://github.com/RadeonOpenCompute/ROCm-docker/blob/master/quick-start.md +AMD GPUs are only supported on Linux platforms due to the lack of a +Windows ROCm driver at the current time. Also be aware that support +for newer AMD GPUs is spotty. Your mileage may vary. -Yeah - I am sorry, but since I am not into PC-Masterrace, I had no better Idea than -looking up if there is a container runtime for ROCm as well xD +It is possible that the ROCm driver is already installed on your +machine. To test, open up a terminal window and issue the following +command: -So please forgive me, but at least the page isn't empty anymore 🙈 +``` +rocm-smi +``` + +If you get a table labeled "ROCm System Management Interface" the +driver is installed and you are done. If you get "command not found," +then the driver needs to be installed. + +Go to AMD's [ROCm Downloads +Guide](https://rocmdocs.amd.com/en/latest/Installation_Guide/Installation_new.html#installation-methods) +and scroll to the _Installation Methods_ section. Find the subsection +for the install method for your preferred Linux distribution, and +issue the commands given in the recipe. + +Annoyingly, the official AMD site does not have a recipe for the most +recent version of Ubuntu, 22.04. However, this [community-contributed +recipe](https://novaspirit.github.io/amdgpu-rocm-ubu22/) is reported +to work well. + +After installation, please run `rocm-smi` a second time to confirm +that the driver is present and the GPU is recognized. You may need to +do a reboot in order to load the driver. + +### Linux Install with a ROCm-docker Container + +If you are comfortable with the Docker containerization system, then +you can build a ROCm docker file. The source code and installation +recipes are available +[Here](https://github.com/RadeonOpenCompute/ROCm-docker/blob/master/quick-start.md) + +### Torch Installation + +When installing torch and torchvision manually with `pip`, remember to provide +the argument `--extra-index-url +https://download.pytorch.org/whl/rocm5.2` as described in the [Manual +Installation Guide](020_INSTALL_MANUAL.md). + +This will be done automatically for you if you use the installer +script. + +Be aware that the torch machine learning library does not seamlessly +interoperate with all AMD GPUs and you may experience garbled images, +black images, or long startup delays before rendering commences. Most +of these issues can be solved by Googling for workarounds. If you have +a problem and find a solution, please post an +[Issue](https://github.com/invoke-ai/InvokeAI/issues) so that other +users benefit and we can update this document.