![]() |
15 hours ago | |
---|---|---|
.dev_scripts | 2 months ago | |
.github | 1 week ago | |
coverage | 6 months ago | |
docker | 1 month ago | |
docs | 2 days ago | |
installer | 1 day ago | |
invokeai | 15 hours ago | |
scripts | 2 weeks ago | |
tests | 6 days ago | |
.dockerignore | 2 months ago | |
.editorconfig | 10 months ago | |
.git-blame-ignore-revs | 2 months ago | |
.gitattributes | 9 months ago | |
.gitignore | 1 month ago | |
.gitmodules | 1 year ago | |
.pre-commit-config.yaml | 2 weeks ago | |
.prettierrc.yaml | 1 year ago | |
CODE_OF_CONDUCT.md | 10 months ago | |
InvokeAI_Statement_of_Values.md | 9 months ago | |
LICENSE | 2 months ago | |
LICENSE-SD1+SD2.txt | 2 months ago | |
LICENSE-SDXL.txt | 2 months ago | |
README.md | 2 weeks ago | |
Stable_Diffusion_v1_Model_Card.md | 9 months ago | |
flake.lock | 1 month ago | |
flake.nix | 1 month ago | |
mkdocs.yml | 2 weeks ago | |
pyproject.toml | 4 days ago |
1a917d94
-e099-4fa1-a70f-7dd8d06910
)
# Invoke AI - Generative AI for Professional Creatives
## Professional Creative Tools for Stable Diffusion, Custom-Trained Models, and more.
To learn more about Invoke AI, get started instantly, or implement our Business solutions, visit [invoke.ai](https://invoke.ai)
[![discord badge]][discord link]
[![latest release badge]][latest release link] [![github stars badge]][github stars link] [![github forks badge]][github forks link]
[![CI checks on main badge]][CI checks on main link] [![latest commit to main badge]][latest commit to main link]
[![github open issues badge]][github open issues link] [![github open prs badge]][github open prs link] [![translation status badge]][translation status link]
InvokeAI is a leading creative engine built to empower professionals and enthusiasts alike. Generate and create stunning visual media using the latest AI-driven technologies. InvokeAI offers an industry leading Web Interface, interactive Command Line Interface, and also serves as the foundation for multiple commercial products.
Quick links: [How to Install] [Discord Server] [Documentation and Tutorials] [Bug Reports] [Discussion, Ideas & Q&A] [Contributing]
Table of Contents 📝
Getting Started
More About Invoke
Supporting the Project
For full installation and upgrade instructions, please see: InvokeAI Installation Overview
If upgrading from version 2.3, please read Migrating a 2.3 root directory to 3.0 first.
Go to the bottom of the Latest Release Page
Download the .zip file for your OS (Windows/macOS/Linux).
Unzip the file.
Windows: double-click on the install.bat
script. macOS: Open a Terminal window, drag the file install.sh
from Finder
into the Terminal, and press return. Linux: run install.sh
.
You'll be asked to confirm the location of the folder in which to install InvokeAI and its image generation model files. Pick a location with at least 15 GB of free memory. More if you plan on installing lots of models.
Wait while the installer does its thing. After installing the software, the installer will launch a script that lets you configure InvokeAI and select a set of starting image generation models.
Find the folder that InvokeAI was installed into (it is not the
same as the unpacked zip file directory!) The default location of this
folder (if you didn't change it in step 5) is ~/invokeai
on
Linux/Mac systems, and C:\Users\YourName\invokeai
on Windows. This directory will contain launcher scripts named invoke.sh
and invoke.bat
.
On Windows systems, double-click on the invoke.bat
file. On
macOS, open a Terminal window, drag invoke.sh
from the folder into
the Terminal, and press return. On Linux, run invoke.sh
Press 2 to open the "browser-based UI", press enter/return, wait a minute or two for Stable Diffusion to start up, then open your browser and go to http://localhost:9090.
Type banana sushi
in the box on the top left and click Invoke
You must have Python 3.9 through 3.11 installed on your machine. Earlier or
later versions are not supported.
Node.js also needs to be installed along with yarn (can be installed with
the command npm install -g yarn
if needed)
Create a directory to install InvokeAI into. You'll need at least 15 GB of free space:
mkdir invokeai
````
3. Create a virtual environment named `.venv` inside this directory and activate it:
```terminal
cd invokeai
python -m venv .venv --prompt InvokeAI
For Linux/Mac users:
source .venv/bin/activate
For Windows users:
.venv\Scripts\activate
For Windows/Linux with an NVIDIA GPU:
pip install "InvokeAI[xformers]" --use-pep517 --extra-index-url https://download.pytorch.org/whl/cu118
For Linux with an AMD GPU:
pip install InvokeAI --use-pep517 --extra-index-url https://download.pytorch.org/whl/rocm5.4.2
For non-GPU systems:
pip install InvokeAI --use-pep517 --extra-index-url https://download.pytorch.org/whl/cpu
For Macintoshes, either Intel or M1/M2:
pip install InvokeAI --use-pep517
invokeai-configure --root .
Don't miss the dot at the end!
invokeai-web
Point your browser to http://localhost:9090 to bring up the web interface.
Type banana sushi
in the box on the top left and click Invoke
.
Be sure to activate the virtual environment each time before re-launching InvokeAI,
using source .venv/bin/activate
or .venv\Scripts\activate
.
This fork is supported across Linux, Windows and Macintosh. Linux users can use either an Nvidia-based card (with CUDA support) or an AMD card (using the ROCm driver). For full installation and upgrade instructions, please see: InvokeAI Installation Overview
The InvokeAI root directory is where the InvokeAI startup file,
installed models, and generated images are stored. It is ordinarily
named invokeai
and located in your home directory. The contents and
layout of this directory has changed between versions 2.3 and 3.0 and
cannot be used directly.
We currently recommend that you use the installer to create a new root
directory named differently from the 2.3 one, e.g. invokeai-3
and
then use a migration script to copy your 2.3 models into the new
location. However, if you choose, you can upgrade this directory in
place. This section gives both recipes.
This is the safer recipe because it leaves your old root directory in place to fall back on.
Follow the instructions above to create and install InvokeAI in a directory that has a different name from the 2.3 invokeai directory. In this example, we will use "invokeai-3"
When you are prompted to select models to install, select a minimal set of models, such as stable-diffusion-v1.5 only.
After installation is complete launch invokeai.sh
(Linux/Mac) or
invokeai.bat
and select option 8 "Open the developers console". This
will take you to the command line.
Issue the command invokeai-migrate3 --from /path/to/v2.3-root --to
/path/to/invokeai-3-root
. Provide the correct --from
and --to
paths for your v2.3 and v3.0 root directories respectively.
This will copy and convert your old models from 2.3 format to 3.0
format and create a new models
directory in the 3.0 directory. The
old models directory (which contains the models selected at install
time) will be renamed models.orig
and can be deleted once you have
confirmed that the migration was successful.
If you wish, you can pass the 2.3 root directory to both --from
and
--to
in order to update in place. Warning: this directory will no
longer be usable with InvokeAI 2.3.
For the adventurous, you may do an in-place upgrade from 2.3 to 3.0 without touching the command line. *This recipe does not work on Windows platforms due to a bug in the Windows version of the 2.3 upgrade script. See the next section for a Windows recipe.
Launch the InvokeAI launcher script in your current v2.3 root directory.
Select option [9] "Update InvokeAI" to bring up the updater dialog.
Select option [1] to upgrade to the latest release.
Once the upgrade is finished you will be returned to the launcher menu. Select option [7] "Re-run the configure script to fix a broken install or to complete a major upgrade".
This will run the configure script against the v2.3 directory and update it to the 3.0 format. The following files will be replaced:
The original versions of these files will be saved with the suffix ".orig" appended to the end. Once you have confirmed that the upgrade worked, you can safely remove these files. Alternatively you can restore a working v2.3 directory by removing the new files and restoring the ".orig" files' original names.
Windows Users can upgrade with the
invoke.sh
or invoke.bat
Type the following commands
pip install "invokeai @ https://github.com/invoke-ai/InvokeAI/archive/refs/tags/v3.0.0" --use-pep517 --upgrade
invokeai-configure --root .
(Replace v3.0.0
with the current release number if this document is out of date).
The first command will install and upgrade new software to run InvokeAI. The second will prepare the 2.3 directory for use with 3.0. You may now launch the WebUI in the usual way, by selecting option [1] from the launcher script
The migration script will migrate your invokeai settings and models, including textual inversion models, LoRAs and merges that you may have installed previously. However it does not migrate the generated images stored in your 2.3-format outputs directory. To do this, you need to run an additional step:
From a working InvokeAI 3.0 root directory, start the launcher and enter menu option [8] to open the "developer's console".
At the developer's console command line, type the command:
invokeai-import-images
(Many kudos to techjedi for contributing this script.)
InvokeAI is supported across Linux, Windows and macOS. Linux users can use either an Nvidia-based card (with CUDA support) or an AMD card (using the ROCm driver).
You will need one of the following:
We do not recommend the GTX 1650 or 1660 series video cards. They are unable to run in half-precision mode and do not have sufficient VRAM to render 512x512 images.
Memory - At least 12 GB Main Memory RAM.
Disk - At least 12 GB of free disk space for the machine learning model, Python, and all its dependencies.
Feature documentation can be reviewed by navigating to the InvokeAI Documentation page
InvokeAI offers a locally hosted Web Server & React Frontend, with an industry leading user experience. The Web-based UI allows for simple and intuitive workflows, and is responsive for use on mobile devices and tablets accessing the web server.
The Unified Canvas is a fully integrated canvas implementation with support for all core generation capabilities, in/outpainting, brush tools, and more. This creative tool unlocks the capability for artists to create with AI as a creative collaborator, and can be used to augment AI-generated imagery, sketches, photography, renders, and more.
InvokeAI offers a fully featured workflow management solution, enabling users to combine the power of nodes based workflows with the easy of a UI. This allows for customizable generation pipelines to be developed and shared by users looking to create specific workflows to support their production use-cases.
Invoke AI provides an organized gallery system for easily storing, accessing, and remixing your content in the Invoke workspace. Images can be dragged/dropped onto any Image-base UI element in the application, and rich metadata within the Image allows for easy recall of key prompts or settings used in your workflow.
For our latest changes, view our Release Notes and the CHANGELOG.
Please check out our Q&A to get solutions for common installation problems and other issues. For more help, please join our Discord
Anyone who wishes to contribute to this project, whether documentation, features, bug fixes, code cleanup, testing, or code reviews, is very much encouraged to do so.
Get started with contributing by reading our Contribution documentation, joining the #dev-chat or the GitHub discussion board.
If you are unfamiliar with how to contribute to GitHub projects, we have a new contributor checklist you can follow to get started contributing: New Contributor Checklist.
We hope you enjoy using our software as much as we enjoy creating it, and we hope that some of those of you who are reading this will elect to become part of our community.
Welcome to InvokeAI!
This fork is a combined effort of various people from across the world. Check out the list of all these amazing people. We thank them for their time, hard work and effort.
For support, please use this repository's GitHub Issues tracking service, or join the Discord.
Original portions of the software are Copyright (c) 2023 by respective contributors.