How to use optimum-intel-internal-testing/tiny-stable-diffusion-with-textual-inversion with Diffusers:
pip install -U diffusers transformers accelerate
import torch from diffusers import DiffusionPipeline # switch to "mps" for apple devices pipe = DiffusionPipeline.from_pretrained("hf-internal-testing/tiny-stable-diffusion-torch", dtype=torch.bfloat16, device_map="cuda") pipe.load_textual_inversion("optimum-intel-internal-testing/tiny-stable-diffusion-with-textual-inversion")
The community tab is the place to discuss and collaborate with the HF community!