How to use from the
Use from the
Diffusers library
pip install -U diffusers transformers accelerate
import torch
from diffusers import DiffusionPipeline

# switch to "mps" for apple devices
pipe = DiffusionPipeline.from_pretrained("stabilityai/stable-diffusion-xl-base-1.0", dtype=torch.bfloat16, device_map="cuda")
pipe.load_lora_weights("PhilSad/outputs")

prompt = "interior design in sks style"
image = pipe(prompt).images[0]

SDXL LoRA DreamBooth - PhilSad/outputs

Model description

These are PhilSad/outputs LoRA adaption weights for stabilityai/stable-diffusion-xl-base-1.0.

The weights were trained using DreamBooth.

LoRA for the text encoder was enabled: False.

Special VAE used for training: None.

Trigger words

You should use interior design in sks style to trigger the image generation.

Download model

Weights for this model are available in Safetensors format.

Download them in the Files & versions tab.

Intended uses & limitations

How to use

# TODO: add an example code snippet for running this diffusion pipeline

Limitations and bias

[TODO: provide examples of latent issues and potential remediations]

Training details

[TODO: describe the data used to train the model]

Downloads last month
5
Inference Providers NEW

Model tree for PhilSad/outputs

Adapter
(8376)
this model

Space using PhilSad/outputs 1