Quantization
Collection
4 items • Updated
import torch
from diffusers import DiffusionPipeline
# switch to "mps" for apple devices
pipe = DiffusionPipeline.from_pretrained("Silan10/flux-quantized-bitsandbytes", dtype=torch.bfloat16, device_map="cuda")
prompt = "Astronaut in a jungle, cold color palette, muted colors, detailed, 8k"
image = pipe(prompt).images[0]Silan10/flux-quantized-bitsandbytes is an 8-bit quantized version of the
black-forest-labs/FLUX.1-dev
text-to-image model. In this version, the transformer, text_encoder and
text_encoder_2 components have been quantized to 8-bit precision using bitsandbytes.
Bitsandbytes quantization uses 8-bit integer representation with dynamic scaling factors. This provides substantial memory savings while maintaining high image quality through mixed-precision computation.
import torch
import os
from diffusers import FluxPipeline
model_path = "Silan10/flux-quantized-bitsandbytes"
print("Loading pipeline...")
pipe = FluxPipeline.from_pretrained(
model_path,
torch_dtype=torch.bfloat16
)
pipe.to("cuda")
print("✓ Pipeline loaded successfully.")
prompt = "Ultra-detailed nighttime cyberpunk city street, several pedestrians in modern clothes, one person in the foreground looking toward the camera, sharp facial features and detailed hair, wet pavement reflecting colorful neon signs, shop windows with small readable text on signs, a gradient sky fading from deep blue to purple, a mix of strong highlights and deep shadows, highly detailed, 4K, cinematic lighting."
print("Generating image...")
image = pipe(
prompt,
num_inference_steps=20,
guidance_scale=3.5,
max_sequence_length=512,
width=1024,
height=1024,
generator=torch.Generator("cpu").manual_seed(42)
).images[0]
image.save("output_bitsandbytes.png")
print("✓ Image generated successfully.")
print("DONE!")
Base model
black-forest-labs/FLUX.1-dev