Update README.md
Browse files
README.md
CHANGED
@@ -133,8 +133,8 @@ except:
|
|
133 |
hf_hub_download(repo_id="briaai/BRIA-4B-Adapt", filename='pipeline_bria.py', local_dir=local_dir)
|
134 |
hf_hub_download(repo_id="briaai/BRIA-4B-Adapt", filename='transformer_bria.py', local_dir=local_dir)
|
135 |
hf_hub_download(repo_id="briaai/BRIA-4B-Adapt", filename='bria_utils.py', local_dir=local_dir)
|
136 |
-
hf_hub_download(repo_id="briaai/BRIA-
|
137 |
-
hf_hub_download(repo_id="briaai/BRIA-
|
138 |
|
139 |
|
140 |
import torch
|
@@ -144,7 +144,7 @@ from pipeline_bria_controlnet import BriaControlNetPipeline
|
|
144 |
import PIL.Image as Image
|
145 |
|
146 |
base_model = 'briaai/BRIA-4B-Adapt'
|
147 |
-
controlnet_model = 'briaai/BRIA-
|
148 |
|
149 |
controlnet = BriaControlNetModel.from_pretrained(controlnet_model, torch_dtype=torch.bfloat16)
|
150 |
controlnet = BriaMultiControlNetModel([controlnet])
|
@@ -152,8 +152,8 @@ controlnet = BriaMultiControlNetModel([controlnet])
|
|
152 |
pipe = BriaControlNetPipeline.from_pretrained(base_model, controlnet=controlnet, torch_dtype=torch.bfloat16, trust_remote_code=True)
|
153 |
pipe.to("cuda")
|
154 |
|
155 |
-
control_image_colorgrid = load_image("https://huggingface.co/briaai/BRIA-
|
156 |
-
control_image_pose = load_image("https://huggingface.co/briaai/BRIA-
|
157 |
|
158 |
control_image = [control_image_colorgrid, control_image_pose]
|
159 |
controlnet_conditioning_scale = [0.5, 0.5]
|
@@ -161,7 +161,7 @@ control_mode = [2, 5]
|
|
161 |
|
162 |
width, height = control_image[0].size
|
163 |
|
164 |
-
prompt = 'Two kids in jackets play near a tent in a forest
|
165 |
|
166 |
generator = torch.Generator(device="cuda").manual_seed(555)
|
167 |
image = pipe(
|
@@ -174,7 +174,7 @@ image = pipe(
|
|
174 |
num_inference_steps=50,
|
175 |
max_sequence_length=128,
|
176 |
guidance_scale=5,
|
177 |
-
generator=generator
|
|
|
178 |
).images[0]
|
179 |
-
|
180 |
```
|
|
|
133 |
hf_hub_download(repo_id="briaai/BRIA-4B-Adapt", filename='pipeline_bria.py', local_dir=local_dir)
|
134 |
hf_hub_download(repo_id="briaai/BRIA-4B-Adapt", filename='transformer_bria.py', local_dir=local_dir)
|
135 |
hf_hub_download(repo_id="briaai/BRIA-4B-Adapt", filename='bria_utils.py', local_dir=local_dir)
|
136 |
+
hf_hub_download(repo_id="briaai/BRIA-3.0-ControlNet-Union", filename='pipeline_bria_controlnet.py', local_dir=local_dir)
|
137 |
+
hf_hub_download(repo_id="briaai/BRIA-3.0-ControlNet-Union", filename='controlnet_bria.py', local_dir=local_dir)
|
138 |
|
139 |
|
140 |
import torch
|
|
|
144 |
import PIL.Image as Image
|
145 |
|
146 |
base_model = 'briaai/BRIA-4B-Adapt'
|
147 |
+
controlnet_model = 'briaai/BRIA-3.0-ControlNet-Union'
|
148 |
|
149 |
controlnet = BriaControlNetModel.from_pretrained(controlnet_model, torch_dtype=torch.bfloat16)
|
150 |
controlnet = BriaMultiControlNetModel([controlnet])
|
|
|
152 |
pipe = BriaControlNetPipeline.from_pretrained(base_model, controlnet=controlnet, torch_dtype=torch.bfloat16, trust_remote_code=True)
|
153 |
pipe.to("cuda")
|
154 |
|
155 |
+
control_image_colorgrid = load_image("https://huggingface.co/briaai/BRIA-3.0-ControlNet-Union/resolve/main/colorgrid.jpg")
|
156 |
+
control_image_pose = load_image("https://huggingface.co/briaai/BRIA-3.0-ControlNet-Union/resolve/main/pose.jpg")
|
157 |
|
158 |
control_image = [control_image_colorgrid, control_image_pose]
|
159 |
controlnet_conditioning_scale = [0.5, 0.5]
|
|
|
161 |
|
162 |
width, height = control_image[0].size
|
163 |
|
164 |
+
prompt = 'Two kids in jackets play near a tent in a forest.'
|
165 |
|
166 |
generator = torch.Generator(device="cuda").manual_seed(555)
|
167 |
image = pipe(
|
|
|
174 |
num_inference_steps=50,
|
175 |
max_sequence_length=128,
|
176 |
guidance_scale=5,
|
177 |
+
generator=generator,
|
178 |
+
negative_prompt="Logo,Watermark,Text,Ugly,Morbid,Extra fingers,Poorly drawn hands,Mutation,Blurry,Extra limbs,Gross proportions,Missing arms,Mutated hands,Long neck,Duplicate"
|
179 |
).images[0]
|
|
|
180 |
```
|