|
|
@ -2321,15 +2321,14 @@ |
|
|
|
" 'use_scale_shift_norm': True,\n", |
|
|
|
" 'use_scale_shift_norm': True,\n", |
|
|
|
" })\n", |
|
|
|
" })\n", |
|
|
|
"\n", |
|
|
|
"\n", |
|
|
|
"secondary_model_ver = 2\n", |
|
|
|
|
|
|
|
"model_default = model_config['image_size']\n", |
|
|
|
"model_default = model_config['image_size']\n", |
|
|
|
"\n", |
|
|
|
"\n", |
|
|
|
"\n", |
|
|
|
"\n", |
|
|
|
"\n", |
|
|
|
"\n", |
|
|
|
"if secondary_model_ver == 2:\n", |
|
|
|
"if use_secondary_model:\n", |
|
|
|
" secondary_model = SecondaryDiffusionImageNet2()\n", |
|
|
|
" secondary_model = SecondaryDiffusionImageNet2()\n", |
|
|
|
" secondary_model.load_state_dict(torch.load(f'{model_path}/secondary_model_imagenet_2.pth', map_location='cpu'))\n", |
|
|
|
" secondary_model.load_state_dict(torch.load(f'{model_path}/secondary_model_imagenet_2.pth', map_location='cpu'))\n", |
|
|
|
"secondary_model.eval().requires_grad_(False).to(device)\n", |
|
|
|
" secondary_model.eval().requires_grad_(False).to(device)\n", |
|
|
|
"\n", |
|
|
|
"\n", |
|
|
|
"clip_models = []\n", |
|
|
|
"clip_models = []\n", |
|
|
|
"if ViTB32 is True: clip_models.append(clip.load('ViT-B/32', jit=False)[0].eval().requires_grad_(False).to(device)) \n", |
|
|
|
"if ViTB32 is True: clip_models.append(clip.load('ViT-B/32', jit=False)[0].eval().requires_grad_(False).to(device)) \n", |
|
|
|