准备环境和资源
stable-diffusion-webui-develop:1.0-pytorch1.13-gpu-py310-cu117-ubuntu22.04步骤一:在DSW中打开教程文件
进入PAI-DSW开发环境。
在页面左上方,选择DSW实例所在的地域。
在左侧导航栏单击工作空间列表,在工作空间列表页面中单击默认工作空间名称,进入对应工作空间内。
在左侧导航栏,选择模型开发与训练>交互式建模(DSW)。
单击需要打开的实例操作列下的打开,进入PAI-DSW实例开发环境。
在Notebook页签的Launcher页面,单击快速开始区域Tool下的DSW Gallery,打开DSW Gallery页面。

在DSW Gallery页面中,搜索并找到用AI重燃亚运经典教程,单击教程卡片中的在DSW中打开。
单击后即会自动将本教程所需的资源和教程文件下载至DSW实例中,并在下载完成后自动打开教程文件。
步骤二:运行教程文件
在打开的教程文件image_restoration.ipynb文件中,您可以直接看到教程文本,您可以在教程文件中直接运行对应的步骤的命令,当成功运行结束一个步骤命令后,再顺次运行下个步骤的命令。本教程包含的操作步骤以及每个步骤的执行结果如下。
Get:1 http://mirrors.cloud.aliyuncs.com/ubuntu jammy InRelease [270 kB]
Get:2 http://mirrors.cloud.aliyuncs.com/ubuntu jammy-updates InRelease [119 kB]
Get:3 http://mirrors.cloud.aliyuncs.com/ubuntu jammy-backports InRelease [109 kB]
Get:4 http://mirrors.cloud.aliyuncs.com/ubuntu jammy-security InRelease [110 kB]
Get:5 http://mirrors.cloud.aliyuncs.com/ubuntu jammy/multiverse Sources [361 kB]
Get:6 http://mirrors.cloud.aliyuncs.com/ubuntu jammy/main Sources [1668 kB]
Get:7 http://mirrors.cloud.aliyuncs.com/ubuntu jammy/restricted Sources [28.2 kB]
Get:8 http://mirrors.cloud.aliyuncs.com/ubuntu jammy/universe Sources [22.0 MB]
Get:9 http://mirrors.cloud.aliyuncs.com/ubuntu jammy/universe amd64 Packages [17.5 MB]
Get:10 http://mirrors.cloud.aliyuncs.com/ubuntu jammy/multiverse amd64 Packages [266 kB]
Get:11 http://mirrors.cloud.aliyuncs.com/ubuntu jammy/restricted amd64 Packages [164 kB]
Get:12 http://mirrors.cloud.aliyuncs.com/ubuntu jammy/main amd64 Packages [1792 kB]
Get:13 http://mirrors.cloud.aliyuncs.com/ubuntu jammy-updates/multiverse Sources [21.0 kB]
Get:14 http://mirrors.cloud.aliyuncs.com/ubuntu jammy-updates/universe Sources [347 kB]
Get:15 http://mirrors.cloud.aliyuncs.com/ubuntu jammy-updates/main Sources [531 kB]
Get:16 http://mirrors.cloud.aliyuncs.com/ubuntu jammy-updates/restricted Sources [56.3 kB]
Get:17 http://mirrors.cloud.aliyuncs.com/ubuntu jammy-updates/restricted amd64 Packages [1015 kB]
Get:18 http://mirrors.cloud.aliyuncs.com/ubuntu jammy-updates/main amd64 Packages [1185 kB]
Get:19 http://mirrors.cloud.aliyuncs.com/ubuntu jammy-updates/universe amd64 Packages [1251 kB]
Get:20 http://mirrors.cloud.aliyuncs.com/ubuntu jammy-updates/multiverse amd64 Packages [49.8 kB]
Get:21 http://mirrors.cloud.aliyuncs.com/ubuntu jammy-backports/main Sources [9392 B]
Get:22 http://mirrors.cloud.aliyuncs.com/ubuntu jammy-backports/universe Sources [10.5 kB]
Get:23 http://mirrors.cloud.aliyuncs.com/ubuntu jammy-backports/main amd64 Packages [50.3 kB]
Get:24 http://mirrors.cloud.aliyuncs.com/ubuntu jammy-backports/universe amd64 Packages [28.1 kB]
Get:25 http://mirrors.cloud.aliyuncs.com/ubuntu jammy-security/restricted Sources [53.4 kB]
Get:26 http://mirrors.cloud.aliyuncs.com/ubuntu jammy-security/universe Sources [202 kB]
Get:27 http://mirrors.cloud.aliyuncs.com/ubuntu jammy-security/multiverse Sources [11.3 kB]
Get:28 http://mirrors.cloud.aliyuncs.com/ubuntu jammy-security/main Sources [270 kB]
Get:29 http://mirrors.cloud.aliyuncs.com/ubuntu jammy-security/main amd64 Packages [915 kB]
Get:30 http://mirrors.cloud.aliyuncs.com/ubuntu jammy-security/multiverse amd64 Packages [44.0 kB]
Get:31 http://mirrors.cloud.aliyuncs.com/ubuntu jammy-security/restricted amd64 Packages [995 kB]
Get:32 http://mirrors.cloud.aliyuncs.com/ubuntu jammy-security/universe amd64 Packages [990 kB]
Fetched 52.4 MB in 2s (26.0 MB/s)
Reading package lists... Done
Reading package lists... Done
Building dependency tree... Done
Reading state information... Done
Suggested packages:
zip
The following NEW packages will be installed:
unzip
0 upgraded, 1 newly installed, 0 to remove and 99 not upgraded.
Need to get 174 kB of archives.
After this operation, 385 kB of additional disk space will be used.
Get:1 http://mirrors.cloud.aliyuncs.com/ubuntu jammy-updates/main amd64 unzip amd64 6.0-26ubuntu3.1 [174 kB]
Fetched 174 kB in 0s (7160 kB/s)
debconf: delaying package configuration, since apt-utils is not installed
Selecting previously unselected package unzip.
(Reading database ... 20089 files and directories currently installed.)
Preparing to unpack .../unzip_6.0-26ubuntu3.1_amd64.deb ...
Unpacking unzip (6.0-26ubuntu3.1) ...
Setting up unzip (6.0-26ubuntu3.1) ...http://pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com/aigc-data/restoration/img/input.zip
cn-hangzhou
--2023-09-04 11:20:55-- http://pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com/aigc-data/restoration/img/input.zip
Resolving pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com (pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com)... 100.118.28.49, 100.118.28.45, 100.118.28.44, ...
Connecting to pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com (pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com)|100.118.28.49|:80... connected.
HTTP request sent, awaiting response... 200 OK
Length: 1657458 (1.6M) [application/zip]
Saving to: ‘input.zip’
input.zip 100%[===================>] 1.58M --.-KB/s in 0.09s
2023-09-04 11:20:56 (16.8 MB/s) - ‘input.zip’ saved [1657458/1657458]
Archive: input.zip
creating: input/
inflating: input/54.jpg
inflating: input/20.jpg
inflating: input/10.jpg
inflating: input/50.jpg
inflating: input/2.jpg
inflating: input/40.jpg
inflating: input/4.png
inflating: input/70.jpg
inflating: input/34.jpg
inflating: input/64.jpg http://pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com/aigc-data/restoration/repo/restormer.zip
cn-hangzhou
--2023-09-04 11:30:30-- http://pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com/aigc-data/restoration/repo/restormer.zip
Resolving pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com (pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com)... 100.118.28.49, 100.118.28.45, 100.118.28.44, ...
Connecting to pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com (pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com)|100.118.28.49|:80... connected.
HTTP request sent, awaiting response... 200 OK
Length: 2859106485 (2.7G) [application/zip]
Saving to: ‘restormer.zip’
restormer.zip 100%[===================>] 2.66G 14.2MB/s in 3m 18s
2023-09-04 11:33:48 (13.8 MB/s) - ‘restormer.zip’ saved [2859106485/2859106485]
Archive: restormer.zip
creating: Restormer/
creating: Restormer/.ipynb_checkpoints/
inflating: Restormer/.ipynb_checkpoints/demo-checkpoint.py
inflating: Restormer/setup.cfg
creating: Restormer/Denoising/
inflating: Restormer/Denoising/test_real_denoising_dnd.py
creating: Restormer/Denoising/pretrained_models/
inflating: Restormer/Denoising/pretrained_models/gaussian_gray_denoising_sigma25.pth
inflating: Restormer/Denoising/pretrained_models/real_denoising.pth
inflating: Restormer/Denoising/pretrained_models/gaussian_color_denoising_blind.pth
inflating: Restormer/Denoising/pretrained_models/gaussian_gray_denoising_sigma15.pth
inflating: Restormer/Denoising/pretrained_models/gaussian_color_denoising_sigma50.pth
inflating: Restormer/Denoising/pretrained_models/gaussian_color_denoising_sigma25.pth
inflating: Restormer/Denoising/pretrained_models/gaussian_color_denoising_sigma15.pth
inflating: Restormer/Denoising/pretrained_models/gaussian_gray_denoising_blind.pth
inflating: Restormer/Denoising/pretrained_models/gaussian_gray_denoising_sigma50.pth
inflating: Restormer/Denoising/test_gaussian_color_denoising.py
inflating: Restormer/Denoising/evaluate_gaussian_gray_denoising.py
inflating: Restormer/Denoising/test_real_denoising_sidd.py
creating: Restormer/Denoising/Datasets/
inflating: Restormer/Denoising/Datasets/README.md
inflating: Restormer/Denoising/generate_patches_sidd.py
inflating: Restormer/Denoising/generate_patches_dfwb.py
creating: Restormer/Denoising/Options/
inflating: Restormer/Denoising/Options/GaussianColorDenoising_RestormerSigma50.yml
inflating: Restormer/Denoising/Options/GaussianGrayDenoising_Restormer.yml
inflating: Restormer/Denoising/Options/GaussianColorDenoising_Restormer.yml
inflating: Restormer/Denoising/Options/GaussianColorDenoising_RestormerSigma15.yml
inflating: Restormer/Denoising/Options/GaussianGrayDenoising_RestormerSigma25.yml
inflating: Restormer/Denoising/Options/GaussianGrayDenoising_RestormerSigma15.yml
inflating: Restormer/Denoising/Options/GaussianGrayDenoising_RestormerSigma50.yml
inflating: Restormer/Denoising/Options/GaussianColorDenoising_RestormerSigma25.yml
inflating: Restormer/Denoising/Options/RealDenoising_Restormer.yml
inflating: Restormer/Denoising/evaluate_gaussian_color_denoising.py
inflating: Restormer/Denoising/README.md
inflating: Restormer/Denoising/test_gaussian_gray_denoising.py
inflating: Restormer/Denoising/evaluate_sidd.m
inflating: Restormer/Denoising/download_data.py
inflating: Restormer/Denoising/utils.py
extracting: Restormer/hat.zip
creating: Restormer/Deraining/
creating: Restormer/Deraining/pretrained_models/
inflating: Restormer/Deraining/pretrained_models/deraining.pth
inflating: Restormer/Deraining/pretrained_models/README.md
inflating: Restormer/Deraining/test.py
creating: Restormer/Deraining/Datasets/
inflating: Restormer/Deraining/Datasets/README.md
creating: Restormer/Deraining/Options/
inflating: Restormer/Deraining/Options/Deraining_Restormer.yml
inflating: Restormer/Deraining/README.md
inflating: Restormer/Deraining/evaluate_PSNR_SSIM.m
inflating: Restormer/Deraining/download_data.py
inflating: Restormer/Deraining/utils.py
creating: Restormer/demo/
creating: Restormer/demo/restored/
creating: Restormer/demo/restored/Single_Image_Defocus_Deblurring/
inflating: Restormer/demo/restored/Single_Image_Defocus_Deblurring/couple.png
inflating: Restormer/demo/restored/Single_Image_Defocus_Deblurring/engagement.png
inflating: Restormer/demo/restored/Single_Image_Defocus_Deblurring/portrait.png
creating: Restormer/demo/degraded/
inflating: Restormer/demo/degraded/couple.jpg
inflating: Restormer/demo/degraded/portrait.jpg
inflating: Restormer/demo/degraded/engagement.jpg
inflating: Restormer/.gitignore
inflating: Restormer/LICENSE.md
inflating: Restormer/demo.py
extracting: Restormer/Resformer_pretrain.zip
creating: Restormer/Motion_Deblurring/
inflating: Restormer/Motion_Deblurring/evaluate_gopro_hide.m
creating: Restormer/Motion_Deblurring/pretrained_models/
inflating: Restormer/Motion_Deblurring/pretrained_models/motion_deblurring.pth
inflating: Restormer/Motion_Deblurring/pretrained_models/README.md
inflating: Restormer/Motion_Deblurring/test.py
creating: Restormer/Motion_Deblurring/Datasets/
inflating: Restormer/Motion_Deblurring/Datasets/README.md
creating: Restormer/Motion_Deblurring/Options/
inflating: Restormer/Motion_Deblurring/Options/Deblurring_Restormer.yml
inflating: Restormer/Motion_Deblurring/README.md
inflating: Restormer/Motion_Deblurring/evaluate_realblur.py
inflating: Restormer/Motion_Deblurring/generate_patches_gopro.py
inflating: Restormer/Motion_Deblurring/download_data.py
inflating: Restormer/Motion_Deblurring/utils.py
creating: Restormer/basicsr/
creating: Restormer/basicsr/models/
inflating: Restormer/basicsr/models/base_model.py
creating: Restormer/basicsr/models/archs/
inflating: Restormer/basicsr/models/archs/restormer_arch.py
inflating: Restormer/basicsr/models/archs/arch_util.py
inflating: Restormer/basicsr/models/archs/__init__.py
inflating: Restormer/basicsr/models/lr_scheduler.py
creating: Restormer/basicsr/models/losses/
inflating: Restormer/basicsr/models/losses/loss_util.py
inflating: Restormer/basicsr/models/losses/__init__.py
inflating: Restormer/basicsr/models/losses/losses.py
inflating: Restormer/basicsr/models/__init__.py
inflating: Restormer/basicsr/models/image_restoration_model.py
inflating: Restormer/basicsr/train.py
inflating: Restormer/basicsr/version.py
inflating: Restormer/basicsr/test.py
creating: Restormer/basicsr/utils/
inflating: Restormer/basicsr/utils/bundle_submissions.py
inflating: Restormer/basicsr/utils/file_client.py
inflating: Restormer/basicsr/utils/face_util.py
inflating: Restormer/basicsr/utils/create_lmdb.py
inflating: Restormer/basicsr/utils/logger.py
inflating: Restormer/basicsr/utils/options.py
inflating: Restormer/basicsr/utils/img_util.py
inflating: Restormer/basicsr/utils/matlab_functions.py
inflating: Restormer/basicsr/utils/download_util.py
inflating: Restormer/basicsr/utils/__init__.py
inflating: Restormer/basicsr/utils/misc.py
inflating: Restormer/basicsr/utils/dist_util.py
inflating: Restormer/basicsr/utils/lmdb_util.py
inflating: Restormer/basicsr/utils/flow_util.py
creating: Restormer/basicsr/metrics/
inflating: Restormer/basicsr/metrics/fid.py
inflating: Restormer/basicsr/metrics/metric_util.py
inflating: Restormer/basicsr/metrics/niqe_pris_params.npz
inflating: Restormer/basicsr/metrics/psnr_ssim.py
inflating: Restormer/basicsr/metrics/niqe.py
inflating: Restormer/basicsr/metrics/__init__.py
creating: Restormer/basicsr/data/
creating: Restormer/basicsr/data/meta_info/
inflating: Restormer/basicsr/data/meta_info/meta_info_Vimeo90K_test_medium_GT.txt
inflating: Restormer/basicsr/data/meta_info/meta_info_REDS4_test_GT.txt
inflating: Restormer/basicsr/data/meta_info/meta_info_Vimeo90K_test_GT.txt
inflating: Restormer/basicsr/data/meta_info/meta_info_DIV2K800sub_GT.txt
inflating: Restormer/basicsr/data/meta_info/meta_info_REDS_GT.txt
inflating: Restormer/basicsr/data/meta_info/meta_info_Vimeo90K_test_slow_GT.txt
inflating: Restormer/basicsr/data/meta_info/meta_info_REDSofficial4_test_GT.txt
inflating: Restormer/basicsr/data/meta_info/meta_info_Vimeo90K_test_fast_GT.txt
inflating: Restormer/basicsr/data/meta_info/meta_info_Vimeo90K_train_GT.txt
inflating: Restormer/basicsr/data/meta_info/meta_info_REDSval_official_test_GT.txt
inflating: Restormer/basicsr/data/paired_image_dataset.py
inflating: Restormer/basicsr/data/data_util.py
inflating: Restormer/basicsr/data/reds_dataset.py
inflating: Restormer/basicsr/data/video_test_dataset.py
inflating: Restormer/basicsr/data/single_image_dataset.py
inflating: Restormer/basicsr/data/vimeo90k_dataset.py
inflating: Restormer/basicsr/data/__init__.py
inflating: Restormer/basicsr/data/prefetch_dataloader.py
inflating: Restormer/basicsr/data/ffhq_dataset.py
inflating: Restormer/basicsr/data/transforms.py
inflating: Restormer/basicsr/data/data_sampler.py
creating: Restormer/pretrained_models_demotion/
inflating: Restormer/pretrained_models_demotion/motion_deblurring.pth
creating: Restormer/.git/
creating: Restormer/.git/logs/
creating: Restormer/.git/logs/refs/
creating: Restormer/.git/logs/refs/remotes/
creating: Restormer/.git/logs/refs/remotes/origin/
inflating: Restormer/.git/logs/refs/remotes/origin/HEAD
creating: Restormer/.git/logs/refs/heads/
inflating: Restormer/.git/logs/refs/heads/main
inflating: Restormer/.git/logs/HEAD
inflating: Restormer/.git/config
creating: Restormer/.git/refs/
creating: Restormer/.git/refs/remotes/
creating: Restormer/.git/refs/remotes/origin/
extracting: Restormer/.git/refs/remotes/origin/HEAD
creating: Restormer/.git/refs/heads/
extracting: Restormer/.git/refs/heads/main
creating: Restormer/.git/refs/tags/
extracting: Restormer/.git/HEAD
creating: Restormer/.git/hooks/
inflating: Restormer/.git/hooks/push-to-checkout.sample
inflating: Restormer/.git/hooks/commit-msg.sample
inflating: Restormer/.git/hooks/applypatch-msg.sample
inflating: Restormer/.git/hooks/pre-receive.sample
inflating: Restormer/.git/hooks/pre-push.sample
inflating: Restormer/.git/hooks/fsmonitor-watchman.sample
inflating: Restormer/.git/hooks/post-update.sample
inflating: Restormer/.git/hooks/update.sample
inflating: Restormer/.git/hooks/pre-merge-commit.sample
inflating: Restormer/.git/hooks/pre-commit.sample
inflating: Restormer/.git/hooks/pre-applypatch.sample
inflating: Restormer/.git/hooks/prepare-commit-msg.sample
inflating: Restormer/.git/hooks/pre-rebase.sample
creating: Restormer/.git/info/
inflating: Restormer/.git/info/exclude
creating: Restormer/.git/objects/
creating: Restormer/.git/objects/pack/
inflating: Restormer/.git/objects/pack/pack-c5acb539fe9f4f374066c96759a798aee30d3def.pack
inflating: Restormer/.git/objects/pack/pack-c5acb539fe9f4f374066c96759a798aee30d3def.idx
creating: Restormer/.git/objects/info/
inflating: Restormer/.git/packed-refs
inflating: Restormer/.git/index
inflating: Restormer/.git/description
creating: Restormer/pretrained_models_defocus_deblur/
inflating: Restormer/pretrained_models_defocus_deblur/single_image_defocus_deblurring.pth
inflating: Restormer/pretrained_models_defocus_deblur/dual_pixel_defocus_deblurring.pth
creating: Restormer/Defocus_Deblurring/
creating: Restormer/Defocus_Deblurring/pretrained_models/
inflating: Restormer/Defocus_Deblurring/pretrained_models/single_image_defocus_deblurring.pth
inflating: Restormer/Defocus_Deblurring/pretrained_models/README.md
inflating: Restormer/Defocus_Deblurring/generate_patches_dpdd.py
creating: Restormer/Defocus_Deblurring/Datasets/
inflating: Restormer/Defocus_Deblurring/Datasets/README.md
creating: Restormer/Defocus_Deblurring/Options/
inflating: Restormer/Defocus_Deblurring/Options/DefocusDeblur_DualPixel_16bit_Restormer.yml
inflating: Restormer/Defocus_Deblurring/Options/DefocusDeblur_Single_8bit_Restormer.yml
inflating: Restormer/Defocus_Deblurring/README.md
inflating: Restormer/Defocus_Deblurring/test_single_image_defocus_deblur.py
inflating: Restormer/Defocus_Deblurring/download_data.py
inflating: Restormer/Defocus_Deblurring/utils.py
inflating: Restormer/Defocus_Deblurring/test_dual_pixel_defocus_deblur.py
inflating: Restormer/README.md
extracting: Restormer/VERSION
inflating: Restormer/setup.py
creating: Restormer/pretrained_models_derain/
inflating: Restormer/pretrained_models_derain/deraining.pth
inflating: Restormer/train.sh
inflating: Restormer/INSTALL.md
creating: Restormer/__MACOSX/
creating: Restormer/__MACOSX/pretrained_models_demotion/
inflating: Restormer/__MACOSX/pretrained_models_demotion/._motion_deblurring.pth
creating: Restormer/__MACOSX/pretrained_models_defocus_deblur/
inflating: Restormer/__MACOSX/pretrained_models_defocus_deblur/._single_image_defocus_deblurring.pth
inflating: Restormer/__MACOSX/pretrained_models_defocus_deblur/._dual_pixel_defocus_deblurring.pth
inflating: Restormer/__MACOSX/._pretrained_models_denoise
inflating: Restormer/__MACOSX/._pretrained_models_defocus_deblur
creating: Restormer/__MACOSX/pretrained_models_derain/
inflating: Restormer/__MACOSX/pretrained_models_derain/._deraining.pth
inflating: Restormer/__MACOSX/._pretrained_models_derain
inflating: Restormer/__MACOSX/._pretrained_models_demotion
creating: Restormer/__MACOSX/pretrained_models_denoise/
inflating: Restormer/__MACOSX/pretrained_models_denoise/._gaussian_gray_denoising_sigma25.pth
inflating: Restormer/__MACOSX/pretrained_models_denoise/._gaussian_gray_denoising_sigma50.pth
inflating: Restormer/__MACOSX/pretrained_models_denoise/._gaussian_color_denoising_sigma50.pth
inflating: Restormer/__MACOSX/pretrained_models_denoise/._gaussian_color_denoising_sigma25.pth
inflating: Restormer/__MACOSX/pretrained_models_denoise/._gaussian_gray_denoising_sigma15.pth
inflating: Restormer/__MACOSX/pretrained_models_denoise/._gaussian_gray_denoising_blind.pth
inflating: Restormer/__MACOSX/pretrained_models_denoise/._gaussian_color_denoising_blind.pth
inflating: Restormer/__MACOSX/pretrained_models_denoise/._real_denoising.pth
inflating: Restormer/__MACOSX/pretrained_models_denoise/._gaussian_color_denoising_sigma15.pth Looking in indexes: https://mirrors.cloud.aliyuncs.com/pypi/simple
Collecting natsort
Downloading https://mirrors.cloud.aliyuncs.com/pypi/packages/ef/82/7a9d0550484a62c6da82858ee9419f3dd1ccc9aa1c26a1e43da3ecd20b0d/natsort-8.4.0-py3-none-any.whl (38 kB)
Installing collected packages: natsort
Successfully installed natsort-8.4.0
WARNING: Running pip as the 'root' user can result in broken permissions and conflicting behaviour with the system package manager. It is recommended to use a virtual environment instead: https://pip.pypa.io/warnings/venv
[notice] A new release of pip is available: 23.0.1 -> 23.2.1
[notice] To update, run: python3 -m pip install --upgrade pip==> Running Motion_Deblurring with weights /mnt/workspace/Restormer/Motion_Deblurring/pretrained_models/motion_deblurring.pth
100%|███████████████████████████████████████████| 10/10 [00:28<00:00, 2.81s/it]
Restored images are saved at results/Motion_Deblurring
==> Running Single_Image_Defocus_Deblurring with weights /mnt/workspace/Restormer/Defocus_Deblurring/pretrained_models/single_image_defocus_deblurring.pth
100%|███████████████████████████████████████████| 10/10 [00:26<00:00, 2.64s/it]
Restored images are saved at results/Single_Image_Defocus_Deblurring
==> Running Deraining with weights /mnt/workspace/Restormer/Deraining/pretrained_models/deraining.pth
100%|███████████████████████████████████████████| 10/10 [00:26<00:00, 2.65s/it]
Restored images are saved at results/Deraining
==> Running Real_Denoising with weights /mnt/workspace/Restormer/Denoising/pretrained_models/real_denoising.pth
100%|███████████████████████████████████████████| 10/10 [00:24<00:00, 2.48s/it]
Restored images are saved at results/Real_Denoising
==> Running Gaussian_Gray_Denoising with weights /mnt/workspace/Restormer/Denoising/pretrained_models/gaussian_gray_denoising_blind.pth
100%|███████████████████████████████████████████| 10/10 [00:24<00:00, 2.45s/it]
Restored images are saved at results/Gaussian_Gray_Denoising
==> Running Gaussian_Color_Denoising with weights /mnt/workspace/Restormer/Denoising/pretrained_models/gaussian_color_denoising_blind.pth
100%|███████████████████████████████████████████| 10/10 [00:24<00:00, 2.50s/it]
Restored images are saved at results/Gaussian_Color_DenoisingLooking in indexes: https://mirrors.cloud.aliyuncs.com/pypi/simple
Collecting modelscope
Downloading https://mirrors.cloud.aliyuncs.com/pypi/packages/2b/17/53845f398e340217ecb2169033f547f640db3663ffb0fb5a6218c5f3bfec/modelscope-1.8.4-py3-none-any.whl (4.9 MB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 4.9/4.9 MB 42.9 MB/s eta 0:00:0000:0100:01
Collecting gast>=0.2.2
Downloading https://mirrors.cloud.aliyuncs.com/pypi/packages/fa/39/5aae571e5a5f4de9c3445dae08a530498e5c53b0e74410eeeb0991c79047/gast-0.5.4-py3-none-any.whl (19 kB)
Requirement already satisfied: urllib3>=1.26 in /usr/local/lib/python3.10/dist-packages (from modelscope) (1.26.15)
Requirement already satisfied: pyyaml in /usr/local/lib/python3.10/dist-packages (from modelscope) (6.0)
Requirement already satisfied: tqdm>=4.64.0 in /usr/local/lib/python3.10/dist-packages (from modelscope) (4.65.0)
Collecting sortedcontainers>=1.5.9
Downloading https://mirrors.cloud.aliyuncs.com/pypi/packages/32/46/9cb0e58b2deb7f82b84065f37f3bffeb12413f947f9388e4cac22c4621ce/sortedcontainers-2.4.0-py2.py3-none-any.whl (29 kB)
Requirement already satisfied: Pillow>=6.2.0 in /usr/local/lib/python3.10/dist-packages (from modelscope) (9.4.0)
Requirement already satisfied: pyarrow!=9.0.0,>=6.0.0 in /usr/local/lib/python3.10/dist-packages (from modelscope) (11.0.0)
Collecting oss2
Downloading https://mirrors.cloud.aliyuncs.com/pypi/packages/4a/e7/08b90651a435acde68c537eebff970011422f61c465f6d1c88c4b3af6774/oss2-2.18.1.tar.gz (274 kB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 274.3/274.3 kB 73.3 MB/s eta 0:00:00
Preparing metadata (setup.py) ... done
Requirement already satisfied: setuptools in /usr/lib/python3/dist-packages (from modelscope) (59.6.0)
Requirement already satisfied: yapf in /usr/local/lib/python3.10/dist-packages (from modelscope) (0.32.0)
Requirement already satisfied: attrs in /usr/local/lib/python3.10/dist-packages (from modelscope) (22.2.0)
Requirement already satisfied: numpy in /usr/local/lib/python3.10/dist-packages (from modelscope) (1.23.3)
Requirement already satisfied: python-dateutil>=2.1 in /usr/local/lib/python3.10/dist-packages (from modelscope) (2.8.2)
Collecting simplejson>=3.3.0
Downloading https://mirrors.cloud.aliyuncs.com/pypi/packages/b8/00/9720ea26c0da200a39b89e25721970f50f4b80bb2ab6de0199324f93c4ca/simplejson-3.19.1-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (137 kB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 137.9/137.9 kB 48.0 MB/s eta 0:00:00
Requirement already satisfied: datasets<=2.13.0,>=2.8.0 in /usr/local/lib/python3.10/dist-packages (from modelscope) (2.11.0)
Requirement already satisfied: einops in /usr/local/lib/python3.10/dist-packages (from modelscope) (0.4.1)
Requirement already satisfied: addict in /usr/local/lib/python3.10/dist-packages (from modelscope) (2.4.0)
Requirement already satisfied: requests>=2.25 in /usr/local/lib/python3.10/dist-packages (from modelscope) (2.25.1)
Requirement already satisfied: pandas in /usr/local/lib/python3.10/dist-packages (from modelscope) (1.5.3)
Requirement already satisfied: filelock>=3.3.0 in /usr/local/lib/python3.10/dist-packages (from modelscope) (3.10.7)
Requirement already satisfied: scipy in /usr/local/lib/python3.10/dist-packages (from modelscope) (1.10.1)
Requirement already satisfied: huggingface-hub<1.0.0,>=0.11.0 in /usr/local/lib/python3.10/dist-packages (from datasets<=2.13.0,>=2.8.0->modelscope) (0.13.3)
Requirement already satisfied: packaging in /usr/local/lib/python3.10/dist-packages (from datasets<=2.13.0,>=2.8.0->modelscope) (23.0)
Requirement already satisfied: multiprocess in /usr/local/lib/python3.10/dist-packages (from datasets<=2.13.0,>=2.8.0->modelscope) (0.70.14)
Requirement already satisfied: xxhash in /usr/local/lib/python3.10/dist-packages (from datasets<=2.13.0,>=2.8.0->modelscope) (3.2.0)
Requirement already satisfied: dill<0.3.7,>=0.3.0 in /usr/local/lib/python3.10/dist-packages (from datasets<=2.13.0,>=2.8.0->modelscope) (0.3.6)
Requirement already satisfied: aiohttp in /usr/local/lib/python3.10/dist-packages (from datasets<=2.13.0,>=2.8.0->modelscope) (3.8.4)
Requirement already satisfied: fsspec[http]>=2021.11.1 in /usr/local/lib/python3.10/dist-packages (from datasets<=2.13.0,>=2.8.0->modelscope) (2023.3.0)
Requirement already satisfied: responses<0.19 in /usr/local/lib/python3.10/dist-packages (from datasets<=2.13.0,>=2.8.0->modelscope) (0.18.0)
Requirement already satisfied: six>=1.5 in /usr/local/lib/python3.10/dist-packages (from python-dateutil>=2.1->modelscope) (1.16.0)
Requirement already satisfied: certifi>=2017.4.17 in /usr/local/lib/python3.10/dist-packages (from requests>=2.25->modelscope) (2022.12.7)
Requirement already satisfied: chardet<5,>=3.0.2 in /usr/local/lib/python3.10/dist-packages (from requests>=2.25->modelscope) (4.0.0)
Requirement already satisfied: idna<3,>=2.5 in /usr/local/lib/python3.10/dist-packages (from requests>=2.25->modelscope) (2.10)
Collecting aliyun-python-sdk-core>=2.13.12
Downloading https://mirrors.cloud.aliyuncs.com/pypi/packages/55/5a/6eec6c6e78817e5ca2afee661f2bbb33dbcfa2ce09a2980b52223323bd2e/aliyun-python-sdk-core-2.13.36.tar.gz (440 kB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 440.5/440.5 kB 26.0 MB/s eta 0:00:00
Preparing metadata (setup.py) ... done
Collecting aliyun-python-sdk-kms>=2.4.1
Downloading https://mirrors.cloud.aliyuncs.com/pypi/packages/13/90/02d05478df643ceac0021bd3db4f19b42dd06c2b73e082569d0d340de70c/aliyun_python_sdk_kms-2.16.1-py2.py3-none-any.whl (70 kB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 70.8/70.8 kB 28.1 MB/s eta 0:00:00
Collecting crcmod>=1.7
Downloading https://mirrors.cloud.aliyuncs.com/pypi/packages/6b/b0/e595ce2a2527e169c3bcd6c33d2473c1918e0b7f6826a043ca1245dd4e5b/crcmod-1.7.tar.gz (89 kB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 89.7/89.7 kB 31.2 MB/s eta 0:00:00
Preparing metadata (setup.py) ... done
Requirement already satisfied: pycryptodome>=3.4.7 in /usr/local/lib/python3.10/dist-packages (from oss2->modelscope) (3.17)
Requirement already satisfied: pytz>=2020.1 in /usr/local/lib/python3.10/dist-packages (from pandas->modelscope) (2023.3)
Requirement already satisfied: cryptography>=2.6.0 in /usr/local/lib/python3.10/dist-packages (from aliyun-python-sdk-core>=2.13.12->oss2->modelscope) (40.0.1)
Collecting jmespath<1.0.0,>=0.9.3
Downloading https://mirrors.cloud.aliyuncs.com/pypi/packages/07/cb/5f001272b6faeb23c1c9e0acc04d48eaaf5c862c17709d20e3469c6e0139/jmespath-0.10.0-py2.py3-none-any.whl (24 kB)
Requirement already satisfied: frozenlist>=1.1.1 in /usr/local/lib/python3.10/dist-packages (from aiohttp->datasets<=2.13.0,>=2.8.0->modelscope) (1.3.3)
Requirement already satisfied: yarl<2.0,>=1.0 in /usr/local/lib/python3.10/dist-packages (from aiohttp->datasets<=2.13.0,>=2.8.0->modelscope) (1.8.2)
Requirement already satisfied: charset-normalizer<4.0,>=2.0 in /usr/local/lib/python3.10/dist-packages (from aiohttp->datasets<=2.13.0,>=2.8.0->modelscope) (3.1.0)
Requirement already satisfied: multidict<7.0,>=4.5 in /usr/local/lib/python3.10/dist-packages (from aiohttp->datasets<=2.13.0,>=2.8.0->modelscope) (6.0.4)
Requirement already satisfied: async-timeout<5.0,>=4.0.0a3 in /usr/local/lib/python3.10/dist-packages (from aiohttp->datasets<=2.13.0,>=2.8.0->modelscope) (4.0.2)
Requirement already satisfied: aiosignal>=1.1.2 in /usr/local/lib/python3.10/dist-packages (from aiohttp->datasets<=2.13.0,>=2.8.0->modelscope) (1.3.1)
Requirement already satisfied: typing-extensions>=3.7.4.3 in /usr/local/lib/python3.10/dist-packages (from huggingface-hub<1.0.0,>=0.11.0->datasets<=2.13.0,>=2.8.0->modelscope) (4.5.0)
Requirement already satisfied: cffi>=1.12 in /usr/local/lib/python3.10/dist-packages (from cryptography>=2.6.0->aliyun-python-sdk-core>=2.13.12->oss2->modelscope) (1.15.1)
Requirement already satisfied: pycparser in /usr/local/lib/python3.10/dist-packages (from cffi>=1.12->cryptography>=2.6.0->aliyun-python-sdk-core>=2.13.12->oss2->modelscope) (2.21)
Building wheels for collected packages: oss2, aliyun-python-sdk-core, crcmod
Building wheel for oss2 (setup.py) ... done
Created wheel for oss2: filename=oss2-2.18.1-py3-none-any.whl size=115202 sha256=f9143856d3968608c696d2556f1f64953f89a31f7931d695e5f09ffe96a10fe4
Stored in directory: /root/.cache/pip/wheels/2b/e2/2d/a55f3aabf369a023a14d1fc570dd3a3824cdd2223f0f73e902
Building wheel for aliyun-python-sdk-core (setup.py) ... done
Created wheel for aliyun-python-sdk-core: filename=aliyun_python_sdk_core-2.13.36-py3-none-any.whl size=533196 sha256=5ec4012574fc4cb6e83ed02aebc59fe9747c1a536e346e07e7eed5073e3a3be9
Stored in directory: /root/.cache/pip/wheels/0b/4f/1c/459b3309c6370bdaa926bb358dbc1b42aa0e0a26c0476ac401
Building wheel for crcmod (setup.py) ... done
Created wheel for crcmod: filename=crcmod-1.7-cp310-cp310-linux_x86_64.whl size=31428 sha256=5078b5859e75ffe6d90a36bac0f4db76801b4db5a04e0f5a5523912ee5e65447
Stored in directory: /root/.cache/pip/wheels/3d/1b/33/88c98fcba7f84f7d448c44de21caf3e98deaebae12cb5104ba
Successfully built oss2 aliyun-python-sdk-core crcmod
Installing collected packages: sortedcontainers, crcmod, simplejson, jmespath, gast, aliyun-python-sdk-core, aliyun-python-sdk-kms, oss2, modelscope
Successfully installed aliyun-python-sdk-core-2.13.36 aliyun-python-sdk-kms-2.16.1 crcmod-1.7 gast-0.5.4 jmespath-0.10.0 modelscope-1.8.4 oss2-2.18.1 simplejson-3.19.1 sortedcontainers-2.4.0
WARNING: Running pip as the 'root' user can result in broken permissions and conflicting behaviour with the system package manager. It is recommended to use a virtual environment instead: https://pip.pypa.io/warnings/venv
[notice] A new release of pip is available: 23.0.1 -> 23.2.1
[notice] To update, run: python3 -m pip install --upgrade pip
http://pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com/aigc-data/restoration/repo/nafnet.zip
cn-hangzhou
--2023-09-04 11:45:09-- http://pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com/aigc-data/restoration/repo/nafnet.zip
Resolving pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com (pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com)... 100.118.28.49, 100.118.28.44, 100.118.28.45, ...
Connecting to pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com (pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com)|100.118.28.49|:80... connected.
HTTP request sent, awaiting response... 200 OK
Length: 943903968 (900M) [application/zip]
Saving to: ‘nafnet.zip’
nafnet.zip 100%[===================>] 900.18M 10.7MB/s in 84s
2023-09-04 11:46:33 (10.7 MB/s) - ‘nafnet.zip’ saved [943903968/943903968]
Archive: nafnet.zip
creating: NAFNet/
creating: NAFNet/pretrain_model/
creating: NAFNet/pretrain_model/cv_nafnet_image-deblur_reds/
inflating: NAFNet/pretrain_model/cv_nafnet_image-deblur_reds/configuration.json
inflating: NAFNet/pretrain_model/cv_nafnet_image-deblur_reds/README.md
creating: NAFNet/pretrain_model/cv_nafnet_image-deblur_reds/data/
inflating: NAFNet/pretrain_model/cv_nafnet_image-deblur_reds/data/deblur.gif
inflating: NAFNet/pretrain_model/cv_nafnet_image-deblur_reds/data/nafnet_arch.png
inflating: NAFNet/pretrain_model/cv_nafnet_image-deblur_reds/data/blurry.jpg
extracting: NAFNet/pretrain_model/cv_nafnet_image-deblur_reds/.mdl
inflating: NAFNet/pretrain_model/cv_nafnet_image-deblur_reds/.msc
inflating: NAFNet/pretrain_model/cv_nafnet_image-deblur_reds/pytorch_model.pt
creating: NAFNet/pretrain_model/cv_nafnet_image-denoise_sidd/
inflating: NAFNet/pretrain_model/cv_nafnet_image-denoise_sidd/configuration.json
inflating: NAFNet/pretrain_model/cv_nafnet_image-denoise_sidd/README.md
creating: NAFNet/pretrain_model/cv_nafnet_image-denoise_sidd/data/
creating: NAFNet/pretrain_model/cv_nafnet_image-denoise_sidd/data/SIDD_example/
creating: NAFNet/pretrain_model/cv_nafnet_image-denoise_sidd/data/SIDD_example/0003_001_S6_00100_00060_3200_H/
inflating: NAFNet/pretrain_model/cv_nafnet_image-denoise_sidd/data/SIDD_example/0003_001_S6_00100_00060_3200_H/0003_GT_SRGB_010.PNG
inflating: NAFNet/pretrain_model/cv_nafnet_image-denoise_sidd/data/SIDD_example/0003_001_S6_00100_00060_3200_H/0003_NOISY_SRGB_010.PNG
inflating: NAFNet/pretrain_model/cv_nafnet_image-denoise_sidd/data/SIDD_example/0003_001_S6_00100_00060_3200_H/0003_GT_SRGB_011.PNG
inflating: NAFNet/pretrain_model/cv_nafnet_image-denoise_sidd/data/SIDD_example/0003_001_S6_00100_00060_3200_H/0003_NOISY_SRGB_011.PNG
creating: NAFNet/pretrain_model/cv_nafnet_image-denoise_sidd/data/SIDD_example/0001_001_S6_00100_00060_3200_L/
inflating: NAFNet/pretrain_model/cv_nafnet_image-denoise_sidd/data/SIDD_example/0001_001_S6_00100_00060_3200_L/0001_NOISY_SRGB_010.PNG
inflating: NAFNet/pretrain_model/cv_nafnet_image-denoise_sidd/data/SIDD_example/0001_001_S6_00100_00060_3200_L/0001_GT_SRGB_010.PNG
inflating: NAFNet/pretrain_model/cv_nafnet_image-denoise_sidd/data/SIDD_example/0001_001_S6_00100_00060_3200_L/0001_GT_SRGB_011.PNG
inflating: NAFNet/pretrain_model/cv_nafnet_image-denoise_sidd/data/SIDD_example/0001_001_S6_00100_00060_3200_L/0001_NOISY_SRGB_011.PNG
creating: NAFNet/pretrain_model/cv_nafnet_image-denoise_sidd/data/SIDD_example/0002_001_S6_00100_00020_3200_N/
inflating: NAFNet/pretrain_model/cv_nafnet_image-denoise_sidd/data/SIDD_example/0002_001_S6_00100_00020_3200_N/0002_GT_SRGB_011.PNG
inflating: NAFNet/pretrain_model/cv_nafnet_image-denoise_sidd/data/SIDD_example/0002_001_S6_00100_00020_3200_N/0002_NOISY_SRGB_010.PNG
inflating: NAFNet/pretrain_model/cv_nafnet_image-denoise_sidd/data/SIDD_example/0002_001_S6_00100_00020_3200_N/0002_NOISY_SRGB_011.PNG
inflating: NAFNet/pretrain_model/cv_nafnet_image-denoise_sidd/data/SIDD_example/0002_001_S6_00100_00020_3200_N/0002_GT_SRGB_010.PNG
creating: NAFNet/pretrain_model/cv_nafnet_image-denoise_sidd/data/SIDD_example/0004_001_S6_00100_00060_4400_L/
inflating: NAFNet/pretrain_model/cv_nafnet_image-denoise_sidd/data/SIDD_example/0004_001_S6_00100_00060_4400_L/0004_NOISY_SRGB_010.PNG
inflating: NAFNet/pretrain_model/cv_nafnet_image-denoise_sidd/data/SIDD_example/0004_001_S6_00100_00060_4400_L/0004_GT_SRGB_010.PNG
inflating: NAFNet/pretrain_model/cv_nafnet_image-denoise_sidd/data/SIDD_example/0004_001_S6_00100_00060_4400_L/0004_GT_SRGB_011.PNG
inflating: NAFNet/pretrain_model/cv_nafnet_image-denoise_sidd/data/SIDD_example/0004_001_S6_00100_00060_4400_L/0004_NOISY_SRGB_011.PNG
inflating: NAFNet/pretrain_model/cv_nafnet_image-denoise_sidd/data/noisy-demo-1.png
inflating: NAFNet/pretrain_model/cv_nafnet_image-denoise_sidd/data/nafnet_arch.png
inflating: NAFNet/pretrain_model/cv_nafnet_image-denoise_sidd/data/noisy-demo-0.png
inflating: NAFNet/pretrain_model/cv_nafnet_image-denoise_sidd/data/denoise.gif
extracting: NAFNet/pretrain_model/cv_nafnet_image-denoise_sidd/.mdl
inflating: NAFNet/pretrain_model/cv_nafnet_image-denoise_sidd/.msc
inflating: NAFNet/pretrain_model/cv_nafnet_image-denoise_sidd/pytorch_model.pt
creating: NAFNet/pretrain_model/cv_nafnet_image-deblur_gopro/
inflating: NAFNet/pretrain_model/cv_nafnet_image-deblur_gopro/configuration.json
inflating: NAFNet/pretrain_model/cv_nafnet_image-deblur_gopro/README.md
creating: NAFNet/pretrain_model/cv_nafnet_image-deblur_gopro/data/
inflating: NAFNet/pretrain_model/cv_nafnet_image-deblur_gopro/data/deblur.gif
inflating: NAFNet/pretrain_model/cv_nafnet_image-deblur_gopro/data/nafnet_arch.png
inflating: NAFNet/pretrain_model/cv_nafnet_image-deblur_gopro/data/blurry.jpg
extracting: NAFNet/pretrain_model/cv_nafnet_image-deblur_gopro/.mdl
inflating: NAFNet/pretrain_model/cv_nafnet_image-deblur_gopro/.msc
inflating: NAFNet/pretrain_model/cv_nafnet_image-deblur_gopro/pytorch_model.pt
inflating: NAFNet/demo.py 2023-09-04 11:47:14,618 - modelscope - INFO - PyTorch version 1.13.1+cu117 Found.
2023-09-04 11:47:14,619 - modelscope - INFO - Loading ast index from /root/.cache/modelscope/ast_indexer
2023-09-04 11:47:14,619 - modelscope - INFO - No valid ast index found from /root/.cache/modelscope/ast_indexer, generating ast index from prebuilt!
2023-09-04 11:47:14,695 - modelscope - INFO - Loading done! Current index file version is 1.8.4, with md5 80fa9349fc3e7b04fcfad511918062b1 and a total number of 902 components indexed
/mnt/workspace/NAFNet
2023-09-04 11:47:15,604 - modelscope - INFO - initiate model from /mnt/workspace/NAFNet/pretrain_model/cv_nafnet_image-deblur_gopro
2023-09-04 11:47:15,604 - modelscope - INFO - initiate model from location /mnt/workspace/NAFNet/pretrain_model/cv_nafnet_image-deblur_gopro.
2023-09-04 11:47:15,605 - modelscope - INFO - initialize model from /mnt/workspace/NAFNet/pretrain_model/cv_nafnet_image-deblur_gopro
2023-09-04 11:47:16,267 - modelscope - INFO - Loading NAFNet model from /mnt/workspace/NAFNet/pretrain_model/cv_nafnet_image-deblur_gopro/pytorch_model.pt, with param key: [params].
2023-09-04 11:47:16,368 - modelscope - INFO - load model done.
2023-09-04 11:47:16,400 - modelscope - INFO - load image denoise model done
Total Image: 10
2023-09-04 11:47:17,517 - modelscope - WARNING - task image-deblurring input definition is missing
2023-09-04 11:47:18,893 - modelscope - WARNING - task image-deblurring output keys are missing
0/10 saved at results/nafnet_deblur/64.jpg
1/10 saved at results/nafnet_deblur/2.jpg
2/10 saved at results/nafnet_deblur/34.jpg
3/10 saved at results/nafnet_deblur/54.jpg
4/10 saved at results/nafnet_deblur/40.jpg
5/10 saved at results/nafnet_deblur/10.jpg
6/10 saved at results/nafnet_deblur/70.jpg
7/10 saved at results/nafnet_deblur/4.png
8/10 saved at results/nafnet_deblur/20.jpg
9/10 saved at results/nafnet_deblur/50.jpg
2023-09-04 11:47:23,689 - modelscope - INFO - PyTorch version 1.13.1+cu117 Found.
2023-09-04 11:47:23,690 - modelscope - INFO - Loading ast index from /root/.cache/modelscope/ast_indexer
2023-09-04 11:47:23,755 - modelscope - INFO - Loading done! Current index file version is 1.8.4, with md5 80fa9349fc3e7b04fcfad511918062b1 and a total number of 902 components indexed
/mnt/workspace/NAFNet
2023-09-04 11:47:24,316 - modelscope - INFO - initiate model from /mnt/workspace/NAFNet/pretrain_model/cv_nafnet_image-denoise_sidd
2023-09-04 11:47:24,316 - modelscope - INFO - initiate model from location /mnt/workspace/NAFNet/pretrain_model/cv_nafnet_image-denoise_sidd.
2023-09-04 11:47:24,317 - modelscope - INFO - initialize model from /mnt/workspace/NAFNet/pretrain_model/cv_nafnet_image-denoise_sidd
2023-09-04 11:47:24,627 - modelscope - INFO - Loading NAFNet model from /mnt/workspace/NAFNet/pretrain_model/cv_nafnet_image-denoise_sidd/pytorch_model.pt, with param key: [params].
2023-09-04 11:47:24,685 - modelscope - INFO - load model done.
2023-09-04 11:47:24,707 - modelscope - INFO - load image denoise model done
Total Image: 10
0/10 saved at results/nafnet_denoise/64.jpg
1/10 saved at results/nafnet_denoise/2.jpg
2/10 saved at results/nafnet_denoise/34.jpg
3/10 saved at results/nafnet_denoise/54.jpg
4/10 saved at results/nafnet_denoise/40.jpg
5/10 saved at results/nafnet_denoise/10.jpg
6/10 saved at results/nafnet_denoise/70.jpg
7/10 saved at results/nafnet_denoise/4.png
8/10 saved at results/nafnet_denoise/20.jpg
9/10 saved at results/nafnet_denoise/50.jpg
2023-09-04 11:47:30,906 - modelscope - INFO - PyTorch version 1.13.1+cu117 Found.
2023-09-04 11:47:30,907 - modelscope - INFO - Loading ast index from /root/.cache/modelscope/ast_indexer
2023-09-04 11:47:30,969 - modelscope - INFO - Loading done! Current index file version is 1.8.4, with md5 80fa9349fc3e7b04fcfad511918062b1 and a total number of 902 components indexed
/mnt/workspace/NAFNet
2023-09-04 11:47:31,522 - modelscope - INFO - initiate model from /mnt/workspace/NAFNet/pretrain_model/cv_nafnet_image-deblur_reds
2023-09-04 11:47:31,522 - modelscope - INFO - initiate model from location /mnt/workspace/NAFNet/pretrain_model/cv_nafnet_image-deblur_reds.
2023-09-04 11:47:31,523 - modelscope - INFO - initialize model from /mnt/workspace/NAFNet/pretrain_model/cv_nafnet_image-deblur_reds
2023-09-04 11:47:32,187 - modelscope - INFO - Loading NAFNet model from /mnt/workspace/NAFNet/pretrain_model/cv_nafnet_image-deblur_reds/pytorch_model.pt, with param key: [params].
2023-09-04 11:47:32,296 - modelscope - INFO - load model done.
2023-09-04 11:47:32,320 - modelscope - INFO - load image denoise model done
Total Image: 10
2023-09-04 11:47:33,402 - modelscope - WARNING - task image-deblurring input definition is missing
2023-09-04 11:47:34,753 - modelscope - WARNING - task image-deblurring output keys are missing
0/10 saved at results/nafnet_de_motion_blur/64.jpg
1/10 saved at results/nafnet_de_motion_blur/2.jpg
2/10 saved at results/nafnet_de_motion_blur/34.jpg
3/10 saved at results/nafnet_de_motion_blur/54.jpg
4/10 saved at results/nafnet_de_motion_blur/40.jpg
5/10 saved at results/nafnet_de_motion_blur/10.jpg
6/10 saved at results/nafnet_de_motion_blur/70.jpg
7/10 saved at results/nafnet_de_motion_blur/4.png
8/10 saved at results/nafnet_de_motion_blur/20.jpg
9/10 saved at results/nafnet_de_motion_blur/50.jpghttp://pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com/aigc-data/restoration/repo/realesrgan.zip
cn-hangzhou
--2023-09-05 01:29:05-- http://pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com/aigc-data/restoration/repo/realesrgan.zip
Resolving pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com (pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com)... 100.118.28.50, 100.118.28.49, 100.118.28.44, ...
Connecting to pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com (pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com)|100.118.28.50|:80... connected.
HTTP request sent, awaiting response... 200 OK
Length: 151719684 (145M) [application/zip]
Saving to: ‘realesrgan.zip’
realesrgan.zip 100%[===================>] 144.69M 11.3MB/s in 13s
2023-09-05 01:29:17 (11.5 MB/s) - ‘realesrgan.zip’ saved [151719684/151719684]
Archive: realesrgan.zip
creating: Real-ESRGAN/
creating: Real-ESRGAN/.ipynb_checkpoints/
inflating: Real-ESRGAN/.ipynb_checkpoints/inference_realesrgan-checkpoint.py
inflating: Real-ESRGAN/setup.cfg
creating: Real-ESRGAN/tests/
inflating: Real-ESRGAN/tests/test_discriminator_arch.py
inflating: Real-ESRGAN/tests/test_model.py
inflating: Real-ESRGAN/tests/test_utils.py
creating: Real-ESRGAN/tests/data/
inflating: Real-ESRGAN/tests/data/meta_info_pair.txt
inflating: Real-ESRGAN/tests/data/test_realesrnet_model.yml
inflating: Real-ESRGAN/tests/data/test_realesrgan_dataset.yml
creating: Real-ESRGAN/tests/data/gt.lmdb/
inflating: Real-ESRGAN/tests/data/gt.lmdb/meta_info.txt
inflating: Real-ESRGAN/tests/data/gt.lmdb/lock.mdb
inflating: Real-ESRGAN/tests/data/gt.lmdb/data.mdb
inflating: Real-ESRGAN/tests/data/test_realesrgan_model.yml
inflating: Real-ESRGAN/tests/data/meta_info_gt.txt
inflating: Real-ESRGAN/tests/data/test_realesrgan_paired_dataset.yml
creating: Real-ESRGAN/tests/data/gt/
inflating: Real-ESRGAN/tests/data/gt/comic.png
inflating: Real-ESRGAN/tests/data/gt/baboon.png
creating: Real-ESRGAN/tests/data/lq.lmdb/
inflating: Real-ESRGAN/tests/data/lq.lmdb/meta_info.txt
inflating: Real-ESRGAN/tests/data/lq.lmdb/lock.mdb
inflating: Real-ESRGAN/tests/data/lq.lmdb/data.mdb
creating: Real-ESRGAN/tests/data/lq/
extracting: Real-ESRGAN/tests/data/lq/comic.png
inflating: Real-ESRGAN/tests/data/lq/baboon.png
inflating: Real-ESRGAN/tests/test_dataset.py
inflating: Real-ESRGAN/requirements.txt
creating: Real-ESRGAN/assets/
inflating: Real-ESRGAN/assets/teaser.jpg
inflating: Real-ESRGAN/assets/realesrgan_logo_ai.png
inflating: Real-ESRGAN/assets/realesrgan_logo_gv.png
inflating: Real-ESRGAN/assets/realesrgan_logo.png
inflating: Real-ESRGAN/assets/realesrgan_logo_gi.png
inflating: Real-ESRGAN/assets/teaser-text.png
inflating: Real-ESRGAN/assets/realesrgan_logo_av.png
creating: Real-ESRGAN/pretrained_model/
inflating: Real-ESRGAN/.gitignore
creating: Real-ESRGAN/.vscode/
inflating: Real-ESRGAN/.vscode/settings.json
inflating: Real-ESRGAN/demo.py
creating: Real-ESRGAN/experiments/
creating: Real-ESRGAN/experiments/pretrained_models/
extracting: Real-ESRGAN/experiments/pretrained_models/README.md
inflating: Real-ESRGAN/cog.yaml
inflating: Real-ESRGAN/LICENSE
creating: Real-ESRGAN/inputs/
inflating: Real-ESRGAN/inputs/wolf_gray.jpg
inflating: Real-ESRGAN/inputs/0030.jpg
creating: Real-ESRGAN/inputs/video/
inflating: Real-ESRGAN/inputs/video/onepiece_demo.mp4
inflating: Real-ESRGAN/inputs/0014.jpg
inflating: Real-ESRGAN/inputs/OST_009.png
inflating: Real-ESRGAN/inputs/tree_alpha_16bit.png
inflating: Real-ESRGAN/inputs/children-alpha.png
inflating: Real-ESRGAN/inputs/00003.png
inflating: Real-ESRGAN/inputs/00017_gray.png
inflating: Real-ESRGAN/inputs/ADE_val_00000114.jpg
inflating: Real-ESRGAN/MANIFEST.in
inflating: Real-ESRGAN/.pre-commit-config.yaml
creating: Real-ESRGAN/.git/
creating: Real-ESRGAN/.git/logs/
creating: Real-ESRGAN/.git/logs/refs/
creating: Real-ESRGAN/.git/logs/refs/remotes/
creating: Real-ESRGAN/.git/logs/refs/remotes/origin/
inflating: Real-ESRGAN/.git/logs/refs/remotes/origin/HEAD
creating: Real-ESRGAN/.git/logs/refs/heads/
inflating: Real-ESRGAN/.git/logs/refs/heads/master
inflating: Real-ESRGAN/.git/logs/HEAD
inflating: Real-ESRGAN/.git/config
creating: Real-ESRGAN/.git/refs/
creating: Real-ESRGAN/.git/refs/remotes/
creating: Real-ESRGAN/.git/refs/remotes/origin/
extracting: Real-ESRGAN/.git/refs/remotes/origin/HEAD
creating: Real-ESRGAN/.git/refs/heads/
extracting: Real-ESRGAN/.git/refs/heads/master
creating: Real-ESRGAN/.git/refs/tags/
extracting: Real-ESRGAN/.git/HEAD
creating: Real-ESRGAN/.git/hooks/
inflating: Real-ESRGAN/.git/hooks/push-to-checkout.sample
inflating: Real-ESRGAN/.git/hooks/commit-msg.sample
inflating: Real-ESRGAN/.git/hooks/applypatch-msg.sample
inflating: Real-ESRGAN/.git/hooks/pre-receive.sample
inflating: Real-ESRGAN/.git/hooks/pre-push.sample
inflating: Real-ESRGAN/.git/hooks/fsmonitor-watchman.sample
inflating: Real-ESRGAN/.git/hooks/post-update.sample
inflating: Real-ESRGAN/.git/hooks/update.sample
inflating: Real-ESRGAN/.git/hooks/pre-merge-commit.sample
inflating: Real-ESRGAN/.git/hooks/pre-commit.sample
inflating: Real-ESRGAN/.git/hooks/pre-applypatch.sample
inflating: Real-ESRGAN/.git/hooks/prepare-commit-msg.sample
inflating: Real-ESRGAN/.git/hooks/pre-rebase.sample
creating: Real-ESRGAN/.git/info/
inflating: Real-ESRGAN/.git/info/exclude
creating: Real-ESRGAN/.git/objects/
creating: Real-ESRGAN/.git/objects/pack/
inflating: Real-ESRGAN/.git/objects/pack/pack-4d4a54ddcab2e146413d01c2262bbf138200efd1.pack
inflating: Real-ESRGAN/.git/objects/pack/pack-4d4a54ddcab2e146413d01c2262bbf138200efd1.idx
creating: Real-ESRGAN/.git/objects/info/
inflating: Real-ESRGAN/.git/packed-refs
inflating: Real-ESRGAN/.git/index
inflating: Real-ESRGAN/.git/description
inflating: Real-ESRGAN/inference_realesrgan_video.py
inflating: Real-ESRGAN/CODE_OF_CONDUCT.md
inflating: Real-ESRGAN/README.md
creating: Real-ESRGAN/realesrgan/
creating: Real-ESRGAN/realesrgan/.ipynb_checkpoints/
inflating: Real-ESRGAN/realesrgan/.ipynb_checkpoints/__init__-checkpoint.py
creating: Real-ESRGAN/realesrgan/models/
inflating: Real-ESRGAN/realesrgan/models/realesrnet_model.py
creating: Real-ESRGAN/realesrgan/models/__pycache__/
inflating: Real-ESRGAN/realesrgan/models/__pycache__/__init__.cpython-310.pyc
inflating: Real-ESRGAN/realesrgan/models/__pycache__/realesrgan_model.cpython-310.pyc
inflating: Real-ESRGAN/realesrgan/models/__pycache__/realesrnet_model.cpython-310.pyc
inflating: Real-ESRGAN/realesrgan/models/realesrgan_model.py
inflating: Real-ESRGAN/realesrgan/models/__init__.py
inflating: Real-ESRGAN/realesrgan/train.py
creating: Real-ESRGAN/realesrgan/__pycache__/
inflating: Real-ESRGAN/realesrgan/__pycache__/__init__.cpython-310.pyc
inflating: Real-ESRGAN/realesrgan/__pycache__/utils.cpython-310.pyc
creating: Real-ESRGAN/realesrgan/archs/
inflating: Real-ESRGAN/realesrgan/archs/discriminator_arch.py
creating: Real-ESRGAN/realesrgan/archs/__pycache__/
inflating: Real-ESRGAN/realesrgan/archs/__pycache__/__init__.cpython-310.pyc
inflating: Real-ESRGAN/realesrgan/archs/__pycache__/srvgg_arch.cpython-310.pyc
inflating: Real-ESRGAN/realesrgan/archs/__pycache__/discriminator_arch.cpython-310.pyc
inflating: Real-ESRGAN/realesrgan/archs/__init__.py
inflating: Real-ESRGAN/realesrgan/archs/srvgg_arch.py
creating: Real-ESRGAN/realesrgan/data/
creating: Real-ESRGAN/realesrgan/data/__pycache__/
inflating: Real-ESRGAN/realesrgan/data/__pycache__/realesrgan_paired_dataset.cpython-310.pyc
inflating: Real-ESRGAN/realesrgan/data/__pycache__/__init__.cpython-310.pyc
inflating: Real-ESRGAN/realesrgan/data/__pycache__/realesrgan_dataset.cpython-310.pyc
inflating: Real-ESRGAN/realesrgan/data/realesrgan_dataset.py
inflating: Real-ESRGAN/realesrgan/data/realesrgan_paired_dataset.py
inflating: Real-ESRGAN/realesrgan/data/__init__.py
inflating: Real-ESRGAN/realesrgan/__init__.py
inflating: Real-ESRGAN/realesrgan/utils.py
inflating: Real-ESRGAN/README_CN.md
extracting: Real-ESRGAN/VERSION
inflating: Real-ESRGAN/setup.py
creating: Real-ESRGAN/.github/
creating: Real-ESRGAN/.github/workflows/
inflating: Real-ESRGAN/.github/workflows/no-response.yml
inflating: Real-ESRGAN/.github/workflows/release.yml
inflating: Real-ESRGAN/.github/workflows/pylint.yml
inflating: Real-ESRGAN/.github/workflows/publish-pip.yml
creating: Real-ESRGAN/docs/
inflating: Real-ESRGAN/docs/anime_comparisons.md
inflating: Real-ESRGAN/docs/FAQ.md
inflating: Real-ESRGAN/docs/ncnn_conversion.md
inflating: Real-ESRGAN/docs/Training_CN.md
inflating: Real-ESRGAN/docs/CONTRIBUTING.md
inflating: Real-ESRGAN/docs/anime_video_model.md
inflating: Real-ESRGAN/docs/feedback.md
inflating: Real-ESRGAN/docs/anime_model.md
inflating: Real-ESRGAN/docs/Training.md
inflating: Real-ESRGAN/docs/model_zoo.md
inflating: Real-ESRGAN/docs/anime_comparisons_CN.md
inflating: Real-ESRGAN/cog_predict.py
creating: Real-ESRGAN/scripts/
inflating: Real-ESRGAN/scripts/pytorch2onnx.py
inflating: Real-ESRGAN/scripts/extract_subimages.py
inflating: Real-ESRGAN/scripts/generate_meta_info.py
inflating: Real-ESRGAN/scripts/generate_multiscale_DF2K.py
inflating: Real-ESRGAN/scripts/generate_meta_info_pairdata.py
creating: Real-ESRGAN/weights/
inflating: Real-ESRGAN/weights/RealESRNet_x4plus.pth
inflating: Real-ESRGAN/weights/README.md
inflating: Real-ESRGAN/weights/RealESRGAN_x4plus.pth
inflating: Real-ESRGAN/weights/RealESRGAN_x4plus_anime_6B.pth
creating: Real-ESRGAN/options/
inflating: Real-ESRGAN/options/train_realesrnet_x2plus.yml
inflating: Real-ESRGAN/options/finetune_realesrgan_x4plus_pairdata.yml
inflating: Real-ESRGAN/options/train_realesrnet_x4plus.yml
inflating: Real-ESRGAN/options/train_realesrgan_x2plus.yml
inflating: Real-ESRGAN/options/train_realesrgan_x4plus.yml
inflating: Real-ESRGAN/options/finetune_realesrgan_x4plus.yml Testing 0 10
Tile 1/1
Testing 1 2
Tile 1/4
Tile 2/4
Tile 3/4
Tile 4/4
Testing 2 20
Tile 1/4
Tile 2/4
Tile 3/4
Tile 4/4
Testing 3 34
Tile 1/4
Tile 2/4
Tile 3/4
Tile 4/4
Testing 4 4
Tile 1/4
Tile 2/4
Tile 3/4
Tile 4/4
Tile 1/4
Tile 2/4
Tile 3/4
Tile 4/4
Testing 5 40
Tile 1/1
Testing 6 50
Tile 1/1
Testing 7 54
Tile 1/4
Tile 2/4
Tile 3/4
Tile 4/4
Testing 8 64
Tile 1/1
Testing 9 70
Tile 1/4
Tile 2/4
Tile 3/4
Tile 4/4
Testing 0 10
Tile 1/1
Testing 1 2
Tile 1/4
Tile 2/4
Tile 3/4
Tile 4/4
Testing 2 20
Tile 1/4
Tile 2/4
Tile 3/4
Tile 4/4
Testing 3 34
Tile 1/4
Tile 2/4
Tile 3/4
Tile 4/4
Testing 4 4
Tile 1/4
Tile 2/4
Tile 3/4
Tile 4/4
Tile 1/4
Tile 2/4
Tile 3/4
Tile 4/4
Testing 5 40
Tile 1/1
Testing 6 50
Tile 1/1
Testing 7 54
Tile 1/4
Tile 2/4
Tile 3/4
Tile 4/4
Testing 8 64
Tile 1/1
Testing 9 70
Tile 1/4
Tile 2/4
Tile 3/4
Tile 4/4
Testing 0 10
Tile 1/1
Testing 1 2
Tile 1/4
Tile 2/4
Tile 3/4
Tile 4/4
Testing 2 20
Tile 1/4
Tile 2/4
Tile 3/4
Tile 4/4
Testing 3 34
Tile 1/4
Tile 2/4
Tile 3/4
Tile 4/4
Testing 4 4
Tile 1/4
Tile 2/4
Tile 3/4
Tile 4/4
Tile 1/4
Tile 2/4
Tile 3/4
Tile 4/4
Testing 5 40
Tile 1/1
Testing 6 50
Tile 1/1
Testing 7 54
Tile 1/4
Tile 2/4
Tile 3/4
Tile 4/4
Testing 8 64
Tile 1/1
Testing 9 70
Tile 1/4
Tile 2/4
Tile 3/4
Tile 4/4http://pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com/aigc-data/restoration/repo/swinir.zip
cn-hangzhou
--2023-09-05 01:34:02-- http://pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com/aigc-data/restoration/repo/swinir.zip
Resolving pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com (pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com)... 100.118.28.44, 100.118.28.49, 100.118.28.45, ...
Connecting to pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com (pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com)|100.118.28.44|:80... connected.
HTTP request sent, awaiting response... 200 OK
Length: 580693761 (554M) [application/zip]
Saving to: ‘swinir.zip’
swinir.zip 100%[===================>] 553.79M 11.4MB/s in 51s
2023-09-05 01:34:54 (10.8 MB/s) - ‘swinir.zip’ saved [580693761/580693761]
Archive: swinir.zip
creating: SwinIR/
creating: SwinIR/.ipynb_checkpoints/
inflating: SwinIR/.ipynb_checkpoints/main_test_swinir-checkpoint.py
inflating: SwinIR/.ipynb_checkpoints/predict-checkpoint.py
creating: SwinIR/models/
creating: SwinIR/models/__pycache__/
inflating: SwinIR/models/__pycache__/network_swinir.cpython-310.pyc
inflating: SwinIR/models/network_swinir.py
creating: SwinIR/pretrained_model/
inflating: SwinIR/pretrained_model/004_grayDN_DFWB_s128w8_SwinIR-M_noise50.pth
inflating: SwinIR/pretrained_model/006_colorCAR_DFWB_s126w7_SwinIR-M_jpeg10.pth
inflating: SwinIR/pretrained_model/004_grayDN_DFWB_s128w8_SwinIR-M_noise25.pth
inflating: SwinIR/pretrained_model/005_colorDN_DFWB_s128w8_SwinIR-M_noise50.pth
inflating: SwinIR/pretrained_model/001_classicalSR_DF2K_s64w8_SwinIR-M_x4.pth
inflating: SwinIR/pretrained_model/006_colorCAR_DFWB_s126w7_SwinIR-M_jpeg40.pth
inflating: SwinIR/pretrained_model/006_colorCAR_DFWB_s126w7_SwinIR-M_jpeg20.pth
inflating: SwinIR/pretrained_model/006_colorCAR_DFWB_s126w7_SwinIR-M_jpeg30.pth
inflating: SwinIR/pretrained_model/005_colorDN_DFWB_s128w8_SwinIR-M_noise15.pth
inflating: SwinIR/pretrained_model/004_grayDN_DFWB_s128w8_SwinIR-M_noise15.pth
inflating: SwinIR/pretrained_model/003_realSR_BSRGAN_DFO_s64w8_SwinIR-M_x4_GAN.pth
inflating: SwinIR/pretrained_model/005_colorDN_DFWB_s128w8_SwinIR-M_noise25.pth
inflating: SwinIR/demo.py
inflating: SwinIR/cog.yaml
inflating: SwinIR/LICENSE
inflating: SwinIR/download-weights.sh
creating: SwinIR/utils/
creating: SwinIR/utils/__pycache__/
inflating: SwinIR/utils/__pycache__/util_calculate_psnr_ssim.cpython-310.pyc
inflating: SwinIR/utils/util_calculate_psnr_ssim.py
creating: SwinIR/.git/
creating: SwinIR/.git/logs/
creating: SwinIR/.git/logs/refs/
creating: SwinIR/.git/logs/refs/remotes/
creating: SwinIR/.git/logs/refs/remotes/origin/
inflating: SwinIR/.git/logs/refs/remotes/origin/HEAD
creating: SwinIR/.git/logs/refs/heads/
inflating: SwinIR/.git/logs/refs/heads/main
inflating: SwinIR/.git/logs/HEAD
inflating: SwinIR/.git/config
creating: SwinIR/.git/refs/
creating: SwinIR/.git/refs/remotes/
creating: SwinIR/.git/refs/remotes/origin/
extracting: SwinIR/.git/refs/remotes/origin/HEAD
creating: SwinIR/.git/refs/heads/
extracting: SwinIR/.git/refs/heads/main
creating: SwinIR/.git/refs/tags/
extracting: SwinIR/.git/HEAD
creating: SwinIR/.git/hooks/
inflating: SwinIR/.git/hooks/push-to-checkout.sample
inflating: SwinIR/.git/hooks/commit-msg.sample
inflating: SwinIR/.git/hooks/applypatch-msg.sample
inflating: SwinIR/.git/hooks/pre-receive.sample
inflating: SwinIR/.git/hooks/pre-push.sample
inflating: SwinIR/.git/hooks/fsmonitor-watchman.sample
inflating: SwinIR/.git/hooks/post-update.sample
inflating: SwinIR/.git/hooks/update.sample
inflating: SwinIR/.git/hooks/pre-merge-commit.sample
inflating: SwinIR/.git/hooks/pre-commit.sample
inflating: SwinIR/.git/hooks/pre-applypatch.sample
inflating: SwinIR/.git/hooks/prepare-commit-msg.sample
inflating: SwinIR/.git/hooks/pre-rebase.sample
creating: SwinIR/.git/info/
inflating: SwinIR/.git/info/exclude
creating: SwinIR/.git/objects/
creating: SwinIR/.git/objects/pack/
inflating: SwinIR/.git/objects/pack/pack-04d185ace674ba8b24be5a9b5fb3304d4c6a4e74.pack
inflating: SwinIR/.git/objects/pack/pack-04d185ace674ba8b24be5a9b5fb3304d4c6a4e74.idx
creating: SwinIR/.git/objects/info/
inflating: SwinIR/.git/packed-refs
inflating: SwinIR/.git/index
inflating: SwinIR/.git/description
inflating: SwinIR/README.md
creating: SwinIR/testsets/
creating: SwinIR/testsets/McMaster/
inflating: SwinIR/testsets/McMaster/17.tif
inflating: SwinIR/testsets/McMaster/9.tif
inflating: SwinIR/testsets/McMaster/3.tif
inflating: SwinIR/testsets/McMaster/5.tif
inflating: SwinIR/testsets/McMaster/16.tif
inflating: SwinIR/testsets/McMaster/4.tif
inflating: SwinIR/testsets/McMaster/12.tif
inflating: SwinIR/testsets/McMaster/6.tif
inflating: SwinIR/testsets/McMaster/14.tif
inflating: SwinIR/testsets/McMaster/7.tif
inflating: SwinIR/testsets/McMaster/8.tif
inflating: SwinIR/testsets/McMaster/13.tif
inflating: SwinIR/testsets/McMaster/1.tif
inflating: SwinIR/testsets/McMaster/18.tif
inflating: SwinIR/testsets/McMaster/11.tif
inflating: SwinIR/testsets/McMaster/2.tif
inflating: SwinIR/testsets/McMaster/15.tif
inflating: SwinIR/testsets/McMaster/10.tif
creating: SwinIR/testsets/classic5/
inflating: SwinIR/testsets/classic5/lena.bmp
inflating: SwinIR/testsets/classic5/baboon.bmp
inflating: SwinIR/testsets/classic5/barbara.bmp
inflating: SwinIR/testsets/classic5/boats.bmp
inflating: SwinIR/testsets/classic5/peppers.bmp
creating: SwinIR/testsets/Set12/
inflating: SwinIR/testsets/Set12/06.png
inflating: SwinIR/testsets/Set12/04.png
inflating: SwinIR/testsets/Set12/03.png
inflating: SwinIR/testsets/Set12/02.png
inflating: SwinIR/testsets/Set12/11.png
inflating: SwinIR/testsets/Set12/07.png
inflating: SwinIR/testsets/Set12/09.png
inflating: SwinIR/testsets/Set12/10.png
inflating: SwinIR/testsets/Set12/12.png
inflating: SwinIR/testsets/Set12/05.png
inflating: SwinIR/testsets/Set12/08.png
inflating: SwinIR/testsets/Set12/01.png
creating: SwinIR/testsets/RealSRSet+5images/
inflating: SwinIR/testsets/RealSRSet+5images/building.png
inflating: SwinIR/testsets/RealSRSet+5images/0030.jpg
inflating: SwinIR/testsets/RealSRSet+5images/painting.png
inflating: SwinIR/testsets/RealSRSet+5images/comic1.png
inflating: SwinIR/testsets/RealSRSet+5images/0014.jpg
inflating: SwinIR/testsets/RealSRSet+5images/OST_009.png
inflating: SwinIR/testsets/RealSRSet+5images/comic2.png
inflating: SwinIR/testsets/RealSRSet+5images/pattern.png
inflating: SwinIR/testsets/RealSRSet+5images/foreman.png
inflating: SwinIR/testsets/RealSRSet+5images/oldphoto3.png
inflating: SwinIR/testsets/RealSRSet+5images/oldphoto6.png
inflating: SwinIR/testsets/RealSRSet+5images/dog.png
inflating: SwinIR/testsets/RealSRSet+5images/Lincoln.png
inflating: SwinIR/testsets/RealSRSet+5images/oldphoto2.png
inflating: SwinIR/testsets/RealSRSet+5images/00003.png
inflating: SwinIR/testsets/RealSRSet+5images/frog.png
inflating: SwinIR/testsets/RealSRSet+5images/butterfly.png
inflating: SwinIR/testsets/RealSRSet+5images/ppt3.png
inflating: SwinIR/testsets/RealSRSet+5images/butterfly2.png
inflating: SwinIR/testsets/RealSRSet+5images/computer.png
extracting: SwinIR/testsets/RealSRSet+5images/chip.png
inflating: SwinIR/testsets/RealSRSet+5images/comic3.png
inflating: SwinIR/testsets/RealSRSet+5images/dped_crop00061.png
inflating: SwinIR/testsets/RealSRSet+5images/tiger.png
inflating: SwinIR/testsets/RealSRSet+5images/ADE_val_00000114.jpg
creating: SwinIR/testsets/Set5/
creating: SwinIR/testsets/Set5/HR/
inflating: SwinIR/testsets/Set5/HR/woman.png
inflating: SwinIR/testsets/Set5/HR/butterfly.png
inflating: SwinIR/testsets/Set5/HR/head.png
inflating: SwinIR/testsets/Set5/HR/baby.png
inflating: SwinIR/testsets/Set5/HR/bird.png
creating: SwinIR/testsets/Set5/LR_bicubic/
creating: SwinIR/testsets/Set5/LR_bicubic/X8/
extracting: SwinIR/testsets/Set5/LR_bicubic/X8/womanx8.png
extracting: SwinIR/testsets/Set5/LR_bicubic/X8/headx8.png
extracting: SwinIR/testsets/Set5/LR_bicubic/X8/birdx8.png
extracting: SwinIR/testsets/Set5/LR_bicubic/X8/babyx8.png
extracting: SwinIR/testsets/Set5/LR_bicubic/X8/butterflyx8.png
creating: SwinIR/testsets/Set5/LR_bicubic/X4/
extracting: SwinIR/testsets/Set5/LR_bicubic/X4/birdx4.png
extracting: SwinIR/testsets/Set5/LR_bicubic/X4/butterflyx4.png
extracting: SwinIR/testsets/Set5/LR_bicubic/X4/babyx4.png
extracting: SwinIR/testsets/Set5/LR_bicubic/X4/womanx4.png
extracting: SwinIR/testsets/Set5/LR_bicubic/X4/headx4.png
creating: SwinIR/testsets/Set5/LR_bicubic/X3/
extracting: SwinIR/testsets/Set5/LR_bicubic/X3/butterflyx3.png
inflating: SwinIR/testsets/Set5/LR_bicubic/X3/babyx3.png
extracting: SwinIR/testsets/Set5/LR_bicubic/X3/womanx3.png
extracting: SwinIR/testsets/Set5/LR_bicubic/X3/birdx3.png
extracting: SwinIR/testsets/Set5/LR_bicubic/X3/headx3.png
creating: SwinIR/testsets/Set5/LR_bicubic/X2/
inflating: SwinIR/testsets/Set5/LR_bicubic/X2/womanx2.png
inflating: SwinIR/testsets/Set5/LR_bicubic/X2/headx2.png
inflating: SwinIR/testsets/Set5/LR_bicubic/X2/butterflyx2.png
inflating: SwinIR/testsets/Set5/LR_bicubic/X2/birdx2.png
inflating: SwinIR/testsets/Set5/LR_bicubic/X2/babyx2.png
creating: SwinIR/figs/
inflating: SwinIR/figs/ETH_SwinIR-L.png
inflating: SwinIR/figs/OST_009_crop_realESRGAN.png
inflating: SwinIR/figs/ETH_realESRGAN.jpg
inflating: SwinIR/figs/SwinIR_archi.png
inflating: SwinIR/figs/OST_009_crop_LR.png
inflating: SwinIR/figs/jepg_compress_artfact_reduction.png
inflating: SwinIR/figs/classic_image_sr.png
inflating: SwinIR/figs/ETH_SwinIR.png
inflating: SwinIR/figs/OST_009_crop_SwinIR.png
inflating: SwinIR/figs/ETH_BSRGAN.png
inflating: SwinIR/figs/color_image_denoising.png
inflating: SwinIR/figs/classic_image_sr_visual.png
inflating: SwinIR/figs/lightweight_image_sr.png
inflating: SwinIR/figs/gray_image_denoising.png
inflating: SwinIR/figs/OST_009_crop_SwinIR-L.png
inflating: SwinIR/figs/OST_009_crop_BSRGAN.png
inflating: SwinIR/figs/real_world_image_sr.png
inflating: SwinIR/figs/ETH_LR.png
creating: SwinIR/model_zoo/
inflating: SwinIR/model_zoo/README.md
inflating: SwinIR/predict.py loading model from /mnt/workspace/SwinIR/pretrained_model/001_classicalSR_DF2K_s64w8_SwinIR-M_x4.pth
/usr/local/lib/python3.10/dist-packages/torch/functional.py:504: UserWarning: torch.meshgrid: in an upcoming release, it will be required to pass the indexing argument. (Triggered internally at ../aten/src/ATen/native/TensorShape.cpp:3190.)
return _VF.meshgrid(tensors, **kwargs) # type: ignore[attr-defined]
results//swinir_classical_sr_x4
Testing 0 10
Testing 1 2
Testing 2 20
Testing 3 34
Testing 4 4
Testing 5 40
Testing 6 50
Testing 7 54
Testing 8 64
Testing 9 70
loading model from /mnt/workspace/SwinIR/pretrained_model/003_realSR_BSRGAN_DFO_s64w8_SwinIR-M_x4_GAN.pth
/usr/local/lib/python3.10/dist-packages/torch/functional.py:504: UserWarning: torch.meshgrid: in an upcoming release, it will be required to pass the indexing argument. (Triggered internally at ../aten/src/ATen/native/TensorShape.cpp:3190.)
return _VF.meshgrid(tensors, **kwargs) # type: ignore[attr-defined]
results//swinir_real_sr_x4
Testing 0 10
Testing 1 2
Testing 2 20
Testing 3 34
Testing 4 4
Testing 5 40
Testing 6 50
Testing 7 54
Testing 8 64
Testing 9 70
loading model from /mnt/workspace/SwinIR/pretrained_model/004_grayDN_DFWB_s128w8_SwinIR-M_noise15.pth
/usr/local/lib/python3.10/dist-packages/torch/functional.py:504: UserWarning: torch.meshgrid: in an upcoming release, it will be required to pass the indexing argument. (Triggered internally at ../aten/src/ATen/native/TensorShape.cpp:3190.)
return _VF.meshgrid(tensors, **kwargs) # type: ignore[attr-defined]
results//swinir_gray_dn_noise15
Testing 0 10 - PSNR: 33.61 dB; SSIM: 0.9556; PSNRB: 0.00 dB;PSNR_Y: 0.00 dB; SSIM_Y: 0.0000; PSNRB_Y: 0.00 dB.
Testing 1 2 - PSNR: 33.54 dB; SSIM: 0.9048; PSNRB: 0.00 dB;PSNR_Y: 0.00 dB; SSIM_Y: 0.0000; PSNRB_Y: 0.00 dB.
Testing 2 20 - PSNR: 32.79 dB; SSIM: 0.9033; PSNRB: 0.00 dB;PSNR_Y: 0.00 dB; SSIM_Y: 0.0000; PSNRB_Y: 0.00 dB.
Testing 3 34 - PSNR: 32.80 dB; SSIM: 0.8973; PSNRB: 0.00 dB;PSNR_Y: 0.00 dB; SSIM_Y: 0.0000; PSNRB_Y: 0.00 dB.
Testing 4 4 - PSNR: 39.13 dB; SSIM: 0.9357; PSNRB: 0.00 dB;PSNR_Y: 0.00 dB; SSIM_Y: 0.0000; PSNRB_Y: 0.00 dB.
Testing 5 40 - PSNR: 31.12 dB; SSIM: 0.9500; PSNRB: 0.00 dB;PSNR_Y: 0.00 dB; SSIM_Y: 0.0000; PSNRB_Y: 0.00 dB.
Testing 6 50 - PSNR: 31.57 dB; SSIM: 0.9437; PSNRB: 0.00 dB;PSNR_Y: 0.00 dB; SSIM_Y: 0.0000; PSNRB_Y: 0.00 dB.
Testing 7 54 - PSNR: 36.47 dB; SSIM: 0.9115; PSNRB: 0.00 dB;PSNR_Y: 0.00 dB; SSIM_Y: 0.0000; PSNRB_Y: 0.00 dB.
Testing 8 64 - PSNR: 35.19 dB; SSIM: 0.9507; PSNRB: 0.00 dB;PSNR_Y: 0.00 dB; SSIM_Y: 0.0000; PSNRB_Y: 0.00 dB.
Testing 9 70 - PSNR: 31.88 dB; SSIM: 0.8856; PSNRB: 0.00 dB;PSNR_Y: 0.00 dB; SSIM_Y: 0.0000; PSNRB_Y: 0.00 dB.
results//swinir_gray_dn_noise15
-- Average PSNR/SSIM(RGB): 33.81 dB; 0.9238
loading model from /mnt/workspace/SwinIR/pretrained_model/005_colorDN_DFWB_s128w8_SwinIR-M_noise15.pth
/usr/local/lib/python3.10/dist-packages/torch/functional.py:504: UserWarning: torch.meshgrid: in an upcoming release, it will be required to pass the indexing argument. (Triggered internally at ../aten/src/ATen/native/TensorShape.cpp:3190.)
return _VF.meshgrid(tensors, **kwargs) # type: ignore[attr-defined]
results//swinir_color_dn_noise15
Testing 0 10 - PSNR: 36.38 dB; SSIM: 0.9705; PSNRB: 0.00 dB;PSNR_Y: 37.74 dB; SSIM_Y: 0.9749; PSNRB_Y: 0.00 dB.
Testing 1 2 - PSNR: 35.28 dB; SSIM: 0.9327; PSNRB: 0.00 dB;PSNR_Y: 37.29 dB; SSIM_Y: 0.9463; PSNRB_Y: 0.00 dB.
Testing 2 20 - PSNR: 35.06 dB; SSIM: 0.9315; PSNRB: 0.00 dB;PSNR_Y: 36.52 dB; SSIM_Y: 0.9432; PSNRB_Y: 0.00 dB.
Testing 3 34 - PSNR: 35.23 dB; SSIM: 0.9307; PSNRB: 0.00 dB;PSNR_Y: 36.58 dB; SSIM_Y: 0.9418; PSNRB_Y: 0.00 dB.
Testing 4 4 - PSNR: 39.07 dB; SSIM: 0.9320; PSNRB: 0.00 dB;PSNR_Y: 41.84 dB; SSIM_Y: 0.9568; PSNRB_Y: 0.00 dB.
Testing 5 40 - PSNR: 34.48 dB; SSIM: 0.9716; PSNRB: 0.00 dB;PSNR_Y: 35.83 dB; SSIM_Y: 0.9751; PSNRB_Y: 0.00 dB.
Testing 6 50 - PSNR: 34.92 dB; SSIM: 0.9648; PSNRB: 0.00 dB;PSNR_Y: 36.27 dB; SSIM_Y: 0.9702; PSNRB_Y: 0.00 dB.
Testing 7 54 - PSNR: 38.24 dB; SSIM: 0.9331; PSNRB: 0.00 dB;PSNR_Y: 39.60 dB; SSIM_Y: 0.9463; PSNRB_Y: 0.00 dB.
Testing 8 64 - PSNR: 37.77 dB; SSIM: 0.9678; PSNRB: 0.00 dB;PSNR_Y: 39.14 dB; SSIM_Y: 0.9733; PSNRB_Y: 0.00 dB.
Testing 9 70 - PSNR: 34.51 dB; SSIM: 0.9226; PSNRB: 0.00 dB;PSNR_Y: 35.85 dB; SSIM_Y: 0.9349; PSNRB_Y: 0.00 dB.
results//swinir_color_dn_noise15
-- Average PSNR/SSIM(RGB): 36.10 dB; 0.9457
-- Average PSNR_Y/SSIM_Y: 37.67 dB; 0.9563
loading model from /mnt/workspace/SwinIR/pretrained_model/006_colorCAR_DFWB_s126w7_SwinIR-M_jpeg10.pth
/usr/local/lib/python3.10/dist-packages/torch/functional.py:504: UserWarning: torch.meshgrid: in an upcoming release, it will be required to pass the indexing argument. (Triggered internally at ../aten/src/ATen/native/TensorShape.cpp:3190.)
return _VF.meshgrid(tensors, **kwargs) # type: ignore[attr-defined]
results//swinir_color_jpeg_car_jpeg10
Testing 0 10
Testing 1 2
Testing 2 20
Testing 3 34
Testing 4 4
Testing 5 40
Testing 6 50
Testing 7 54
Testing 8 64
Testing 9 70 http://pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com/aigc-data/restoration/repo/hat.zip
cn-hangzhou
--2023-09-05 01:47:39-- http://pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com/aigc-data/restoration/repo/hat.zip
Resolving pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com (pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com)... 100.118.28.49, 100.118.28.45, 100.118.28.44, ...
Connecting to pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com (pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com)|100.118.28.49|:80... connected.
HTTP request sent, awaiting response... 200 OK
Length: 788324861 (752M) [application/zip]
Saving to: ‘hat.zip’
hat.zip 100%[===================>] 751.80M 12.1MB/s in 68s
2023-09-05 01:48:47 (11.1 MB/s) - ‘hat.zip’ saved [788324861/788324861]
Archive: hat.zip
creating: HAT/
creating: HAT/.ipynb_checkpoints/
inflating: HAT/.ipynb_checkpoints/predict-checkpoint.py
inflating: HAT/.ipynb_checkpoints/README-checkpoint.md
inflating: HAT/setup.cfg
extracting: HAT/requirements.txt
creating: HAT/pretrained_model/
inflating: HAT/pretrained_model/HAT-L_SRx4_ImageNet-pretrain.pth
inflating: HAT/pretrained_model/HAT-L_SRx2_ImageNet-pretrain.pth
inflating: HAT/pretrained_model/Real_HAT_GAN_sharper.pth
inflating: HAT/pretrained_model/Real_HAT_GAN_SRx4.pth
inflating: HAT/.gitignore
creating: HAT/datasets/
inflating: HAT/datasets/README.md
creating: HAT/results/
creating: HAT/results/HAT_GAN_Real_SRx4_archived_20230811_075007/
inflating: HAT/results/HAT_GAN_Real_SRx4_archived_20230811_075007/test_HAT_GAN_Real_SRx4_20230811_074626.log
creating: HAT/results/HAT_GAN_Real_SRx4_archived_20230811_075007/visualization/
creating: HAT/results/HAT_GAN_Real_SRx4_archived_20230811_075007/visualization/custom/
inflating: HAT/results/HAT_GAN_Real_SRx4_archived_20230811_075007/visualization/custom/2_HAT_GAN_Real_SRx4.png
creating: HAT/results/HAT_GAN_Real_SRx4_archived_20230811_075650/
inflating: HAT/results/HAT_GAN_Real_SRx4_archived_20230811_075650/test_HAT_GAN_Real_SRx4_20230811_075049.log
creating: HAT/results/HAT_GAN_Real_SRx4_archived_20230811_075650/visualization/
creating: HAT/results/HAT_GAN_Real_SRx4_archived_20230811_075650/visualization/custom/
creating: HAT/results/HAT_GAN_Real_SRx4_archived_20230811_075650/visualization/custom/.ipynb_checkpoints/
inflating: HAT/results/HAT_GAN_Real_SRx4_archived_20230811_075650/visualization/custom/.ipynb_checkpoints/IMG_1452_HAT_GAN_Real_SRx4-checkpoint.png
inflating: HAT/results/HAT_GAN_Real_SRx4_archived_20230811_075650/visualization/custom/.ipynb_checkpoints/2_HAT_GAN_Real_SRx4-checkpoint.png
inflating: HAT/results/HAT_GAN_Real_SRx4_archived_20230811_075650/visualization/custom/IMG_1452_HAT_GAN_Real_SRx4.png
inflating: HAT/results/HAT_GAN_Real_SRx4_archived_20230811_075650/visualization/custom/2_HAT_GAN_Real_SRx4.png
creating: HAT/results/HAT_GAN_Real_SRx4_archived_20230811_081135/
creating: HAT/results/HAT_GAN_Real_SRx4_archived_20230811_081135/visualization/
creating: HAT/results/HAT_GAN_Real_SRx4_archived_20230811_074626/
creating: HAT/results/HAT_GAN_Real_SRx4_archived_20230811_074626/visualization/
inflating: HAT/results/HAT_GAN_Real_SRx4_archived_20230811_074626/test_HAT_GAN_Real_SRx4_20230811_074612.log
creating: HAT/results/HAT_GAN_Real_SRx4/
creating: HAT/results/HAT_GAN_Real_SRx4/visualization/
creating: HAT/results/HAT_GAN_Real_SRx4/visualization/custom/
inflating: HAT/results/HAT_GAN_Real_SRx4/visualization/custom/IMG_1452_HAT_GAN_Real_SRx4.png
inflating: HAT/results/HAT_GAN_Real_SRx4/visualization/custom/2_HAT_GAN_Real_SRx4.png
creating: HAT/results/HAT_GAN_Real_SRx4_archived_20230811_081052/
inflating: HAT/results/HAT_GAN_Real_SRx4_archived_20230811_081052/test_HAT_GAN_Real_SRx4_20230811_075714.log
creating: HAT/results/HAT_GAN_Real_SRx4_archived_20230811_081052/visualization/
creating: HAT/results/HAT_GAN_Real_SRx4_archived_20230811_074612/
inflating: HAT/results/HAT_GAN_Real_SRx4_archived_20230811_074612/test_HAT_GAN_Real_SRx4_20230811_074331.log
creating: HAT/results/HAT_GAN_Real_SRx4_archived_20230811_074612/visualization/
creating: HAT/results/HAT_GAN_Real_SRx4_archived_20230811_081900/
creating: HAT/results/HAT_GAN_Real_SRx4_archived_20230811_081900/visualization/
creating: HAT/results/HAT_GAN_Real_SRx4_archived_20230811_081900/visualization/custom/
inflating: HAT/results/HAT_GAN_Real_SRx4_archived_20230811_081900/visualization/custom/IMG_1452_HAT_GAN_Real_SRx4.png
inflating: HAT/results/HAT_GAN_Real_SRx4_archived_20230811_081900/visualization/custom/2_HAT_GAN_Real_SRx4.png
creating: HAT/results/HAT_GAN_Real_SRx4_archived_20230811_082002/
creating: HAT/results/HAT_GAN_Real_SRx4_archived_20230811_082002/visualization/
extracting: HAT/results/README.md
creating: HAT/results/HAT_SRx4_ImageNet-pretrain/
inflating: HAT/results/HAT_SRx4_ImageNet-pretrain/test_HAT_SRx4_ImageNet-pretrain_20230811_074251.log
creating: HAT/results/HAT_SRx4_ImageNet-pretrain/visualization/
creating: HAT/results/HAT_GAN_Real_SRx4_archived_20230811_082148/
creating: HAT/results/HAT_GAN_Real_SRx4_archived_20230811_082148/visualization/
creating: HAT/results/HAT_GAN_Real_SRx4_archived_20230811_082240/
creating: HAT/results/HAT_GAN_Real_SRx4_archived_20230811_082240/visualization/
creating: HAT/results/HAT_GAN_Real_SRx4_archived_20230811_081942/
creating: HAT/results/HAT_GAN_Real_SRx4_archived_20230811_081942/visualization/
creating: HAT/results/HAT_GAN_Real_SRx4_archived_20230811_075049/
inflating: HAT/results/HAT_GAN_Real_SRx4_archived_20230811_075049/test_HAT_GAN_Real_SRx4_20230811_075007.log
creating: HAT/results/HAT_GAN_Real_SRx4_archived_20230811_075049/visualization/
creating: HAT/results/HAT_GAN_Real_SRx4_archived_20230811_075049/visualization/custom/
inflating: HAT/results/HAT_GAN_Real_SRx4_archived_20230811_075049/visualization/custom/2_HAT_GAN_Real_SRx4.png
creating: HAT/results/HAT_GAN_Real_SRx4_archived_20230811_081209/
creating: HAT/results/HAT_GAN_Real_SRx4_archived_20230811_081209/visualization/
creating: HAT/results/HAT_GAN_Real_SRx4_archived_20230811_075714/
inflating: HAT/results/HAT_GAN_Real_SRx4_archived_20230811_075714/test_HAT_GAN_Real_SRx4_20230811_075650.log
creating: HAT/results/HAT_GAN_Real_SRx4_archived_20230811_075714/visualization/
creating: HAT/experiments/
creating: HAT/experiments/pretrained_models/
extracting: HAT/experiments/pretrained_models/README.md
inflating: HAT/cog.yaml
inflating: HAT/LICENSE
creating: HAT/hat/
creating: HAT/hat/.ipynb_checkpoints/
inflating: HAT/hat/.ipynb_checkpoints/__init__-checkpoint.py
inflating: HAT/hat/.ipynb_checkpoints/test-checkpoint.py
creating: HAT/hat/models/
creating: HAT/hat/models/.ipynb_checkpoints/
inflating: HAT/hat/models/.ipynb_checkpoints/__init__-checkpoint.py
inflating: HAT/hat/models/.ipynb_checkpoints/hat_model-checkpoint.py
inflating: HAT/hat/models/realhatgan_model.py
creating: HAT/hat/models/__pycache__/
inflating: HAT/hat/models/__pycache__/__init__.cpython-310.pyc
inflating: HAT/hat/models/__pycache__/hat_model.cpython-310.pyc
inflating: HAT/hat/models/__pycache__/realhatmse_model.cpython-310.pyc
inflating: HAT/hat/models/__pycache__/realhatgan_model.cpython-310.pyc
inflating: HAT/hat/models/hat_model.py
inflating: HAT/hat/models/realhatmse_model.py
inflating: HAT/hat/models/__init__.py
inflating: HAT/hat/train.py
creating: HAT/hat/results/
creating: HAT/hat/results/HAT_GAN_Real_SRx4/
inflating: HAT/hat/results/HAT_GAN_Real_SRx4/test_HAT_GAN_Real_SRx4_20230811_075405.log
creating: HAT/hat/results/HAT_GAN_Real_SRx4/visualization/
inflating: HAT/hat/test.py
creating: HAT/hat/archs/
creating: HAT/hat/archs/.ipynb_checkpoints/
inflating: HAT/hat/archs/.ipynb_checkpoints/__init__-checkpoint.py
creating: HAT/hat/archs/__pycache__/
inflating: HAT/hat/archs/__pycache__/__init__.cpython-310.pyc
inflating: HAT/hat/archs/__pycache__/hat_arch.cpython-310.pyc
inflating: HAT/hat/archs/hat_arch.py
inflating: HAT/hat/archs/__init__.py
creating: HAT/hat/data/
creating: HAT/hat/data/meta_info/
inflating: HAT/hat/data/meta_info/meta_info_DF2Ksub_GT.txt
creating: HAT/hat/data/.ipynb_checkpoints/
inflating: HAT/hat/data/.ipynb_checkpoints/__init__-checkpoint.py
inflating: HAT/hat/data/.ipynb_checkpoints/imagenet_paired_dataset-checkpoint.py
creating: HAT/hat/data/__pycache__/
inflating: HAT/hat/data/__pycache__/__init__.cpython-310.pyc
inflating: HAT/hat/data/__pycache__/imagenet_paired_dataset.cpython-310.pyc
inflating: HAT/hat/data/__pycache__/realesrgan_dataset.cpython-310.pyc
inflating: HAT/hat/data/realesrgan_dataset.py
inflating: HAT/hat/data/imagenet_paired_dataset.py
inflating: HAT/hat/data/__init__.py
inflating: HAT/hat/__init__.py
creating: HAT/.git/
creating: HAT/.git/logs/
creating: HAT/.git/logs/refs/
creating: HAT/.git/logs/refs/remotes/
creating: HAT/.git/logs/refs/remotes/origin/
inflating: HAT/.git/logs/refs/remotes/origin/HEAD
creating: HAT/.git/logs/refs/heads/
inflating: HAT/.git/logs/refs/heads/main
inflating: HAT/.git/logs/HEAD
inflating: HAT/.git/config
creating: HAT/.git/refs/
creating: HAT/.git/refs/remotes/
creating: HAT/.git/refs/remotes/origin/
extracting: HAT/.git/refs/remotes/origin/HEAD
creating: HAT/.git/refs/heads/
extracting: HAT/.git/refs/heads/main
creating: HAT/.git/refs/tags/
extracting: HAT/.git/HEAD
creating: HAT/.git/hooks/
inflating: HAT/.git/hooks/push-to-checkout.sample
inflating: HAT/.git/hooks/commit-msg.sample
inflating: HAT/.git/hooks/applypatch-msg.sample
inflating: HAT/.git/hooks/pre-receive.sample
inflating: HAT/.git/hooks/pre-push.sample
inflating: HAT/.git/hooks/fsmonitor-watchman.sample
inflating: HAT/.git/hooks/post-update.sample
inflating: HAT/.git/hooks/update.sample
inflating: HAT/.git/hooks/pre-merge-commit.sample
inflating: HAT/.git/hooks/pre-commit.sample
inflating: HAT/.git/hooks/pre-applypatch.sample
inflating: HAT/.git/hooks/prepare-commit-msg.sample
inflating: HAT/.git/hooks/pre-rebase.sample
creating: HAT/.git/info/
inflating: HAT/.git/info/exclude
creating: HAT/.git/objects/
creating: HAT/.git/objects/pack/
inflating: HAT/.git/objects/pack/pack-bf02a359cd3a677a5831135490add5f47b36243f.idx
inflating: HAT/.git/objects/pack/pack-bf02a359cd3a677a5831135490add5f47b36243f.pack
creating: HAT/.git/objects/info/
inflating: HAT/.git/packed-refs
inflating: HAT/.git/index
inflating: HAT/.git/description
inflating: HAT/README.md
creating: HAT/figures/
creating: HAT/figures/.ipynb_checkpoints/
inflating: HAT/figures/.ipynb_checkpoints/Comparison-checkpoint.png
inflating: HAT/figures/Comparison.png
inflating: HAT/figures/Performance_comparison.png
inflating: HAT/figures/Visual_Results.png
extracting: HAT/VERSION
inflating: HAT/setup.py
inflating: HAT/predict.py
creating: HAT/options/
creating: HAT/options/train/
inflating: HAT/options/train/train_HAT-L_SRx4_finetune_from_ImageNet_pretrain.yml
inflating: HAT/options/train/train_HAT-L_SRx2_ImageNet_from_scratch.yml
inflating: HAT/options/train/train_HAT-L_SRx2_finetune_from_ImageNet_pretrain.yml
inflating: HAT/options/train/train_Real_HAT_GAN_SRx4_finetune_from_mse_model.yml
inflating: HAT/options/train/train_HAT_SRx3_ImageNet_from_scratch.yml
inflating: HAT/options/train/train_HAT_SRx4_finetune_from_ImageNet_pretrain.yml
inflating: HAT/options/train/train_HAT_SRx2_finetune_from_ImageNet_pretrain.yml
inflating: HAT/options/train/train_HAT-L_SRx3_finetune_from_ImageNet_pretrain.yml
inflating: HAT/options/train/train_HAT_SRx4_finetune_from_SRx2.yml
inflating: HAT/options/train/train_HAT-S_SRx3_from_scratch.yml
inflating: HAT/options/train/train_HAT-S_SRx2_from_scratch.yml
inflating: HAT/options/train/train_HAT_SRx2_from_scratch.yml
inflating: HAT/options/train/train_HAT_SRx3_from_scratch.yml
inflating: HAT/options/train/train_HAT-L_SRx3_ImageNet_from_scratch.yml
inflating: HAT/options/train/train_Real_HAT_SRx4_mse_model.yml
inflating: HAT/options/train/train_HAT-L_SRx4_ImageNet_from_scratch.yml
inflating: HAT/options/train/train_HAT_SRx4_ImageNet_from_scratch.yml
inflating: HAT/options/train/train_HAT-S_SRx4_finetune_from_SRx2.yml
inflating: HAT/options/train/train_HAT_SRx2_ImageNet_from_scratch.yml
inflating: HAT/options/train/train_HAT_SRx3_finetune_from_ImageNet_pretrain.yml
creating: HAT/options/test/
inflating: HAT/options/test/HAT_SRx4_ImageNet-pretrain.yml
creating: HAT/options/test/.ipynb_checkpoints/
inflating: HAT/options/test/.ipynb_checkpoints/HAT_SRx2_ImageNet-pretrain-checkpoint.yml
inflating: HAT/options/test/.ipynb_checkpoints/HAT-L_SRx4_ImageNet-pretrain-checkpoint.yml
inflating: HAT/options/test/.ipynb_checkpoints/HAT-L_SRx2_ImageNet-pretrain-checkpoint.yml
inflating: HAT/options/test/.ipynb_checkpoints/HAT_GAN_Real_SRx4-checkpoint.yml
inflating: HAT/options/test/HAT_SRx2_ImageNet-pretrain.yml
inflating: HAT/options/test/HAT_GAN_Real_SRx4.yml
inflating: HAT/options/test/HAT-S_SRx2.yml
inflating: HAT/options/test/HAT_SRx4_ImageNet-LR.yml
inflating: HAT/options/test/HAT-L_SRx4_ImageNet-pretrain.yml
inflating: HAT/options/test/HAT-S_SRx3.yml
inflating: HAT/options/test/HAT_tile_example.yml
inflating: HAT/options/test/HAT_SRx3.yml
inflating: HAT/options/test/HAT_SRx3_ImageNet-pretrain.yml
inflating: HAT/options/test/HAT-S_SRx4.yml
inflating: HAT/options/test/HAT-L_SRx3_ImageNet-pretrain.yml
inflating: HAT/options/test/HAT_SRx4.yml
inflating: HAT/options/test/HAT-L_SRx2_ImageNet-pretrain.yml
inflating: HAT/options/test/HAT_SRx2.yml /mnt/workspace/HAT
/usr/local/lib/python3.10/dist-packages/torch/functional.py:504: UserWarning: torch.meshgrid: in an upcoming release, it will be required to pass the indexing argument. (Triggered internally at ../aten/src/ATen/native/TensorShape.cpp:3190.)
return _VF.meshgrid(tensors, **kwargs) # type: ignore[attr-defined]
Tile 1/4
Tile 2/4
Tile 3/4
Tile 4/4
Tile 1/16
Tile 2/16
Tile 3/16
Tile 4/16
Tile 5/16
Tile 6/16
Tile 7/16
Tile 8/16
Tile 9/16
Tile 10/16
Tile 11/16
Tile 12/16
Tile 13/16
Tile 14/16
Tile 15/16
Tile 16/16
Tile 1/16
Tile 2/16
Tile 3/16
Tile 4/16
Tile 5/16
Tile 6/16
Tile 7/16
Tile 8/16
Tile 9/16
Tile 10/16
Tile 11/16
Tile 12/16
Tile 13/16
Tile 14/16
Tile 15/16
Tile 16/16
Tile 1/16
Tile 2/16
Tile 3/16
Tile 4/16
Tile 5/16
Tile 6/16
Tile 7/16
Tile 8/16
Tile 9/16
Tile 10/16
Tile 11/16
Tile 12/16
Tile 13/16
Tile 14/16
Tile 15/16
Tile 16/16
Tile 1/12
Tile 2/12
Tile 3/12
Tile 4/12
Tile 5/12
Tile 6/12
Tile 7/12
Tile 8/12
Tile 9/12
Tile 10/12
Tile 11/12
Tile 12/12
Tile 1/4
Tile 2/4
Tile 3/4
Tile 4/4
Tile 1/4
Tile 2/4
Tile 3/4
Tile 4/4
Tile 1/16
Tile 2/16
Tile 3/16
Tile 4/16
Tile 5/16
Tile 6/16
Tile 7/16
Tile 8/16
Tile 9/16
Tile 10/16
Tile 11/16
Tile 12/16
Tile 13/16
Tile 14/16
Tile 15/16
Tile 16/16
Tile 1/4
Tile 2/4
Tile 3/4
Tile 4/4
Tile 1/16
Tile 2/16
Tile 3/16
Tile 4/16
Tile 5/16
Tile 6/16
Tile 7/16
Tile 8/16
Tile 9/16
Tile 10/16
Tile 11/16
Tile 12/16
Tile 13/16
Tile 14/16
Tile 15/16
Tile 16/16
/mnt/workspace/HAT
/usr/local/lib/python3.10/dist-packages/torch/functional.py:504: UserWarning: torch.meshgrid: in an upcoming release, it will be required to pass the indexing argument. (Triggered internally at ../aten/src/ATen/native/TensorShape.cpp:3190.)
return _VF.meshgrid(tensors, **kwargs) # type: ignore[attr-defined]
Tile 1/4
Tile 2/4
Tile 3/4
Tile 4/4
Tile 1/16
Tile 2/16
Tile 3/16
Tile 4/16
Tile 5/16
Tile 6/16
Tile 7/16
Tile 8/16
Tile 9/16
Tile 10/16
Tile 11/16
Tile 12/16
Tile 13/16
Tile 14/16
Tile 15/16
Tile 16/16
Tile 1/16
Tile 2/16
Tile 3/16
Tile 4/16
Tile 5/16
Tile 6/16
Tile 7/16
Tile 8/16
Tile 9/16
Tile 10/16
Tile 11/16
Tile 12/16
Tile 13/16
Tile 14/16
Tile 15/16
Tile 16/16
Tile 1/16
Tile 2/16
Tile 3/16
Tile 4/16
Tile 5/16
Tile 6/16
Tile 7/16
Tile 8/16
Tile 9/16
Tile 10/16
Tile 11/16
Tile 12/16
Tile 13/16
Tile 14/16
Tile 15/16
Tile 16/16
Tile 1/12
Tile 2/12
Tile 3/12
Tile 4/12
Tile 5/12
Tile 6/12
Tile 7/12
Tile 8/12
Tile 9/12
Tile 10/12
Tile 11/12
Tile 12/12
Tile 1/4
Tile 2/4
Tile 3/4
Tile 4/4
Tile 1/4
Tile 2/4
Tile 3/4
Tile 4/4
Tile 1/16
Tile 2/16
Tile 3/16
Tile 4/16
Tile 5/16
Tile 6/16
Tile 7/16
Tile 8/16
Tile 9/16
Tile 10/16
Tile 11/16
Tile 12/16
Tile 13/16
Tile 14/16
Tile 15/16
Tile 16/16
Tile 1/4
Tile 2/4
Tile 3/4
Tile 4/4
Tile 1/16
Tile 2/16
Tile 3/16
Tile 4/16
Tile 5/16
Tile 6/16
Tile 7/16
Tile 8/16
Tile 9/16
Tile 10/16
Tile 11/16
Tile 12/16
Tile 13/16
Tile 14/16
Tile 15/16
Tile 16/16
/mnt/workspace/HAT
/usr/local/lib/python3.10/dist-packages/torch/functional.py:504: UserWarning: torch.meshgrid: in an upcoming release, it will be required to pass the indexing argument. (Triggered internally at ../aten/src/ATen/native/TensorShape.cpp:3190.)
return _VF.meshgrid(tensors, **kwargs) # type: ignore[attr-defined]
Tile 1/4
Tile 2/4
Tile 3/4
Tile 4/4
Tile 1/16
Tile 2/16
Tile 3/16
Tile 4/16
Tile 5/16
Tile 6/16
Tile 7/16
Tile 8/16
Tile 9/16
Tile 10/16
Tile 11/16
Tile 12/16
Tile 13/16
Tile 14/16
Tile 15/16
Tile 16/16
Tile 1/16
Tile 2/16
Tile 3/16
Tile 4/16
Tile 5/16
Tile 6/16
Tile 7/16
Tile 8/16
Tile 9/16
Tile 10/16
Tile 11/16
Tile 12/16
Tile 13/16
Tile 14/16
Tile 15/16
Tile 16/16
Tile 1/16
Tile 2/16
Tile 3/16
Tile 4/16
Tile 5/16
Tile 6/16
Tile 7/16
Tile 8/16
Tile 9/16
Tile 10/16
Tile 11/16
Tile 12/16
Tile 13/16
Tile 14/16
Tile 15/16
Tile 16/16
Tile 1/12
Tile 2/12
Tile 3/12
Tile 4/12
Tile 5/12
Tile 6/12
Tile 7/12
Tile 8/12
Tile 9/12
Tile 10/12
Tile 11/12
Tile 12/12
Tile 1/4
Tile 2/4
Tile 3/4
Tile 4/4
Tile 1/4
Tile 2/4
Tile 3/4
Tile 4/4
Tile 1/16
Tile 2/16
Tile 3/16
Tile 4/16
Tile 5/16
Tile 6/16
Tile 7/16
Tile 8/16
Tile 9/16
Tile 10/16
Tile 11/16
Tile 12/16
Tile 13/16
Tile 14/16
Tile 15/16
Tile 16/16
Tile 1/4
Tile 2/4
Tile 3/4
Tile 4/4
Tile 1/16
Tile 2/16
Tile 3/16
Tile 4/16
Tile 5/16
Tile 6/16
Tile 7/16
Tile 8/16
Tile 9/16
Tile 10/16
Tile 11/16
Tile 12/16
Tile 13/16
Tile 14/16
Tile 15/16
Tile 16/16
/mnt/workspace/HAT
/usr/local/lib/python3.10/dist-packages/torch/functional.py:504: UserWarning: torch.meshgrid: in an upcoming release, it will be required to pass the indexing argument. (Triggered internally at ../aten/src/ATen/native/TensorShape.cpp:3190.)
return _VF.meshgrid(tensors, **kwargs) # type: ignore[attr-defined]
Tile 1/4
Tile 2/4
Tile 3/4
Tile 4/4
Tile 1/16
Tile 2/16
Tile 3/16
Tile 4/16
Tile 5/16
Tile 6/16
Tile 7/16
Tile 8/16
Tile 9/16
Tile 10/16
Tile 11/16
Tile 12/16
Tile 13/16
Tile 14/16
Tile 15/16
Tile 16/16
Tile 1/16
Tile 2/16
Tile 3/16
Tile 4/16
Tile 5/16
Tile 6/16
Tile 7/16
Tile 8/16
Tile 9/16
Tile 10/16
Tile 11/16
Tile 12/16
Tile 13/16
Tile 14/16
Tile 15/16
Tile 16/16
Tile 1/16
Tile 2/16
Tile 3/16
Tile 4/16
Tile 5/16
Tile 6/16
Tile 7/16
Tile 8/16
Tile 9/16
Tile 10/16
Tile 11/16
Tile 12/16
Tile 13/16
Tile 14/16
Tile 15/16
Tile 16/16
Tile 1/12
Tile 2/12
Tile 3/12
Tile 4/12
Tile 5/12
Tile 6/12
Tile 7/12
Tile 8/12
Tile 9/12
Tile 10/12
Tile 11/12
Tile 12/12
Tile 1/4
Tile 2/4
Tile 3/4
Tile 4/4
Tile 1/4
Tile 2/4
Tile 3/4
Tile 4/4
Tile 1/16
Tile 2/16
Tile 3/16
Tile 4/16
Tile 5/16
Tile 6/16
Tile 7/16
Tile 8/16
Tile 9/16
Tile 10/16
Tile 11/16
Tile 12/16
Tile 13/16
Tile 14/16
Tile 15/16
Tile 16/16
Tile 1/4
Tile 2/4
Tile 3/4
Tile 4/4
Tile 1/16
Tile 2/16
Tile 3/16
Tile 4/16
Tile 5/16
Tile 6/16
Tile 7/16
Tile 8/16
Tile 9/16
Tile 10/16
Tile 11/16
Tile 12/16
Tile 13/16
Tile 14/16
Tile 15/16
Tile 16/16
http://pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com/aigc-data/restoration/repo/codeformer.zip
cn-hangzhou
--2023-09-05 02:21:39-- http://pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com/aigc-data/restoration/repo/codeformer.zip
Resolving pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com (pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com)... 100.118.28.44, 100.118.28.50, 100.118.28.49, ...
Connecting to pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com (pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com)|100.118.28.44|:80... connected.
HTTP request sent, awaiting response... 200 OK
Length: 627352702 (598M) [application/zip]
Saving to: ‘codeformer.zip’
codeformer.zip 100%[===================>] 598.29M 12.1MB/s in 55s
2023-09-05 02:22:34 (10.9 MB/s) - ‘codeformer.zip’ saved [627352702/627352702]
Archive: codeformer.zip
creating: CodeFormer/
inflating: CodeFormer/requirements.txt
creating: CodeFormer/assets/
inflating: CodeFormer/assets/network.jpg
inflating: CodeFormer/assets/inpainting_result1.png
inflating: CodeFormer/assets/inpainting_result2.png
inflating: CodeFormer/assets/imgsli_2.jpg
inflating: CodeFormer/assets/imgsli_1.jpg
inflating: CodeFormer/assets/restoration_result2.png
inflating: CodeFormer/assets/restoration_result4.png
inflating: CodeFormer/assets/color_enhancement_result1.png
inflating: CodeFormer/assets/imgsli_3.jpg
inflating: CodeFormer/assets/CodeFormer_logo.png
inflating: CodeFormer/assets/color_enhancement_result2.png
inflating: CodeFormer/assets/restoration_result1.png
inflating: CodeFormer/assets/restoration_result3.png
inflating: CodeFormer/.gitignore
inflating: CodeFormer/demo.py
inflating: CodeFormer/LICENSE
creating: CodeFormer/basicsr/
creating: CodeFormer/basicsr/.ipynb_checkpoints/
inflating: CodeFormer/basicsr/.ipynb_checkpoints/__init__-checkpoint.py
creating: CodeFormer/basicsr/models/
inflating: CodeFormer/basicsr/models/codeformer_joint_model.py
creating: CodeFormer/basicsr/models/__pycache__/
inflating: CodeFormer/basicsr/models/__pycache__/__init__.cpython-310.pyc
inflating: CodeFormer/basicsr/models/__pycache__/codeformer_joint_model.cpython-310.pyc
inflating: CodeFormer/basicsr/models/__pycache__/codeformer_model.cpython-310.pyc
inflating: CodeFormer/basicsr/models/__pycache__/sr_model.cpython-310.pyc
inflating: CodeFormer/basicsr/models/__pycache__/vqgan_model.cpython-310.pyc
inflating: CodeFormer/basicsr/models/__pycache__/base_model.cpython-310.pyc
inflating: CodeFormer/basicsr/models/__pycache__/codeformer_idx_model.cpython-310.pyc
inflating: CodeFormer/basicsr/models/__pycache__/lr_scheduler.cpython-310.pyc
inflating: CodeFormer/basicsr/models/codeformer_model.py
inflating: CodeFormer/basicsr/models/sr_model.py
inflating: CodeFormer/basicsr/models/codeformer_idx_model.py
inflating: CodeFormer/basicsr/models/base_model.py
inflating: CodeFormer/basicsr/models/lr_scheduler.py
inflating: CodeFormer/basicsr/models/__init__.py
inflating: CodeFormer/basicsr/models/vqgan_model.py
inflating: CodeFormer/basicsr/train.py
creating: CodeFormer/basicsr/__pycache__/
inflating: CodeFormer/basicsr/__pycache__/__init__.cpython-310.pyc
inflating: CodeFormer/basicsr/__pycache__/train.cpython-310.pyc
creating: CodeFormer/basicsr/archs/
creating: CodeFormer/basicsr/archs/__pycache__/
inflating: CodeFormer/basicsr/archs/__pycache__/__init__.cpython-310.pyc
inflating: CodeFormer/basicsr/archs/__pycache__/rrdbnet_arch.cpython-310.pyc
inflating: CodeFormer/basicsr/archs/__pycache__/vgg_arch.cpython-310.pyc
inflating: CodeFormer/basicsr/archs/__pycache__/vqgan_arch.cpython-310.pyc
inflating: CodeFormer/basicsr/archs/__pycache__/codeformer_arch.cpython-310.pyc
inflating: CodeFormer/basicsr/archs/__pycache__/arcface_arch.cpython-310.pyc
inflating: CodeFormer/basicsr/archs/__pycache__/arch_util.cpython-310.pyc
inflating: CodeFormer/basicsr/archs/arcface_arch.py
inflating: CodeFormer/basicsr/archs/vgg_arch.py
inflating: CodeFormer/basicsr/archs/arch_util.py
inflating: CodeFormer/basicsr/archs/vqgan_arch.py
inflating: CodeFormer/basicsr/archs/__init__.py
inflating: CodeFormer/basicsr/archs/codeformer_arch.py
inflating: CodeFormer/basicsr/archs/rrdbnet_arch.py
creating: CodeFormer/basicsr/ops/
creating: CodeFormer/basicsr/ops/__pycache__/
inflating: CodeFormer/basicsr/ops/__pycache__/__init__.cpython-310.pyc
creating: CodeFormer/basicsr/ops/fused_act/
inflating: CodeFormer/basicsr/ops/fused_act/fused_act.py
creating: CodeFormer/basicsr/ops/fused_act/src/
inflating: CodeFormer/basicsr/ops/fused_act/src/fused_bias_act_kernel.cu
inflating: CodeFormer/basicsr/ops/fused_act/src/fused_bias_act.cpp
inflating: CodeFormer/basicsr/ops/fused_act/__init__.py
creating: CodeFormer/basicsr/ops/dcn/
creating: CodeFormer/basicsr/ops/dcn/__pycache__/
inflating: CodeFormer/basicsr/ops/dcn/__pycache__/__init__.cpython-310.pyc
inflating: CodeFormer/basicsr/ops/dcn/__pycache__/deform_conv.cpython-310.pyc
inflating: CodeFormer/basicsr/ops/dcn/deform_conv.py
creating: CodeFormer/basicsr/ops/dcn/src/
inflating: CodeFormer/basicsr/ops/dcn/src/deform_conv_cuda_kernel.cu
inflating: CodeFormer/basicsr/ops/dcn/src/deform_conv_cuda.cpp
inflating: CodeFormer/basicsr/ops/dcn/src/deform_conv_ext.cpp
inflating: CodeFormer/basicsr/ops/dcn/__init__.py
creating: CodeFormer/basicsr/ops/upfirdn2d/
inflating: CodeFormer/basicsr/ops/upfirdn2d/upfirdn2d.py
creating: CodeFormer/basicsr/ops/upfirdn2d/src/
inflating: CodeFormer/basicsr/ops/upfirdn2d/src/upfirdn2d_kernel.cu
inflating: CodeFormer/basicsr/ops/upfirdn2d/src/upfirdn2d.cpp
inflating: CodeFormer/basicsr/ops/upfirdn2d/__init__.py
extracting: CodeFormer/basicsr/ops/__init__.py
creating: CodeFormer/basicsr/utils/
creating: CodeFormer/basicsr/utils/__pycache__/
inflating: CodeFormer/basicsr/utils/__pycache__/__init__.cpython-310.pyc
inflating: CodeFormer/basicsr/utils/__pycache__/dist_util.cpython-310.pyc
inflating: CodeFormer/basicsr/utils/__pycache__/img_util.cpython-310.pyc
inflating: CodeFormer/basicsr/utils/__pycache__/download_util.cpython-310.pyc
inflating: CodeFormer/basicsr/utils/__pycache__/logger.cpython-310.pyc
inflating: CodeFormer/basicsr/utils/__pycache__/registry.cpython-310.pyc
inflating: CodeFormer/basicsr/utils/__pycache__/options.cpython-310.pyc
inflating: CodeFormer/basicsr/utils/__pycache__/file_client.cpython-310.pyc
inflating: CodeFormer/basicsr/utils/__pycache__/misc.cpython-310.pyc
inflating: CodeFormer/basicsr/utils/__pycache__/matlab_functions.cpython-310.pyc
inflating: CodeFormer/basicsr/utils/__pycache__/realesrgan_utils.cpython-310.pyc
inflating: CodeFormer/basicsr/utils/file_client.py
inflating: CodeFormer/basicsr/utils/logger.py
inflating: CodeFormer/basicsr/utils/options.py
inflating: CodeFormer/basicsr/utils/video_util.py
inflating: CodeFormer/basicsr/utils/img_util.py
inflating: CodeFormer/basicsr/utils/matlab_functions.py
inflating: CodeFormer/basicsr/utils/download_util.py
inflating: CodeFormer/basicsr/utils/__init__.py
inflating: CodeFormer/basicsr/utils/realesrgan_utils.py
inflating: CodeFormer/basicsr/utils/misc.py
inflating: CodeFormer/basicsr/utils/dist_util.py
inflating: CodeFormer/basicsr/utils/lmdb_util.py
inflating: CodeFormer/basicsr/utils/registry.py
creating: CodeFormer/basicsr/metrics/
creating: CodeFormer/basicsr/metrics/__pycache__/
inflating: CodeFormer/basicsr/metrics/__pycache__/__init__.cpython-310.pyc
inflating: CodeFormer/basicsr/metrics/__pycache__/psnr_ssim.cpython-310.pyc
inflating: CodeFormer/basicsr/metrics/__pycache__/metric_util.cpython-310.pyc
inflating: CodeFormer/basicsr/metrics/metric_util.py
inflating: CodeFormer/basicsr/metrics/psnr_ssim.py
inflating: CodeFormer/basicsr/metrics/__init__.py
extracting: CodeFormer/basicsr/VERSION
inflating: CodeFormer/basicsr/setup.py
creating: CodeFormer/basicsr/data/
inflating: CodeFormer/basicsr/data/paired_image_dataset.py
inflating: CodeFormer/basicsr/data/gaussian_kernels.py
inflating: CodeFormer/basicsr/data/data_util.py
creating: CodeFormer/basicsr/data/__pycache__/
inflating: CodeFormer/basicsr/data/__pycache__/__init__.cpython-310.pyc
inflating: CodeFormer/basicsr/data/__pycache__/prefetch_dataloader.cpython-310.pyc
inflating: CodeFormer/basicsr/data/__pycache__/gaussian_kernels.cpython-310.pyc
inflating: CodeFormer/basicsr/data/__pycache__/ffhq_blind_dataset.cpython-310.pyc
inflating: CodeFormer/basicsr/data/__pycache__/paired_image_dataset.cpython-310.pyc
inflating: CodeFormer/basicsr/data/__pycache__/data_util.cpython-310.pyc
inflating: CodeFormer/basicsr/data/__pycache__/ffhq_blind_joint_dataset.cpython-310.pyc
inflating: CodeFormer/basicsr/data/__pycache__/transforms.cpython-310.pyc
inflating: CodeFormer/basicsr/data/__pycache__/data_sampler.cpython-310.pyc
inflating: CodeFormer/basicsr/data/__init__.py
inflating: CodeFormer/basicsr/data/ffhq_blind_dataset.py
inflating: CodeFormer/basicsr/data/ffhq_blind_joint_dataset.py
inflating: CodeFormer/basicsr/data/prefetch_dataloader.py
inflating: CodeFormer/basicsr/data/transforms.py
inflating: CodeFormer/basicsr/data/data_sampler.py
creating: CodeFormer/basicsr/losses/
creating: CodeFormer/basicsr/losses/__pycache__/
inflating: CodeFormer/basicsr/losses/__pycache__/__init__.cpython-310.pyc
inflating: CodeFormer/basicsr/losses/__pycache__/losses.cpython-310.pyc
inflating: CodeFormer/basicsr/losses/__pycache__/loss_util.cpython-310.pyc
inflating: CodeFormer/basicsr/losses/loss_util.py
inflating: CodeFormer/basicsr/losses/__init__.py
inflating: CodeFormer/basicsr/losses/losses.py
inflating: CodeFormer/basicsr/__init__.py
creating: CodeFormer/facelib/
creating: CodeFormer/facelib/parsing/
inflating: CodeFormer/facelib/parsing/resnet.py
creating: CodeFormer/facelib/parsing/__pycache__/
inflating: CodeFormer/facelib/parsing/__pycache__/__init__.cpython-310.pyc
inflating: CodeFormer/facelib/parsing/__pycache__/bisenet.cpython-310.pyc
inflating: CodeFormer/facelib/parsing/__pycache__/parsenet.cpython-310.pyc
inflating: CodeFormer/facelib/parsing/__pycache__/resnet.cpython-310.pyc
inflating: CodeFormer/facelib/parsing/__init__.py
inflating: CodeFormer/facelib/parsing/parsenet.py
inflating: CodeFormer/facelib/parsing/bisenet.py
creating: CodeFormer/facelib/utils/
creating: CodeFormer/facelib/utils/__pycache__/
inflating: CodeFormer/facelib/utils/__pycache__/__init__.cpython-310.pyc
inflating: CodeFormer/facelib/utils/__pycache__/misc.cpython-310.pyc
inflating: CodeFormer/facelib/utils/__pycache__/face_restoration_helper.cpython-310.pyc
inflating: CodeFormer/facelib/utils/__pycache__/face_utils.cpython-310.pyc
inflating: CodeFormer/facelib/utils/face_utils.py
inflating: CodeFormer/facelib/utils/__init__.py
inflating: CodeFormer/facelib/utils/misc.py
inflating: CodeFormer/facelib/utils/face_restoration_helper.py
creating: CodeFormer/facelib/detection/
creating: CodeFormer/facelib/detection/__pycache__/
inflating: CodeFormer/facelib/detection/__pycache__/__init__.cpython-310.pyc
inflating: CodeFormer/facelib/detection/__pycache__/matlab_cp2tform.cpython-310.pyc
inflating: CodeFormer/facelib/detection/__pycache__/align_trans.cpython-310.pyc
creating: CodeFormer/facelib/detection/retinaface/
inflating: CodeFormer/facelib/detection/retinaface/retinaface_net.py
creating: CodeFormer/facelib/detection/retinaface/__pycache__/
inflating: CodeFormer/facelib/detection/retinaface/__pycache__/retinaface_utils.cpython-310.pyc
inflating: CodeFormer/facelib/detection/retinaface/__pycache__/retinaface.cpython-310.pyc
inflating: CodeFormer/facelib/detection/retinaface/__pycache__/retinaface_net.cpython-310.pyc
inflating: CodeFormer/facelib/detection/retinaface/retinaface_utils.py
inflating: CodeFormer/facelib/detection/retinaface/retinaface.py
inflating: CodeFormer/facelib/detection/align_trans.py
inflating: CodeFormer/facelib/detection/__init__.py
inflating: CodeFormer/facelib/detection/matlab_cp2tform.py
creating: CodeFormer/facelib/detection/yolov5face/
creating: CodeFormer/facelib/detection/yolov5face/models/
creating: CodeFormer/facelib/detection/yolov5face/models/__pycache__/
inflating: CodeFormer/facelib/detection/yolov5face/models/__pycache__/__init__.cpython-310.pyc
inflating: CodeFormer/facelib/detection/yolov5face/models/__pycache__/yolo.cpython-310.pyc
inflating: CodeFormer/facelib/detection/yolov5face/models/__pycache__/common.cpython-310.pyc
inflating: CodeFormer/facelib/detection/yolov5face/models/__pycache__/experimental.cpython-310.pyc
inflating: CodeFormer/facelib/detection/yolov5face/models/yolov5l.yaml
inflating: CodeFormer/facelib/detection/yolov5face/models/experimental.py
inflating: CodeFormer/facelib/detection/yolov5face/models/yolov5n.yaml
inflating: CodeFormer/facelib/detection/yolov5face/models/yolo.py
inflating: CodeFormer/facelib/detection/yolov5face/models/common.py
extracting: CodeFormer/facelib/detection/yolov5face/models/__init__.py
creating: CodeFormer/facelib/detection/yolov5face/__pycache__/
inflating: CodeFormer/facelib/detection/yolov5face/__pycache__/__init__.cpython-310.pyc
inflating: CodeFormer/facelib/detection/yolov5face/__pycache__/face_detector.cpython-310.pyc
inflating: CodeFormer/facelib/detection/yolov5face/face_detector.py
creating: CodeFormer/facelib/detection/yolov5face/utils/
inflating: CodeFormer/facelib/detection/yolov5face/utils/torch_utils.py
creating: CodeFormer/facelib/detection/yolov5face/utils/__pycache__/
inflating: CodeFormer/facelib/detection/yolov5face/utils/__pycache__/__init__.cpython-310.pyc
inflating: CodeFormer/facelib/detection/yolov5face/utils/__pycache__/autoanchor.cpython-310.pyc
inflating: CodeFormer/facelib/detection/yolov5face/utils/__pycache__/torch_utils.cpython-310.pyc
inflating: CodeFormer/facelib/detection/yolov5face/utils/__pycache__/general.cpython-310.pyc
inflating: CodeFormer/facelib/detection/yolov5face/utils/__pycache__/datasets.cpython-310.pyc
inflating: CodeFormer/facelib/detection/yolov5face/utils/datasets.py
inflating: CodeFormer/facelib/detection/yolov5face/utils/autoanchor.py
inflating: CodeFormer/facelib/detection/yolov5face/utils/extract_ckpt.py
extracting: CodeFormer/facelib/detection/yolov5face/utils/__init__.py
inflating: CodeFormer/facelib/detection/yolov5face/utils/general.py
extracting: CodeFormer/facelib/detection/yolov5face/__init__.py
creating: CodeFormer/inputs/
creating: CodeFormer/inputs/cropped_faces/
inflating: CodeFormer/inputs/cropped_faces/0934.png
inflating: CodeFormer/inputs/cropped_faces/Solvay_conference_1927_2_16.png
inflating: CodeFormer/inputs/cropped_faces/0770.png
inflating: CodeFormer/inputs/cropped_faces/0342.png
inflating: CodeFormer/inputs/cropped_faces/0500.png
inflating: CodeFormer/inputs/cropped_faces/0368.png
inflating: CodeFormer/inputs/cropped_faces/Solvay_conference_1927_0018.png
inflating: CodeFormer/inputs/cropped_faces/0345.png
inflating: CodeFormer/inputs/cropped_faces/0720.png
inflating: CodeFormer/inputs/cropped_faces/0444.png
inflating: CodeFormer/inputs/cropped_faces/0599.png
inflating: CodeFormer/inputs/cropped_faces/0478.png
inflating: CodeFormer/inputs/cropped_faces/0763.png
inflating: CodeFormer/inputs/cropped_faces/0240.png
inflating: CodeFormer/inputs/cropped_faces/0729.png
inflating: CodeFormer/inputs/cropped_faces/0717.png
inflating: CodeFormer/inputs/cropped_faces/0412.png
inflating: CodeFormer/inputs/cropped_faces/0777.png
inflating: CodeFormer/inputs/cropped_faces/0143.png
inflating: CodeFormer/inputs/cropped_faces/0885.png
creating: CodeFormer/inputs/masked_faces/
inflating: CodeFormer/inputs/masked_faces/00664.png
inflating: CodeFormer/inputs/masked_faces/00588.png
inflating: CodeFormer/inputs/masked_faces/00108.png
inflating: CodeFormer/inputs/masked_faces/00169.png
inflating: CodeFormer/inputs/masked_faces/00105.png
creating: CodeFormer/inputs/gray_faces/
inflating: CodeFormer/inputs/gray_faces/169_John_Lennon_00.png
inflating: CodeFormer/inputs/gray_faces/111_Alexa_Chung_00.png
inflating: CodeFormer/inputs/gray_faces/161_Zac_Efron_00.png
inflating: CodeFormer/inputs/gray_faces/158_Jimmy_Fallon_00.png
inflating: CodeFormer/inputs/gray_faces/170_Marilyn_Monroe_00.png
inflating: CodeFormer/inputs/gray_faces/099_Victoria_Beckham_00.png
inflating: CodeFormer/inputs/gray_faces/067_David_Beckham_00.png
inflating: CodeFormer/inputs/gray_faces/Einstein02.png
inflating: CodeFormer/inputs/gray_faces/Hepburn01.png
inflating: CodeFormer/inputs/gray_faces/Hepburn02.png
inflating: CodeFormer/inputs/gray_faces/132_Robert_Downey_Jr_00.png
inflating: CodeFormer/inputs/gray_faces/089_Miley_Cyrus_00.png
inflating: CodeFormer/inputs/gray_faces/Einstein01.png
creating: CodeFormer/inputs/whole_imgs/
inflating: CodeFormer/inputs/whole_imgs/01.jpg
inflating: CodeFormer/inputs/whole_imgs/04.jpg
inflating: CodeFormer/inputs/whole_imgs/06.png
inflating: CodeFormer/inputs/whole_imgs/05.jpg
inflating: CodeFormer/inputs/whole_imgs/03.jpg
inflating: CodeFormer/inputs/whole_imgs/02.png
inflating: CodeFormer/inputs/whole_imgs/00.jpg
creating: CodeFormer/web-demos/
creating: CodeFormer/web-demos/replicate/
inflating: CodeFormer/web-demos/replicate/cog.yaml
inflating: CodeFormer/web-demos/replicate/predict.py
creating: CodeFormer/web-demos/hugging_face/
inflating: CodeFormer/web-demos/hugging_face/app.py
inflating: CodeFormer/inference_colorization.py
inflating: CodeFormer/inference_inpainting.py
creating: CodeFormer/.git/
creating: CodeFormer/.git/logs/
creating: CodeFormer/.git/logs/refs/
creating: CodeFormer/.git/logs/refs/remotes/
creating: CodeFormer/.git/logs/refs/remotes/origin/
inflating: CodeFormer/.git/logs/refs/remotes/origin/HEAD
creating: CodeFormer/.git/logs/refs/heads/
inflating: CodeFormer/.git/logs/refs/heads/master
inflating: CodeFormer/.git/logs/HEAD
inflating: CodeFormer/.git/config
creating: CodeFormer/.git/refs/
creating: CodeFormer/.git/refs/remotes/
creating: CodeFormer/.git/refs/remotes/origin/
extracting: CodeFormer/.git/refs/remotes/origin/HEAD
creating: CodeFormer/.git/refs/heads/
extracting: CodeFormer/.git/refs/heads/master
creating: CodeFormer/.git/refs/tags/
extracting: CodeFormer/.git/HEAD
creating: CodeFormer/.git/hooks/
inflating: CodeFormer/.git/hooks/push-to-checkout.sample
inflating: CodeFormer/.git/hooks/commit-msg.sample
inflating: CodeFormer/.git/hooks/applypatch-msg.sample
inflating: CodeFormer/.git/hooks/pre-receive.sample
inflating: CodeFormer/.git/hooks/pre-push.sample
inflating: CodeFormer/.git/hooks/fsmonitor-watchman.sample
inflating: CodeFormer/.git/hooks/post-update.sample
inflating: CodeFormer/.git/hooks/update.sample
inflating: CodeFormer/.git/hooks/pre-merge-commit.sample
inflating: CodeFormer/.git/hooks/pre-commit.sample
inflating: CodeFormer/.git/hooks/pre-applypatch.sample
inflating: CodeFormer/.git/hooks/prepare-commit-msg.sample
inflating: CodeFormer/.git/hooks/pre-rebase.sample
creating: CodeFormer/.git/info/
inflating: CodeFormer/.git/info/exclude
creating: CodeFormer/.git/objects/
creating: CodeFormer/.git/objects/pack/
inflating: CodeFormer/.git/objects/pack/pack-dcebf9a56fbde6deb0dec2536e47337200fdef01.pack
inflating: CodeFormer/.git/objects/pack/pack-dcebf9a56fbde6deb0dec2536e47337200fdef01.idx
creating: CodeFormer/.git/objects/info/
inflating: CodeFormer/.git/packed-refs
creating: CodeFormer/.git/branches/
inflating: CodeFormer/.git/index
inflating: CodeFormer/.git/description
inflating: CodeFormer/README.md
creating: CodeFormer/docs/
inflating: CodeFormer/docs/train.md
inflating: CodeFormer/docs/train_CN.md
inflating: CodeFormer/docs/history_changelog.md
creating: CodeFormer/scripts/
inflating: CodeFormer/scripts/generate_latent_gt.py
inflating: CodeFormer/scripts/download_pretrained_models_from_gdrive.py
inflating: CodeFormer/scripts/download_pretrained_models.py
inflating: CodeFormer/scripts/crop_align_face.py
inflating: CodeFormer/scripts/inference_vqgan.py
creating: CodeFormer/weights/
inflating: CodeFormer/weights/RealESRGAN_x2plus.pth
creating: CodeFormer/weights/CodeFormer/
inflating: CodeFormer/weights/CodeFormer/codeformer.pth
extracting: CodeFormer/weights/CodeFormer/.gitkeep
creating: CodeFormer/weights/dlib/
creating: CodeFormer/weights/facelib/
inflating: CodeFormer/weights/facelib/detection_Resnet50_Final.pth
inflating: CodeFormer/weights/facelib/parsing_parsenet.pth
extracting: CodeFormer/weights/facelib/.gitkeep
inflating: CodeFormer/weights/README.md
creating: CodeFormer/options/
inflating: CodeFormer/options/CodeFormer_colorization.yml
inflating: CodeFormer/options/VQGAN_512_ds32_nearest_stage1.yml
inflating: CodeFormer/options/CodeFormer_stage2.yml
inflating: CodeFormer/options/CodeFormer_inpainting.yml
inflating: CodeFormer/options/CodeFormer_stage3.yml Face detection model: retinaface_resnet50
Background upsampling: False, Face upsampling: False
[1/10] Processing: 10.jpg
Grayscale input: True
detect 0 faces
[2/10] Processing: 2.jpg
detect 14 faces
[3/10] Processing: 20.jpg
detect 0 faces
[4/10] Processing: 34.jpg
Grayscale input: True
detect 3 faces
[5/10] Processing: 4.png
detect 1 faces
[6/10] Processing: 40.jpg
Grayscale input: True
detect 1 faces
[7/10] Processing: 50.jpg
Grayscale input: True
detect 1 faces
[8/10] Processing: 54.jpg
Grayscale input: True
detect 1 faces
[9/10] Processing: 64.jpg
Grayscale input: True
detect 1 faces
[10/10] Processing: 70.jpg
Grayscale input: True
detect 9 faces
All results are saved in results/codeformer_0.5
Face detection model: retinaface_resnet50
Background upsampling: True, Face upsampling: False
[1/10] Processing: 10.jpg
Grayscale input: True
detect 0 faces
[2/10] Processing: 2.jpg
detect 14 faces
[3/10] Processing: 20.jpg
detect 0 faces
[4/10] Processing: 34.jpg
Grayscale input: True
detect 3 faces
[5/10] Processing: 4.png
detect 1 faces
[6/10] Processing: 40.jpg
Grayscale input: True
detect 1 faces
[7/10] Processing: 50.jpg
Grayscale input: True
detect 1 faces
[8/10] Processing: 54.jpg
Grayscale input: True
detect 1 faces
[9/10] Processing: 64.jpg
Grayscale input: True
detect 1 faces
[10/10] Processing: 70.jpg
Grayscale input: True
detect 9 faces
All results are saved in results/codeformer_0.5_bgup
Face detection model: retinaface_resnet50
Background upsampling: True, Face upsampling: True
[1/10] Processing: 10.jpg
Grayscale input: True
detect 0 faces
[2/10] Processing: 2.jpg
detect 14 faces
[3/10] Processing: 20.jpg
detect 0 faces
[4/10] Processing: 34.jpg
Grayscale input: True
detect 3 faces
[5/10] Processing: 4.png
detect 1 faces
[6/10] Processing: 40.jpg
Grayscale input: True
detect 1 faces
[7/10] Processing: 50.jpg
Grayscale input: True
detect 1 faces
[8/10] Processing: 54.jpg
Grayscale input: True
detect 1 faces
[9/10] Processing: 64.jpg
Grayscale input: True
detect 1 faces
[10/10] Processing: 70.jpg
Grayscale input: True
detect 9 faces
All results are saved in results/codeformer_0.5_bgup_faceup
Face detection model: retinaface_resnet50
Background upsampling: True, Face upsampling: True
[1/10] Processing: 10.jpg
Grayscale input: True
detect 0 faces
[2/10] Processing: 2.jpg
detect 14 faces
[3/10] Processing: 20.jpg
detect 0 faces
[4/10] Processing: 34.jpg
Grayscale input: True
detect 3 faces
[5/10] Processing: 4.png
detect 1 faces
[6/10] Processing: 40.jpg
Grayscale input: True
detect 1 faces
[7/10] Processing: 50.jpg
Grayscale input: True
detect 1 faces
[8/10] Processing: 54.jpg
Grayscale input: True
detect 1 faces
[9/10] Processing: 64.jpg
Grayscale input: True
detect 1 faces
[10/10] Processing: 70.jpg
Grayscale input: True
detect 9 faces
All results are saved in results/codeformer_1.0_bgup_faceup
Face detection model: retinaface_resnet50
Background upsampling: True, Face upsampling: True
[1/10] Processing: 10.jpg
Grayscale input: True
detect 0 faces
[2/10] Processing: 2.jpg
detect 14 faces
[3/10] Processing: 20.jpg
detect 0 faces
[4/10] Processing: 34.jpg
Grayscale input: True
detect 3 faces
Input is a 16-bit image
Input is a 16-bit image
[5/10] Processing: 4.png
detect 1 faces
[6/10] Processing: 40.jpg
Grayscale input: True
detect 1 faces
[7/10] Processing: 50.jpg
Grayscale input: True
detect 1 faces
[8/10] Processing: 54.jpg
Grayscale input: True
detect 1 faces
[9/10] Processing: 64.jpg
Grayscale input: True
detect 1 faces
[10/10] Processing: 70.jpg
Grayscale input: True
detect 9 faces
All results are saved in results/codeformer_0.0_bgup_faceupLooking in indexes: https://mirrors.cloud.aliyuncs.com/pypi/simple
Requirement already satisfied: modelscope in /usr/local/lib/python3.10/dist-packages (1.8.4)
Requirement already satisfied: oss2 in /usr/local/lib/python3.10/dist-packages (from modelscope) (2.18.1)
Requirement already satisfied: pandas in /usr/local/lib/python3.10/dist-packages (from modelscope) (1.5.3)
Requirement already satisfied: pyyaml in /usr/local/lib/python3.10/dist-packages (from modelscope) (6.0)
Requirement already satisfied: datasets<=2.13.0,>=2.8.0 in /usr/local/lib/python3.10/dist-packages (from modelscope) (2.11.0)
Requirement already satisfied: einops in /usr/local/lib/python3.10/dist-packages (from modelscope) (0.4.1)
Requirement already satisfied: urllib3>=1.26 in /usr/local/lib/python3.10/dist-packages (from modelscope) (1.26.15)
Requirement already satisfied: Pillow>=6.2.0 in /usr/local/lib/python3.10/dist-packages (from modelscope) (9.4.0)
Requirement already satisfied: filelock>=3.3.0 in /usr/local/lib/python3.10/dist-packages (from modelscope) (3.10.7)
Requirement already satisfied: python-dateutil>=2.1 in /usr/local/lib/python3.10/dist-packages (from modelscope) (2.8.2)
Requirement already satisfied: scipy in /usr/local/lib/python3.10/dist-packages (from modelscope) (1.10.1)
Requirement already satisfied: yapf in /usr/local/lib/python3.10/dist-packages (from modelscope) (0.32.0)
Requirement already satisfied: simplejson>=3.3.0 in /usr/local/lib/python3.10/dist-packages (from modelscope) (3.19.1)
Requirement already satisfied: setuptools in /usr/lib/python3/dist-packages (from modelscope) (59.6.0)
Requirement already satisfied: numpy in /usr/local/lib/python3.10/dist-packages (from modelscope) (1.23.3)
Requirement already satisfied: tqdm>=4.64.0 in /usr/local/lib/python3.10/dist-packages (from modelscope) (4.65.0)
Requirement already satisfied: requests>=2.25 in /usr/local/lib/python3.10/dist-packages (from modelscope) (2.25.1)
Requirement already satisfied: gast>=0.2.2 in /usr/local/lib/python3.10/dist-packages (from modelscope) (0.5.4)
Requirement already satisfied: addict in /usr/local/lib/python3.10/dist-packages (from modelscope) (2.4.0)
Requirement already satisfied: sortedcontainers>=1.5.9 in /usr/local/lib/python3.10/dist-packages (from modelscope) (2.4.0)
Requirement already satisfied: pyarrow!=9.0.0,>=6.0.0 in /usr/local/lib/python3.10/dist-packages (from modelscope) (11.0.0)
Requirement already satisfied: attrs in /usr/local/lib/python3.10/dist-packages (from modelscope) (22.2.0)
Requirement already satisfied: dill<0.3.7,>=0.3.0 in /usr/local/lib/python3.10/dist-packages (from datasets<=2.13.0,>=2.8.0->modelscope) (0.3.6)
Requirement already satisfied: huggingface-hub<1.0.0,>=0.11.0 in /usr/local/lib/python3.10/dist-packages (from datasets<=2.13.0,>=2.8.0->modelscope) (0.13.3)
Requirement already satisfied: fsspec[http]>=2021.11.1 in /usr/local/lib/python3.10/dist-packages (from datasets<=2.13.0,>=2.8.0->modelscope) (2023.3.0)
Requirement already satisfied: packaging in /usr/local/lib/python3.10/dist-packages (from datasets<=2.13.0,>=2.8.0->modelscope) (23.0)
Requirement already satisfied: aiohttp in /usr/local/lib/python3.10/dist-packages (from datasets<=2.13.0,>=2.8.0->modelscope) (3.8.4)
Requirement already satisfied: xxhash in /usr/local/lib/python3.10/dist-packages (from datasets<=2.13.0,>=2.8.0->modelscope) (3.2.0)
Requirement already satisfied: responses<0.19 in /usr/local/lib/python3.10/dist-packages (from datasets<=2.13.0,>=2.8.0->modelscope) (0.18.0)
Requirement already satisfied: multiprocess in /usr/local/lib/python3.10/dist-packages (from datasets<=2.13.0,>=2.8.0->modelscope) (0.70.14)
Requirement already satisfied: six>=1.5 in /usr/local/lib/python3.10/dist-packages (from python-dateutil>=2.1->modelscope) (1.16.0)
Requirement already satisfied: chardet<5,>=3.0.2 in /usr/local/lib/python3.10/dist-packages (from requests>=2.25->modelscope) (4.0.0)
Requirement already satisfied: idna<3,>=2.5 in /usr/local/lib/python3.10/dist-packages (from requests>=2.25->modelscope) (2.10)
Requirement already satisfied: certifi>=2017.4.17 in /usr/local/lib/python3.10/dist-packages (from requests>=2.25->modelscope) (2022.12.7)
Requirement already satisfied: pycryptodome>=3.4.7 in /usr/local/lib/python3.10/dist-packages (from oss2->modelscope) (3.17)
Requirement already satisfied: aliyun-python-sdk-core>=2.13.12 in /usr/local/lib/python3.10/dist-packages (from oss2->modelscope) (2.13.36)
Requirement already satisfied: crcmod>=1.7 in /usr/local/lib/python3.10/dist-packages (from oss2->modelscope) (1.7)
Requirement already satisfied: aliyun-python-sdk-kms>=2.4.1 in /usr/local/lib/python3.10/dist-packages (from oss2->modelscope) (2.16.1)
Requirement already satisfied: pytz>=2020.1 in /usr/local/lib/python3.10/dist-packages (from pandas->modelscope) (2023.3)
Requirement already satisfied: jmespath<1.0.0,>=0.9.3 in /usr/local/lib/python3.10/dist-packages (from aliyun-python-sdk-core>=2.13.12->oss2->modelscope) (0.10.0)
Requirement already satisfied: cryptography>=2.6.0 in /usr/local/lib/python3.10/dist-packages (from aliyun-python-sdk-core>=2.13.12->oss2->modelscope) (40.0.1)
Requirement already satisfied: aiosignal>=1.1.2 in /usr/local/lib/python3.10/dist-packages (from aiohttp->datasets<=2.13.0,>=2.8.0->modelscope) (1.3.1)
Requirement already satisfied: charset-normalizer<4.0,>=2.0 in /usr/local/lib/python3.10/dist-packages (from aiohttp->datasets<=2.13.0,>=2.8.0->modelscope) (3.1.0)
Requirement already satisfied: multidict<7.0,>=4.5 in /usr/local/lib/python3.10/dist-packages (from aiohttp->datasets<=2.13.0,>=2.8.0->modelscope) (6.0.4)
Requirement already satisfied: async-timeout<5.0,>=4.0.0a3 in /usr/local/lib/python3.10/dist-packages (from aiohttp->datasets<=2.13.0,>=2.8.0->modelscope) (4.0.2)
Requirement already satisfied: frozenlist>=1.1.1 in /usr/local/lib/python3.10/dist-packages (from aiohttp->datasets<=2.13.0,>=2.8.0->modelscope) (1.3.3)
Requirement already satisfied: yarl<2.0,>=1.0 in /usr/local/lib/python3.10/dist-packages (from aiohttp->datasets<=2.13.0,>=2.8.0->modelscope) (1.8.2)
Requirement already satisfied: typing-extensions>=3.7.4.3 in /usr/local/lib/python3.10/dist-packages (from huggingface-hub<1.0.0,>=0.11.0->datasets<=2.13.0,>=2.8.0->modelscope) (4.5.0)
Requirement already satisfied: cffi>=1.12 in /usr/local/lib/python3.10/dist-packages (from cryptography>=2.6.0->aliyun-python-sdk-core>=2.13.12->oss2->modelscope) (1.15.1)
Requirement already satisfied: pycparser in /usr/local/lib/python3.10/dist-packages (from cffi>=1.12->cryptography>=2.6.0->aliyun-python-sdk-core>=2.13.12->oss2->modelscope) (2.21)
WARNING: Running pip as the 'root' user can result in broken permissions and conflicting behaviour with the system package manager. It is recommended to use a virtual environment instead: https://pip.pypa.io/warnings/venv
[notice] A new release of pip is available: 23.0.1 -> 23.2.1
[notice] To update, run: python3 -m pip install --upgrade pip
http://pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com/aigc-data/restoration/repo/color.zip
cn-hangzhou
--2023-09-05 02:30:04-- http://pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com/aigc-data/restoration/repo/color.zip
Resolving pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com (pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com)... 100.118.28.45, 100.118.28.44, 100.118.28.50, ...
Connecting to pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com (pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com)|100.118.28.45|:80... connected.
HTTP request sent, awaiting response... 200 OK
Length: 1659285229 (1.5G) [application/zip]
Saving to: ‘color.zip’
color.zip 100%[===================>] 1.54G 12.9MB/s in 2m 10s
2023-09-05 02:32:14 (12.2 MB/s) - ‘color.zip’ saved [1659285229/1659285229]
Archive: color.zip
creating: Colorization/
creating: Colorization/.ipynb_checkpoints/
extracting: Colorization/.ipynb_checkpoints/demo-checkpoint.py
creating: Colorization/pretrain/
creating: Colorization/pretrain/cv_ddcolor_image-colorization/
inflating: Colorization/pretrain/cv_ddcolor_image-colorization/pytorch_model.pt
inflating: Colorization/pretrain/cv_ddcolor_image-colorization/.mdl
inflating: Colorization/pretrain/cv_ddcolor_image-colorization/.msc
inflating: Colorization/pretrain/cv_ddcolor_image-colorization/README.md
creating: Colorization/pretrain/cv_ddcolor_image-colorization/resources/
inflating: Colorization/pretrain/cv_ddcolor_image-colorization/resources/ddcolor_arch.jpg
inflating: Colorization/pretrain/cv_ddcolor_image-colorization/resources/demo2.jpg
inflating: Colorization/pretrain/cv_ddcolor_image-colorization/resources/demo3.jpg
inflating: Colorization/pretrain/cv_ddcolor_image-colorization/resources/demo.jpg
inflating: Colorization/pretrain/cv_ddcolor_image-colorization/configuration.json
creating: Colorization/pretrain/cv_csrnet_image-color-enhance-models/
inflating: Colorization/pretrain/cv_csrnet_image-color-enhance-models/pytorch_model.pt
extracting: Colorization/pretrain/cv_csrnet_image-color-enhance-models/.mdl
inflating: Colorization/pretrain/cv_csrnet_image-color-enhance-models/.msc
inflating: Colorization/pretrain/cv_csrnet_image-color-enhance-models/README.md
inflating: Colorization/pretrain/cv_csrnet_image-color-enhance-models/configuration.json
creating: Colorization/pretrain/cv_csrnet_image-color-enhance-models/data/
inflating: Colorization/pretrain/cv_csrnet_image-color-enhance-models/data/csrnet_1.png
inflating: Colorization/pretrain/cv_csrnet_image-color-enhance-models/data/1.png
creating: Colorization/pretrain/cv_unet_image-colorization/
inflating: Colorization/pretrain/cv_unet_image-colorization/pytorch_model.pt
extracting: Colorization/pretrain/cv_unet_image-colorization/.mdl
inflating: Colorization/pretrain/cv_unet_image-colorization/.msc
inflating: Colorization/pretrain/cv_unet_image-colorization/README.md
inflating: Colorization/pretrain/cv_unet_image-colorization/configuration.json
creating: Colorization/pretrain/cv_unet_image-colorization/description/
inflating: Colorization/pretrain/cv_unet_image-colorization/description/demo2.jpg
inflating: Colorization/pretrain/cv_unet_image-colorization/description/demo3.png
inflating: Colorization/pretrain/cv_unet_image-colorization/description/demo.jpg
inflating: Colorization/pretrain/cv_unet_image-colorization/description/deoldify_arch.png
inflating: Colorization/demo.py 2023-09-05 02:36:59,816 - modelscope - INFO - PyTorch version 1.13.1+cu117 Found.
2023-09-05 02:36:59,818 - modelscope - INFO - Loading ast index from /root/.cache/modelscope/ast_indexer
2023-09-05 02:36:59,890 - modelscope - INFO - Loading done! Current index file version is 1.8.4, with md5 80fa9349fc3e7b04fcfad511918062b1 and a total number of 902 components indexed
2023-09-05 02:37:00,925 - modelscope - INFO - initiate model from /mnt/workspace/Colorization/pretrain/cv_ddcolor_image-colorization
2023-09-05 02:37:00,926 - modelscope - INFO - initiate model from location /mnt/workspace/Colorization/pretrain/cv_ddcolor_image-colorization.
2023-09-05 02:37:00,926 - modelscope - INFO - initialize model from /mnt/workspace/Colorization/pretrain/cv_ddcolor_image-colorization
2023-09-05 02:37:05,703 - modelscope - INFO - Loading DDColor model from /mnt/workspace/Colorization/pretrain/cv_ddcolor_image-colorization/pytorch_model.pt, with param key: [params].
2023-09-05 02:37:05,906 - modelscope - INFO - load model done.
2023-09-05 02:37:05,927 - modelscope - WARNING - No preprocessor field found in cfg.
2023-09-05 02:37:05,927 - modelscope - WARNING - No val key and type key found in preprocessor domain of configuration.json file.
2023-09-05 02:37:05,927 - modelscope - WARNING - Cannot find available config to build preprocessor at mode inference, current config: {'model_dir': '/mnt/workspace/Colorization/pretrain/cv_ddcolor_image-colorization'}. trying to build by task and model information.
2023-09-05 02:37:05,927 - modelscope - WARNING - No preprocessor key ('ddcolor', 'image-colorization') found in PREPROCESSOR_MAP, skip building preprocessor.
2023-09-05 02:37:05,948 - modelscope - INFO - load model done
Total Image: 10
0/10 saved at results/DDC/64.jpg
1/10 saved at results/DDC/2.jpg
2/10 saved at results/DDC/34.jpg
3/10 saved at results/DDC/54.jpg
4/10 saved at results/DDC/40.jpg
5/10 saved at results/DDC/10.jpg
6/10 saved at results/DDC/70.jpg
7/10 saved at results/DDC/4.png
8/10 saved at results/DDC/20.jpg
9/10 saved at results/DDC/50.jpg
2023-09-05 02:37:11,906 - modelscope - INFO - PyTorch version 1.13.1+cu117 Found.
2023-09-05 02:37:11,907 - modelscope - INFO - Loading ast index from /root/.cache/modelscope/ast_indexer
2023-09-05 02:37:11,979 - modelscope - INFO - Loading done! Current index file version is 1.8.4, with md5 80fa9349fc3e7b04fcfad511918062b1 and a total number of 902 components indexed
2023-09-05 02:37:13,020 - modelscope - INFO - initiate model from /mnt/workspace/Colorization/pretrain/cv_ddcolor_image-colorization
2023-09-05 02:37:13,020 - modelscope - INFO - initiate model from location /mnt/workspace/Colorization/pretrain/cv_ddcolor_image-colorization.
2023-09-05 02:37:13,021 - modelscope - INFO - initialize model from /mnt/workspace/Colorization/pretrain/cv_ddcolor_image-colorization
2023-09-05 02:37:18,528 - modelscope - INFO - Loading DDColor model from /mnt/workspace/Colorization/pretrain/cv_ddcolor_image-colorization/pytorch_model.pt, with param key: [params].
2023-09-05 02:37:18,774 - modelscope - INFO - load model done.
2023-09-05 02:37:18,797 - modelscope - WARNING - No preprocessor field found in cfg.
2023-09-05 02:37:18,797 - modelscope - WARNING - No val key and type key found in preprocessor domain of configuration.json file.
2023-09-05 02:37:18,797 - modelscope - WARNING - Cannot find available config to build preprocessor at mode inference, current config: {'model_dir': '/mnt/workspace/Colorization/pretrain/cv_ddcolor_image-colorization'}. trying to build by task and model information.
2023-09-05 02:37:18,797 - modelscope - WARNING - No preprocessor key ('ddcolor', 'image-colorization') found in PREPROCESSOR_MAP, skip building preprocessor.
2023-09-05 02:37:18,819 - modelscope - INFO - load model done
2023-09-05 02:37:18,821 - modelscope - INFO - initiate model from /mnt/workspace/Colorization/pretrain/cv_csrnet_image-color-enhance-models
2023-09-05 02:37:18,821 - modelscope - INFO - initiate model from location /mnt/workspace/Colorization/pretrain/cv_csrnet_image-color-enhance-models.
2023-09-05 02:37:18,822 - modelscope - INFO - initialize model from /mnt/workspace/Colorization/pretrain/cv_csrnet_image-color-enhance-models
2023-09-05 02:37:19,861 - modelscope - INFO - Loading CSRNet model from /mnt/workspace/Colorization/pretrain/cv_csrnet_image-color-enhance-models/pytorch_model.pt, with param key: [params].
2023-09-05 02:37:19,863 - modelscope - INFO - load model done.
Total Image: 10
use_enhance
0/10 saved at results/DDC/enhance_64.jpg
use_enhance
1/10 saved at results/DDC/enhance_2.jpg
use_enhance
2/10 saved at results/DDC/enhance_34.jpg
use_enhance
3/10 saved at results/DDC/enhance_54.jpg
use_enhance
4/10 saved at results/DDC/enhance_40.jpg
use_enhance
5/10 saved at results/DDC/enhance_10.jpg
use_enhance
6/10 saved at results/DDC/enhance_70.jpg
use_enhance
7/10 saved at results/DDC/enhance_4.png
use_enhance
8/10 saved at results/DDC/enhance_20.jpg
use_enhance
9/10 saved at results/DDC/enhance_50.jpg
2023-09-05 02:37:25,090 - modelscope - INFO - PyTorch version 1.13.1+cu117 Found.
2023-09-05 02:37:25,091 - modelscope - INFO - Loading ast index from /root/.cache/modelscope/ast_indexer
2023-09-05 02:37:25,150 - modelscope - INFO - Loading done! Current index file version is 1.8.4, with md5 80fa9349fc3e7b04fcfad511918062b1 and a total number of 902 components indexed
2023-09-05 02:37:25,728 - modelscope - INFO - initiate model from /mnt/workspace/Colorization/pretrain/cv_unet_image-colorization
2023-09-05 02:37:25,728 - modelscope - INFO - initiate model from location /mnt/workspace/Colorization/pretrain/cv_unet_image-colorization.
2023-09-05 02:37:25,730 - modelscope - WARNING - No preprocessor field found in cfg.
2023-09-05 02:37:25,730 - modelscope - WARNING - No val key and type key found in preprocessor domain of configuration.json file.
2023-09-05 02:37:25,730 - modelscope - WARNING - Cannot find available config to build preprocessor at mode inference, current config: {'model_dir': '/mnt/workspace/Colorization/pretrain/cv_unet_image-colorization'}. trying to build by task and model information.
2023-09-05 02:37:25,730 - modelscope - WARNING - Find task: image-colorization, model type: None. Insufficient information to build preprocessor, skip building preprocessor
/usr/local/lib/python3.10/dist-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
warnings.warn(
/usr/local/lib/python3.10/dist-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=ResNet101_Weights.IMAGENET1K_V1`. You can also use `weights=ResNet101_Weights.DEFAULT` to get the most up-to-date weights.
warnings.warn(msg)
Downloading: "https://download.pytorch.org/models/resnet101-63fe2227.pth" to /root/.cache/torch/hub/checkpoints/resnet101-63fe2227.pth
100%|████████████████████████████████████████| 171M/171M [00:08<00:00, 21.7MB/s]
2023-09-05 02:37:40,123 - modelscope - INFO - load model done
Total Image: 10
0/10 saved at results/DeOldify/64.jpg
1/10 saved at results/DeOldify/2.jpg
2/10 saved at results/DeOldify/34.jpg
3/10 saved at results/DeOldify/54.jpg
4/10 saved at results/DeOldify/40.jpg
5/10 saved at results/DeOldify/10.jpg
6/10 saved at results/DeOldify/70.jpg
7/10 saved at results/DeOldify/4.png
8/10 saved at results/DeOldify/20.jpg
9/10 saved at results/DeOldify/50.jpg
2023-09-05 02:37:44,830 - modelscope - INFO - PyTorch version 1.13.1+cu117 Found.
2023-09-05 02:37:44,831 - modelscope - INFO - Loading ast index from /root/.cache/modelscope/ast_indexer
2023-09-05 02:37:44,898 - modelscope - INFO - Loading done! Current index file version is 1.8.4, with md5 80fa9349fc3e7b04fcfad511918062b1 and a total number of 902 components indexed
2023-09-05 02:37:45,464 - modelscope - INFO - initiate model from /mnt/workspace/Colorization/pretrain/cv_unet_image-colorization
2023-09-05 02:37:45,464 - modelscope - INFO - initiate model from location /mnt/workspace/Colorization/pretrain/cv_unet_image-colorization.
2023-09-05 02:37:45,466 - modelscope - WARNING - No preprocessor field found in cfg.
2023-09-05 02:37:45,466 - modelscope - WARNING - No val key and type key found in preprocessor domain of configuration.json file.
2023-09-05 02:37:45,466 - modelscope - WARNING - Cannot find available config to build preprocessor at mode inference, current config: {'model_dir': '/mnt/workspace/Colorization/pretrain/cv_unet_image-colorization'}. trying to build by task and model information.
2023-09-05 02:37:45,466 - modelscope - WARNING - Find task: image-colorization, model type: None. Insufficient information to build preprocessor, skip building preprocessor
/usr/local/lib/python3.10/dist-packages/torchvision/models/_utils.py:208: UserWarning: The parameter 'pretrained' is deprecated since 0.13 and may be removed in the future, please use 'weights' instead.
warnings.warn(
/usr/local/lib/python3.10/dist-packages/torchvision/models/_utils.py:223: UserWarning: Arguments other than a weight enum or `None` for 'weights' are deprecated since 0.13 and may be removed in the future. The current behavior is equivalent to passing `weights=ResNet101_Weights.IMAGENET1K_V1`. You can also use `weights=ResNet101_Weights.DEFAULT` to get the most up-to-date weights.
warnings.warn(msg)
2023-09-05 02:37:49,758 - modelscope - INFO - load model done
2023-09-05 02:37:49,760 - modelscope - INFO - initiate model from /mnt/workspace/Colorization/pretrain/cv_csrnet_image-color-enhance-models
2023-09-05 02:37:49,760 - modelscope - INFO - initiate model from location /mnt/workspace/Colorization/pretrain/cv_csrnet_image-color-enhance-models.
2023-09-05 02:37:49,761 - modelscope - INFO - initialize model from /mnt/workspace/Colorization/pretrain/cv_csrnet_image-color-enhance-models
2023-09-05 02:37:49,766 - modelscope - INFO - Loading CSRNet model from /mnt/workspace/Colorization/pretrain/cv_csrnet_image-color-enhance-models/pytorch_model.pt, with param key: [params].
2023-09-05 02:37:49,768 - modelscope - INFO - load model done.
Total Image: 10
use_enhance
0/10 saved at results/DeOldify/enhance_64.jpg
use_enhance
1/10 saved at results/DeOldify/enhance_2.jpg
use_enhance
2/10 saved at results/DeOldify/enhance_34.jpg
use_enhance
3/10 saved at results/DeOldify/enhance_54.jpg
use_enhance
4/10 saved at results/DeOldify/enhance_40.jpg
use_enhance
5/10 saved at results/DeOldify/enhance_10.jpg
use_enhance
6/10 saved at results/DeOldify/enhance_70.jpg
use_enhance
7/10 saved at results/DeOldify/enhance_4.png
use_enhance
8/10 saved at results/DeOldify/enhance_20.jpg
use_enhance
9/10 saved at results/DeOldify/enhance_50.jpghttp://pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com/aigc-data/restoration/repo/sam_unicolor.zip
cn-hangzhou
--2023-09-05 02:44:17-- http://pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com/aigc-data/restoration/repo/sam_unicolor.zip
Resolving pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com (pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com)... 100.118.28.44, 100.118.28.45, 100.118.28.49, ...
Connecting to pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com (pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com)|100.118.28.44|:80... connected.
HTTP request sent, awaiting response... 200 OK
Length: 9102984978 (8.5G) [application/zip]
Saving to: ‘sam_unicolor.zip’
sam_unicolor.zip 100%[===================>] 8.48G 17.6MB/s in 8m 33s
2023-09-05 02:52:51 (16.9 MB/s) - ‘sam_unicolor.zip’ saved [9102984978/9102984978]
Archive: sam_unicolor.zip
creating: sample/
inflating: sample/utils_func.py
creating: sample/.ipynb_checkpoints/
inflating: sample/.ipynb_checkpoints/sample-checkpoint.ipynb
creating: sample/SAM/
inflating: sample/SAM/CONTRIBUTING.md
creating: sample/SAM/demo/
creating: sample/SAM/demo/src/
creating: sample/SAM/demo/src/components/
inflating: sample/SAM/demo/src/components/Stage.tsx
creating: sample/SAM/demo/src/components/hooks/
inflating: sample/SAM/demo/src/components/hooks/createContext.tsx
inflating: sample/SAM/demo/src/components/hooks/context.tsx
inflating: sample/SAM/demo/src/components/Tool.tsx
creating: sample/SAM/demo/src/components/helpers/
inflating: sample/SAM/demo/src/components/helpers/maskUtils.tsx
inflating: sample/SAM/demo/src/components/helpers/scaleHelper.tsx
inflating: sample/SAM/demo/src/components/helpers/Interfaces.tsx
inflating: sample/SAM/demo/src/components/helpers/onnxModelAPI.tsx
creating: sample/SAM/demo/src/assets/
creating: sample/SAM/demo/src/assets/data/
inflating: sample/SAM/demo/src/assets/data/dogs.jpg
creating: sample/SAM/demo/src/assets/scss/
inflating: sample/SAM/demo/src/assets/scss/App.scss
inflating: sample/SAM/demo/src/assets/index.html
inflating: sample/SAM/demo/src/index.tsx
inflating: sample/SAM/demo/src/App.tsx
creating: sample/SAM/demo/configs/
creating: sample/SAM/demo/configs/webpack/
inflating: sample/SAM/demo/configs/webpack/common.js
inflating: sample/SAM/demo/configs/webpack/prod.js
inflating: sample/SAM/demo/configs/webpack/dev.js
inflating: sample/SAM/demo/tsconfig.json
inflating: sample/SAM/demo/tailwind.config.js
inflating: sample/SAM/demo/README.md
inflating: sample/SAM/demo/postcss.config.js
inflating: sample/SAM/demo/package.json
creating: sample/SAM/assets/
inflating: sample/SAM/assets/masks2.jpg
inflating: sample/SAM/assets/minidemo.gif
inflating: sample/SAM/assets/model_diagram.png
inflating: sample/SAM/assets/notebook1.png
inflating: sample/SAM/assets/masks1.png
inflating: sample/SAM/assets/notebook2.png
inflating: sample/SAM/nohup.out
inflating: sample/SAM/setup.cfg
inflating: sample/SAM/.gitignore
inflating: sample/SAM/CODE_OF_CONDUCT.md
creating: sample/SAM/notebooks/
inflating: sample/SAM/notebooks/part.png
inflating: sample/SAM/notebooks/onnx_model_example.ipynb
inflating: sample/SAM/notebooks/save_0664.png
creating: sample/SAM/notebooks/.ipynb_checkpoints/
inflating: sample/SAM/notebooks/.ipynb_checkpoints/automatic_mask_generator_example-checkpoint.ipynb
inflating: sample/SAM/notebooks/.ipynb_checkpoints/predictor_example-checkpoint.ipynb
inflating: sample/SAM/notebooks/save_part.png
inflating: sample/SAM/notebooks/save.png
inflating: sample/SAM/notebooks/save_0656.png
inflating: sample/SAM/notebooks/save_0660.png
inflating: sample/SAM/notebooks/predictor_example.ipynb
inflating: sample/SAM/notebooks/part_1.png
creating: sample/SAM/notebooks/images/
inflating: sample/SAM/notebooks/images/truck.jpg
inflating: sample/SAM/notebooks/images/000.png
inflating: sample/SAM/notebooks/images/dog.jpg
inflating: sample/SAM/notebooks/images/groceries.jpg
inflating: sample/SAM/notebooks/automatic_mask_generator_example.ipynb
inflating: sample/SAM/README.md
inflating: sample/SAM/LICENSE
creating: sample/SAM/segment_anything/
inflating: sample/SAM/segment_anything/__init__.py
inflating: sample/SAM/segment_anything/predictor.py
inflating: sample/SAM/segment_anything/automatic_mask_generator.py
creating: sample/SAM/segment_anything/utils/
inflating: sample/SAM/segment_anything/utils/__init__.py
inflating: sample/SAM/segment_anything/utils/amg.py
inflating: sample/SAM/segment_anything/utils/onnx.py
inflating: sample/SAM/segment_anything/utils/transforms.py
creating: sample/SAM/segment_anything/utils/__pycache__/
inflating: sample/SAM/segment_anything/utils/__pycache__/__init__.cpython-38.pyc
inflating: sample/SAM/segment_anything/utils/__pycache__/transforms.cpython-38.pyc
inflating: sample/SAM/segment_anything/utils/__pycache__/amg.cpython-38.pyc
creating: sample/SAM/segment_anything/modeling/
inflating: sample/SAM/segment_anything/modeling/__init__.py
inflating: sample/SAM/segment_anything/modeling/mask_decoder.py
inflating: sample/SAM/segment_anything/modeling/prompt_encoder.py
inflating: sample/SAM/segment_anything/modeling/transformer.py
inflating: sample/SAM/segment_anything/modeling/common.py
inflating: sample/SAM/segment_anything/modeling/image_encoder.py
inflating: sample/SAM/segment_anything/modeling/sam.py
creating: sample/SAM/segment_anything/modeling/__pycache__/
inflating: sample/SAM/segment_anything/modeling/__pycache__/__init__.cpython-38.pyc
inflating: sample/SAM/segment_anything/modeling/__pycache__/sam.cpython-38.pyc
inflating: sample/SAM/segment_anything/modeling/__pycache__/image_encoder.cpython-38.pyc
inflating: sample/SAM/segment_anything/modeling/__pycache__/common.cpython-38.pyc
inflating: sample/SAM/segment_anything/modeling/__pycache__/transformer.cpython-38.pyc
inflating: sample/SAM/segment_anything/modeling/__pycache__/mask_decoder.cpython-38.pyc
inflating: sample/SAM/segment_anything/modeling/__pycache__/prompt_encoder.cpython-38.pyc
inflating: sample/SAM/segment_anything/build_sam.py
creating: sample/SAM/segment_anything/__pycache__/
inflating: sample/SAM/segment_anything/__pycache__/__init__.cpython-38.pyc
inflating: sample/SAM/segment_anything/__pycache__/automatic_mask_generator.cpython-38.pyc
inflating: sample/SAM/segment_anything/__pycache__/predictor.cpython-38.pyc
inflating: sample/SAM/segment_anything/__pycache__/build_sam.cpython-38.pyc
inflating: sample/SAM/linter.sh
inflating: sample/SAM/setup.py
creating: sample/SAM/.git/
creating: sample/SAM/.git/logs/
inflating: sample/SAM/.git/logs/HEAD
creating: sample/SAM/.git/logs/refs/
creating: sample/SAM/.git/logs/refs/remotes/
creating: sample/SAM/.git/logs/refs/remotes/origin/
inflating: sample/SAM/.git/logs/refs/remotes/origin/HEAD
creating: sample/SAM/.git/logs/refs/heads/
inflating: sample/SAM/.git/logs/refs/heads/main
creating: sample/SAM/.git/objects/
creating: sample/SAM/.git/objects/info/
creating: sample/SAM/.git/objects/pack/
inflating: sample/SAM/.git/objects/pack/pack-6bddc3bea3b5846071a7a94ed6811aed34aa66d9.pack
inflating: sample/SAM/.git/objects/pack/pack-6bddc3bea3b5846071a7a94ed6811aed34aa66d9.idx
inflating: sample/SAM/.git/packed-refs
creating: sample/SAM/.git/hooks/
inflating: sample/SAM/.git/hooks/pre-commit.sample
inflating: sample/SAM/.git/hooks/pre-rebase.sample
inflating: sample/SAM/.git/hooks/post-update.sample
inflating: sample/SAM/.git/hooks/update.sample
inflating: sample/SAM/.git/hooks/pre-merge-commit.sample
inflating: sample/SAM/.git/hooks/pre-push.sample
inflating: sample/SAM/.git/hooks/fsmonitor-watchman.sample
inflating: sample/SAM/.git/hooks/prepare-commit-msg.sample
inflating: sample/SAM/.git/hooks/applypatch-msg.sample
inflating: sample/SAM/.git/hooks/pre-receive.sample
inflating: sample/SAM/.git/hooks/commit-msg.sample
inflating: sample/SAM/.git/hooks/pre-applypatch.sample
inflating: sample/SAM/.git/config
inflating: sample/SAM/.git/description
creating: sample/SAM/.git/branches/
creating: sample/SAM/.git/info/
inflating: sample/SAM/.git/info/exclude
extracting: sample/SAM/.git/HEAD
creating: sample/SAM/.git/refs/
creating: sample/SAM/.git/refs/remotes/
creating: sample/SAM/.git/refs/remotes/origin/
extracting: sample/SAM/.git/refs/remotes/origin/HEAD
creating: sample/SAM/.git/refs/heads/
extracting: sample/SAM/.git/refs/heads/main
creating: sample/SAM/.git/refs/tags/
inflating: sample/SAM/.git/index
inflating: sample/SAM/.flake8
creating: sample/SAM/scripts/
inflating: sample/SAM/scripts/amg.py
inflating: sample/SAM/scripts/export_onnx_model.py
inflating: sample/color_table.yaml
creating: sample/ImageMatch/
creating: sample/ImageMatch/data/
inflating: sample/ImageMatch/data/vgg19_gray.pth
inflating: sample/ImageMatch/data/vgg19_conv.pth
creating: sample/ImageMatch/lib/
extracting: sample/ImageMatch/lib/__init__.py
inflating: sample/ImageMatch/lib/functional.py
inflating: sample/ImageMatch/lib/TestTransforms.py
inflating: sample/ImageMatch/lib/TrainTransforms.py
inflating: sample/ImageMatch/lib/FeatVGG.py
inflating: sample/ImageMatch/lib/videoloader.py
inflating: sample/ImageMatch/lib/videoloader_imagenet.py
creating: sample/ImageMatch/lib/.vscode/
extracting: sample/ImageMatch/lib/.vscode/settings.json
creating: sample/ImageMatch/lib/__pycache__/
inflating: sample/ImageMatch/lib/__pycache__/__init__.cpython-38.pyc
inflating: sample/ImageMatch/lib/__pycache__/TestTransforms.cpython-38.pyc
inflating: sample/ImageMatch/lib/__pycache__/functional.cpython-38.pyc
inflating: sample/ImageMatch/lib/VGGFeatureLoss.py
creating: sample/ImageMatch/utils/
extracting: sample/ImageMatch/utils/__init__.py
inflating: sample/ImageMatch/utils/util_distortion.py
inflating: sample/ImageMatch/utils/vgg_util.py
inflating: sample/ImageMatch/utils/flowlib.py
inflating: sample/ImageMatch/utils/util_tensorboard.py
inflating: sample/ImageMatch/utils/warping.py
inflating: sample/ImageMatch/utils/util.py
creating: sample/ImageMatch/utils/__pycache__/
inflating: sample/ImageMatch/utils/__pycache__/__init__.cpython-38.pyc
inflating: sample/ImageMatch/utils/__pycache__/util_distortion.cpython-38.pyc
inflating: sample/ImageMatch/utils/__pycache__/util.cpython-38.pyc
inflating: sample/ImageMatch/utils/tb_image_recorder.py
inflating: sample/ImageMatch/warp.py
creating: sample/ImageMatch/tensorboardX/
inflating: sample/ImageMatch/tensorboardX/__init__.py
creating: sample/ImageMatch/tensorboardX/src/
extracting: sample/ImageMatch/tensorboardX/src/__init__.py
inflating: sample/ImageMatch/tensorboardX/src/node_def_pb2.py
inflating: sample/ImageMatch/tensorboardX/src/summary_pb2.py
inflating: sample/ImageMatch/tensorboardX/src/versions_pb2.py
inflating: sample/ImageMatch/tensorboardX/src/resource_handle_pb2.py
inflating: sample/ImageMatch/tensorboardX/src/tensor_shape_pb2.py
inflating: sample/ImageMatch/tensorboardX/src/types_pb2.py
inflating: sample/ImageMatch/tensorboardX/src/attr_value_pb2.py
inflating: sample/ImageMatch/tensorboardX/src/plugin_pr_curve_pb2.py
inflating: sample/ImageMatch/tensorboardX/src/event_pb2.py
inflating: sample/ImageMatch/tensorboardX/src/tensor_pb2.py
inflating: sample/ImageMatch/tensorboardX/src/graph_pb2.py
inflating: sample/ImageMatch/tensorboardX/graph.py
inflating: sample/ImageMatch/tensorboardX/crc32c.py
inflating: sample/ImageMatch/tensorboardX/x2num.py
inflating: sample/ImageMatch/tensorboardX/summary.py
inflating: sample/ImageMatch/tensorboardX/record_writer.py
inflating: sample/ImageMatch/tensorboardX/event_file_writer.py
inflating: sample/ImageMatch/tensorboardX/writer.py
inflating: sample/ImageMatch/tensorboardX/graph_onnx.py
inflating: sample/ImageMatch/tensorboardX/embedding.py
creating: sample/ImageMatch/models/
inflating: sample/ImageMatch/models/ContextualLoss.py
inflating: sample/ImageMatch/models/NonlocalNet.py
inflating: sample/ImageMatch/models/ColorVidNet.py
inflating: sample/ImageMatch/models/FrameColor.py
inflating: sample/ImageMatch/models/spectral_normalization.py
inflating: sample/ImageMatch/models/GAN_models.py
creating: sample/ImageMatch/models/__pycache__/
inflating: sample/ImageMatch/models/__pycache__/FrameColor.cpython-38.pyc
inflating: sample/ImageMatch/models/__pycache__/NonlocalNet.cpython-38.pyc
inflating: sample/ImageMatch/models/__pycache__/vgg19_gray.cpython-38.pyc
inflating: sample/ImageMatch/models/__pycache__/ColorVidNet.cpython-38.pyc
inflating: sample/ImageMatch/models/vgg19_gray.py
creating: sample/ImageMatch/checkpoints/
creating: sample/ImageMatch/checkpoints/video_moredata_l1/
inflating: sample/ImageMatch/checkpoints/video_moredata_l1/nonlocal_net_iter_76000.pth
inflating: sample/ImageMatch/checkpoints/video_moredata_l1/colornet_iter_76000.pth
inflating: sample/ImageMatch/checkpoints/video_moredata_l1/discriminator_iter_76000.pth
creating: sample/ImageMatch/__pycache__/
inflating: sample/ImageMatch/__pycache__/warp.cpython-38.pyc
inflating: sample/colorizer.py
inflating: sample/sample.ipynb
creating: sample/images/
inflating: sample/images/12.jpg
inflating: sample/images/1_exp.jpg
inflating: sample/images/1.jpg
inflating: sample/images/10.jpg
inflating: sample/sam_vit_h_4b8939.pth
creating: sample/unsorted_codes/
inflating: sample/unsorted_codes/clip_score.py
inflating: sample/unsorted_codes/test_heatmap.py
inflating: sample/unsorted_codes/ideep_stroke_file.py
inflating: sample/unsorted_codes/metrics.ipynb
inflating: sample/unsorted_codes/generate_strokes.py
inflating: sample/unsorted_codes/html_images.py
inflating: sample/unsorted_codes/ablation.ipynb
inflating: sample/unsorted_codes/experiment.ipynb
inflating: sample/unsorted_codes/ideep_stroke.py
inflating: sample/unsorted_codes/sample_exemplar.py
inflating: sample/unsorted_codes/sample.ipynb
inflating: sample/unsorted_codes/clip_segment.py
inflating: sample/unsorted_codes/sample_func.py
inflating: sample/unsorted_codes/sample_uncond.py
creating: sample/__pycache__/
inflating: sample/__pycache__/colorizer.cpython-38.pyc
inflating: sample/__pycache__/utils_func.cpython-38.pyc
creating: unicolor/
creating: unicolor/sample/
inflating: unicolor/sample/color_table.yaml
inflating: unicolor/sample/sam_vit_h_4b8939.pth
creating: unicolor/sample/unsorted_codes/
inflating: unicolor/sample/unsorted_codes/experiment.ipynb
inflating: unicolor/sample/unsorted_codes/sample.ipynb
inflating: unicolor/sample/unsorted_codes/html_images.py
inflating: unicolor/sample/unsorted_codes/ideep_stroke.py
inflating: unicolor/sample/unsorted_codes/sample_uncond.py
inflating: unicolor/sample/unsorted_codes/metrics.ipynb
inflating: unicolor/sample/unsorted_codes/test_heatmap.py
inflating: unicolor/sample/unsorted_codes/clip_score.py
inflating: unicolor/sample/unsorted_codes/ideep_stroke_file.py
inflating: unicolor/sample/unsorted_codes/ablation.ipynb
inflating: unicolor/sample/unsorted_codes/sample_exemplar.py
inflating: unicolor/sample/unsorted_codes/sample_func.py
inflating: unicolor/sample/unsorted_codes/clip_segment.py
inflating: unicolor/sample/unsorted_codes/generate_strokes.py
inflating: unicolor/sample/sample.ipynb
creating: unicolor/sample/.ipynb_checkpoints/
inflating: unicolor/sample/.ipynb_checkpoints/utils_func-checkpoint.py
inflating: unicolor/sample/.ipynb_checkpoints/colorizer-checkpoint.py
inflating: unicolor/sample/.ipynb_checkpoints/sample-checkpoint.ipynb
creating: unicolor/sample/ImageMatch/
creating: unicolor/sample/ImageMatch/tensorboardX/
inflating: unicolor/sample/ImageMatch/tensorboardX/writer.py
inflating: unicolor/sample/ImageMatch/tensorboardX/graph_onnx.py
inflating: unicolor/sample/ImageMatch/tensorboardX/x2num.py
creating: unicolor/sample/ImageMatch/tensorboardX/src/
inflating: unicolor/sample/ImageMatch/tensorboardX/src/plugin_pr_curve_pb2.py
inflating: unicolor/sample/ImageMatch/tensorboardX/src/graph_pb2.py
inflating: unicolor/sample/ImageMatch/tensorboardX/src/resource_handle_pb2.py
inflating: unicolor/sample/ImageMatch/tensorboardX/src/tensor_shape_pb2.py
inflating: unicolor/sample/ImageMatch/tensorboardX/src/summary_pb2.py
inflating: unicolor/sample/ImageMatch/tensorboardX/src/types_pb2.py
inflating: unicolor/sample/ImageMatch/tensorboardX/src/versions_pb2.py
inflating: unicolor/sample/ImageMatch/tensorboardX/src/event_pb2.py
inflating: unicolor/sample/ImageMatch/tensorboardX/src/node_def_pb2.py
inflating: unicolor/sample/ImageMatch/tensorboardX/src/attr_value_pb2.py
extracting: unicolor/sample/ImageMatch/tensorboardX/src/__init__.py
inflating: unicolor/sample/ImageMatch/tensorboardX/src/tensor_pb2.py
inflating: unicolor/sample/ImageMatch/tensorboardX/record_writer.py
inflating: unicolor/sample/ImageMatch/tensorboardX/crc32c.py
inflating: unicolor/sample/ImageMatch/tensorboardX/embedding.py
inflating: unicolor/sample/ImageMatch/tensorboardX/graph.py
inflating: unicolor/sample/ImageMatch/tensorboardX/event_file_writer.py
inflating: unicolor/sample/ImageMatch/tensorboardX/__init__.py
inflating: unicolor/sample/ImageMatch/tensorboardX/summary.py
creating: unicolor/sample/ImageMatch/checkpoints/
creating: unicolor/sample/ImageMatch/checkpoints/video_moredata_l1/
inflating: unicolor/sample/ImageMatch/checkpoints/video_moredata_l1/nonlocal_net_iter_76000.pth
inflating: unicolor/sample/ImageMatch/checkpoints/video_moredata_l1/discriminator_iter_76000.pth
inflating: unicolor/sample/ImageMatch/checkpoints/video_moredata_l1/colornet_iter_76000.pth
creating: unicolor/sample/ImageMatch/models/
inflating: unicolor/sample/ImageMatch/models/ColorVidNet.py
inflating: unicolor/sample/ImageMatch/models/NonlocalNet.py
inflating: unicolor/sample/ImageMatch/models/FrameColor.py
inflating: unicolor/sample/ImageMatch/models/spectral_normalization.py
inflating: unicolor/sample/ImageMatch/models/ContextualLoss.py
creating: unicolor/sample/ImageMatch/models/__pycache__/
inflating: unicolor/sample/ImageMatch/models/__pycache__/FrameColor.cpython-310.pyc
inflating: unicolor/sample/ImageMatch/models/__pycache__/ColorVidNet.cpython-38.pyc
inflating: unicolor/sample/ImageMatch/models/__pycache__/vgg19_gray.cpython-310.pyc
inflating: unicolor/sample/ImageMatch/models/__pycache__/FrameColor.cpython-38.pyc
inflating: unicolor/sample/ImageMatch/models/__pycache__/ColorVidNet.cpython-310.pyc
inflating: unicolor/sample/ImageMatch/models/__pycache__/NonlocalNet.cpython-310.pyc
inflating: unicolor/sample/ImageMatch/models/__pycache__/NonlocalNet.cpython-38.pyc
inflating: unicolor/sample/ImageMatch/models/__pycache__/vgg19_gray.cpython-38.pyc
inflating: unicolor/sample/ImageMatch/models/GAN_models.py
inflating: unicolor/sample/ImageMatch/models/vgg19_gray.py
creating: unicolor/sample/ImageMatch/utils/
inflating: unicolor/sample/ImageMatch/utils/vgg_util.py
inflating: unicolor/sample/ImageMatch/utils/flowlib.py
inflating: unicolor/sample/ImageMatch/utils/util.py
inflating: unicolor/sample/ImageMatch/utils/util_tensorboard.py
inflating: unicolor/sample/ImageMatch/utils/warping.py
inflating: unicolor/sample/ImageMatch/utils/util_distortion.py
creating: unicolor/sample/ImageMatch/utils/__pycache__/
inflating: unicolor/sample/ImageMatch/utils/__pycache__/util.cpython-310.pyc
inflating: unicolor/sample/ImageMatch/utils/__pycache__/__init__.cpython-310.pyc
inflating: unicolor/sample/ImageMatch/utils/__pycache__/util.cpython-38.pyc
inflating: unicolor/sample/ImageMatch/utils/__pycache__/__init__.cpython-38.pyc
inflating: unicolor/sample/ImageMatch/utils/__pycache__/util_distortion.cpython-38.pyc
inflating: unicolor/sample/ImageMatch/utils/__pycache__/util_distortion.cpython-310.pyc
extracting: unicolor/sample/ImageMatch/utils/__init__.py
inflating: unicolor/sample/ImageMatch/utils/tb_image_recorder.py
creating: unicolor/sample/ImageMatch/__pycache__/
inflating: unicolor/sample/ImageMatch/__pycache__/warp.cpython-38.pyc
inflating: unicolor/sample/ImageMatch/__pycache__/warp.cpython-310.pyc
creating: unicolor/sample/ImageMatch/data/
inflating: unicolor/sample/ImageMatch/data/vgg19_gray.pth
inflating: unicolor/sample/ImageMatch/data/vgg19_conv.pth
creating: unicolor/sample/ImageMatch/lib/
inflating: unicolor/sample/ImageMatch/lib/FeatVGG.py
inflating: unicolor/sample/ImageMatch/lib/TestTransforms.py
inflating: unicolor/sample/ImageMatch/lib/TrainTransforms.py
inflating: unicolor/sample/ImageMatch/lib/functional.py
inflating: unicolor/sample/ImageMatch/lib/VGGFeatureLoss.py
creating: unicolor/sample/ImageMatch/lib/__pycache__/
inflating: unicolor/sample/ImageMatch/lib/__pycache__/__init__.cpython-310.pyc
inflating: unicolor/sample/ImageMatch/lib/__pycache__/__init__.cpython-38.pyc
inflating: unicolor/sample/ImageMatch/lib/__pycache__/TestTransforms.cpython-310.pyc
inflating: unicolor/sample/ImageMatch/lib/__pycache__/functional.cpython-310.pyc
inflating: unicolor/sample/ImageMatch/lib/__pycache__/TestTransforms.cpython-38.pyc
inflating: unicolor/sample/ImageMatch/lib/__pycache__/functional.cpython-38.pyc
inflating: unicolor/sample/ImageMatch/lib/videoloader_imagenet.py
creating: unicolor/sample/ImageMatch/lib/.vscode/
extracting: unicolor/sample/ImageMatch/lib/.vscode/settings.json
extracting: unicolor/sample/ImageMatch/lib/__init__.py
inflating: unicolor/sample/ImageMatch/lib/videoloader.py
inflating: unicolor/sample/ImageMatch/warp.py
creating: unicolor/sample/SAM/
inflating: unicolor/sample/SAM/LICENSE
creating: unicolor/sample/SAM/segment_anything/
inflating: unicolor/sample/SAM/segment_anything/predictor.py
creating: unicolor/sample/SAM/segment_anything/.ipynb_checkpoints/
inflating: unicolor/sample/SAM/segment_anything/.ipynb_checkpoints/__init__-checkpoint.py
inflating: unicolor/sample/SAM/segment_anything/.ipynb_checkpoints/predictor-checkpoint.py
inflating: unicolor/sample/SAM/segment_anything/.ipynb_checkpoints/build_sam-checkpoint.py
inflating: unicolor/sample/SAM/segment_anything/.ipynb_checkpoints/automatic_mask_generator-checkpoint.py
inflating: unicolor/sample/SAM/segment_anything/amg.py
inflating: unicolor/sample/SAM/segment_anything/automatic_mask_generator.py
creating: unicolor/sample/SAM/segment_anything/modeling/
inflating: unicolor/sample/SAM/segment_anything/modeling/prompt_encoder.py
inflating: unicolor/sample/SAM/segment_anything/modeling/image_encoder.py
inflating: unicolor/sample/SAM/segment_anything/modeling/mask_decoder.py
inflating: unicolor/sample/SAM/segment_anything/modeling/transformer.py
inflating: unicolor/sample/SAM/segment_anything/modeling/sam.py
creating: unicolor/sample/SAM/segment_anything/modeling/__pycache__/
inflating: unicolor/sample/SAM/segment_anything/modeling/__pycache__/transformer.cpython-38.pyc
inflating: unicolor/sample/SAM/segment_anything/modeling/__pycache__/common.cpython-310.pyc
inflating: unicolor/sample/SAM/segment_anything/modeling/__pycache__/image_encoder.cpython-38.pyc
inflating: unicolor/sample/SAM/segment_anything/modeling/__pycache__/sam.cpython-38.pyc
inflating: unicolor/sample/SAM/segment_anything/modeling/__pycache__/__init__.cpython-310.pyc
inflating: unicolor/sample/SAM/segment_anything/modeling/__pycache__/__init__.cpython-38.pyc
inflating: unicolor/sample/SAM/segment_anything/modeling/__pycache__/mask_decoder.cpython-38.pyc
inflating: unicolor/sample/SAM/segment_anything/modeling/__pycache__/image_encoder.cpython-310.pyc
inflating: unicolor/sample/SAM/segment_anything/modeling/__pycache__/sam.cpython-310.pyc
inflating: unicolor/sample/SAM/segment_anything/modeling/__pycache__/prompt_encoder.cpython-38.pyc
inflating: unicolor/sample/SAM/segment_anything/modeling/__pycache__/common.cpython-38.pyc
inflating: unicolor/sample/SAM/segment_anything/modeling/__pycache__/mask_decoder.cpython-310.pyc
inflating: unicolor/sample/SAM/segment_anything/modeling/__pycache__/prompt_encoder.cpython-310.pyc
inflating: unicolor/sample/SAM/segment_anything/modeling/__pycache__/transformer.cpython-310.pyc
inflating: unicolor/sample/SAM/segment_anything/modeling/__init__.py
inflating: unicolor/sample/SAM/segment_anything/modeling/common.py
creating: unicolor/sample/SAM/segment_anything/utils/
creating: unicolor/sample/SAM/segment_anything/utils/.ipynb_checkpoints/
inflating: unicolor/sample/SAM/segment_anything/utils/.ipynb_checkpoints/transforms-checkpoint.py
inflating: unicolor/sample/SAM/segment_anything/utils/.ipynb_checkpoints/amg-checkpoint.py
inflating: unicolor/sample/SAM/segment_anything/utils/amg.py
inflating: unicolor/sample/SAM/segment_anything/utils/transforms.py
creating: unicolor/sample/SAM/segment_anything/utils/__pycache__/
inflating: unicolor/sample/SAM/segment_anything/utils/__pycache__/__init__.cpython-38.pyc
inflating: unicolor/sample/SAM/segment_anything/utils/__pycache__/amg.cpython-38.pyc
inflating: unicolor/sample/SAM/segment_anything/utils/__pycache__/transforms.cpython-38.pyc
inflating: unicolor/sample/SAM/segment_anything/utils/__init__.py
inflating: unicolor/sample/SAM/segment_anything/utils/onnx.py
inflating: unicolor/sample/SAM/segment_anything/transforms.py
inflating: unicolor/sample/SAM/segment_anything/build_sam.py
creating: unicolor/sample/SAM/segment_anything/__pycache__/
inflating: unicolor/sample/SAM/segment_anything/__pycache__/amg.cpython-310.pyc
inflating: unicolor/sample/SAM/segment_anything/__pycache__/__init__.cpython-310.pyc
inflating: unicolor/sample/SAM/segment_anything/__pycache__/build_sam.cpython-38.pyc
inflating: unicolor/sample/SAM/segment_anything/__pycache__/__init__.cpython-38.pyc
inflating: unicolor/sample/SAM/segment_anything/__pycache__/predictor.cpython-38.pyc
inflating: unicolor/sample/SAM/segment_anything/__pycache__/automatic_mask_generator.cpython-310.pyc
inflating: unicolor/sample/SAM/segment_anything/__pycache__/automatic_mask_generator.cpython-38.pyc
inflating: unicolor/sample/SAM/segment_anything/__pycache__/transforms.cpython-310.pyc
inflating: unicolor/sample/SAM/segment_anything/__pycache__/build_sam.cpython-310.pyc
inflating: unicolor/sample/SAM/segment_anything/__pycache__/predictor.cpython-310.pyc
inflating: unicolor/sample/SAM/segment_anything/__init__.py
creating: unicolor/sample/SAM/demo/
creating: unicolor/sample/SAM/demo/src/
inflating: unicolor/sample/SAM/demo/src/App.tsx
inflating: unicolor/sample/SAM/demo/src/index.tsx
creating: unicolor/sample/SAM/demo/src/components/
inflating: unicolor/sample/SAM/demo/src/components/Tool.tsx
creating: unicolor/sample/SAM/demo/src/components/helpers/
inflating: unicolor/sample/SAM/demo/src/components/helpers/Interfaces.tsx
inflating: unicolor/sample/SAM/demo/src/components/helpers/onnxModelAPI.tsx
inflating: unicolor/sample/SAM/demo/src/components/helpers/maskUtils.tsx
inflating: unicolor/sample/SAM/demo/src/components/helpers/scaleHelper.tsx
creating: unicolor/sample/SAM/demo/src/components/hooks/
inflating: unicolor/sample/SAM/demo/src/components/hooks/context.tsx
inflating: unicolor/sample/SAM/demo/src/components/hooks/createContext.tsx
inflating: unicolor/sample/SAM/demo/src/components/Stage.tsx
creating: unicolor/sample/SAM/demo/src/assets/
inflating: unicolor/sample/SAM/demo/src/assets/index.html
creating: unicolor/sample/SAM/demo/src/assets/data/
inflating: unicolor/sample/SAM/demo/src/assets/data/dogs.jpg
creating: unicolor/sample/SAM/demo/src/assets/scss/
inflating: unicolor/sample/SAM/demo/src/assets/scss/App.scss
inflating: unicolor/sample/SAM/demo/tailwind.config.js
inflating: unicolor/sample/SAM/demo/README.md
inflating: unicolor/sample/SAM/demo/package.json
inflating: unicolor/sample/SAM/demo/tsconfig.json
inflating: unicolor/sample/SAM/demo/postcss.config.js
creating: unicolor/sample/SAM/demo/configs/
creating: unicolor/sample/SAM/demo/configs/webpack/
inflating: unicolor/sample/SAM/demo/configs/webpack/common.js
inflating: unicolor/sample/SAM/demo/configs/webpack/prod.js
inflating: unicolor/sample/SAM/demo/configs/webpack/dev.js
inflating: unicolor/sample/SAM/CONTRIBUTING.md
inflating: unicolor/sample/SAM/.gitignore
inflating: unicolor/sample/SAM/linter.sh
creating: unicolor/sample/SAM/notebooks/
inflating: unicolor/sample/SAM/notebooks/save_0656.png
creating: unicolor/sample/SAM/notebooks/.ipynb_checkpoints/
inflating: unicolor/sample/SAM/notebooks/.ipynb_checkpoints/predictor_example-checkpoint.ipynb
inflating: unicolor/sample/SAM/notebooks/.ipynb_checkpoints/automatic_mask_generator_example-checkpoint.ipynb
inflating: unicolor/sample/SAM/notebooks/onnx_model_example.ipynb
inflating: unicolor/sample/SAM/notebooks/automatic_mask_generator_example.ipynb
inflating: unicolor/sample/SAM/notebooks/save.png
inflating: unicolor/sample/SAM/notebooks/save_0664.png
inflating: unicolor/sample/SAM/notebooks/predictor_example.ipynb
inflating: unicolor/sample/SAM/notebooks/save_0660.png
inflating: unicolor/sample/SAM/notebooks/part_1.png
inflating: unicolor/sample/SAM/notebooks/save_part.png
creating: unicolor/sample/SAM/notebooks/images/
inflating: unicolor/sample/SAM/notebooks/images/truck.jpg
inflating: unicolor/sample/SAM/notebooks/images/000.png
inflating: unicolor/sample/SAM/notebooks/images/groceries.jpg
inflating: unicolor/sample/SAM/notebooks/images/dog.jpg
inflating: unicolor/sample/SAM/notebooks/part.png
inflating: unicolor/sample/SAM/README.md
creating: unicolor/sample/SAM/scripts/
inflating: unicolor/sample/SAM/scripts/amg.py
inflating: unicolor/sample/SAM/scripts/export_onnx_model.py
inflating: unicolor/sample/SAM/nohup.out
inflating: unicolor/sample/SAM/setup.py
creating: unicolor/sample/SAM/.git/
creating: unicolor/sample/SAM/.git/branches/
inflating: unicolor/sample/SAM/.git/description
creating: unicolor/sample/SAM/.git/info/
inflating: unicolor/sample/SAM/.git/info/exclude
inflating: unicolor/sample/SAM/.git/config
creating: unicolor/sample/SAM/.git/logs/
creating: unicolor/sample/SAM/.git/logs/refs/
creating: unicolor/sample/SAM/.git/logs/refs/remotes/
creating: unicolor/sample/SAM/.git/logs/refs/remotes/origin/
inflating: unicolor/sample/SAM/.git/logs/refs/remotes/origin/HEAD
creating: unicolor/sample/SAM/.git/logs/refs/heads/
inflating: unicolor/sample/SAM/.git/logs/refs/heads/main
inflating: unicolor/sample/SAM/.git/logs/HEAD
creating: unicolor/sample/SAM/.git/hooks/
inflating: unicolor/sample/SAM/.git/hooks/prepare-commit-msg.sample
inflating: unicolor/sample/SAM/.git/hooks/pre-push.sample
inflating: unicolor/sample/SAM/.git/hooks/pre-receive.sample
inflating: unicolor/sample/SAM/.git/hooks/applypatch-msg.sample
inflating: unicolor/sample/SAM/.git/hooks/update.sample
inflating: unicolor/sample/SAM/.git/hooks/commit-msg.sample
inflating: unicolor/sample/SAM/.git/hooks/pre-rebase.sample
inflating: unicolor/sample/SAM/.git/hooks/pre-applypatch.sample
inflating: unicolor/sample/SAM/.git/hooks/post-update.sample
inflating: unicolor/sample/SAM/.git/hooks/fsmonitor-watchman.sample
inflating: unicolor/sample/SAM/.git/hooks/pre-merge-commit.sample
inflating: unicolor/sample/SAM/.git/hooks/pre-commit.sample
creating: unicolor/sample/SAM/.git/objects/
creating: unicolor/sample/SAM/.git/objects/info/
creating: unicolor/sample/SAM/.git/objects/pack/
inflating: unicolor/sample/SAM/.git/objects/pack/pack-6bddc3bea3b5846071a7a94ed6811aed34aa66d9.pack
inflating: unicolor/sample/SAM/.git/objects/pack/pack-6bddc3bea3b5846071a7a94ed6811aed34aa66d9.idx
inflating: unicolor/sample/SAM/.git/index
inflating: unicolor/sample/SAM/.git/packed-refs
creating: unicolor/sample/SAM/.git/refs/
creating: unicolor/sample/SAM/.git/refs/remotes/
creating: unicolor/sample/SAM/.git/refs/remotes/origin/
extracting: unicolor/sample/SAM/.git/refs/remotes/origin/HEAD
creating: unicolor/sample/SAM/.git/refs/tags/
creating: unicolor/sample/SAM/.git/refs/heads/
extracting: unicolor/sample/SAM/.git/refs/heads/main
extracting: unicolor/sample/SAM/.git/HEAD
inflating: unicolor/sample/SAM/.flake8
inflating: unicolor/sample/SAM/CODE_OF_CONDUCT.md
creating: unicolor/sample/SAM/assets/
inflating: unicolor/sample/SAM/assets/masks2.jpg
inflating: unicolor/sample/SAM/assets/masks1.png
inflating: unicolor/sample/SAM/assets/notebook1.png
inflating: unicolor/sample/SAM/assets/minidemo.gif
inflating: unicolor/sample/SAM/assets/model_diagram.png
inflating: unicolor/sample/SAM/assets/notebook2.png
inflating: unicolor/sample/SAM/setup.cfg
inflating: unicolor/sample/utils_func.py
inflating: unicolor/sample/colorizer.py
creating: unicolor/sample/__pycache__/
inflating: unicolor/sample/__pycache__/utils_func.cpython-310.pyc
inflating: unicolor/sample/__pycache__/utils_func.cpython-38.pyc
inflating: unicolor/sample/__pycache__/colorizer.cpython-38.pyc
inflating: unicolor/sample/__pycache__/colorizer.cpython-310.pyc
creating: unicolor/sample/images/
inflating: unicolor/sample/images/1.jpg
inflating: unicolor/sample/images/1_exp.jpg
inflating: unicolor/sample/images/12.jpg
inflating: unicolor/sample/images/10.jpg
creating: unicolor/framework/
creating: unicolor/framework/datasets/
inflating: unicolor/framework/datasets/image_dataset.py
inflating: unicolor/framework/datasets/data_prepare.py
inflating: unicolor/framework/datasets/mask.py
inflating: unicolor/framework/datasets/utils.py
inflating: unicolor/framework/train_vqgan.py
creating: unicolor/framework/chroma_vqgan/
creating: unicolor/framework/chroma_vqgan/models/
inflating: unicolor/framework/chroma_vqgan/models/lpips.py
inflating: unicolor/framework/chroma_vqgan/models/util.py
inflating: unicolor/framework/chroma_vqgan/models/module.py
inflating: unicolor/framework/chroma_vqgan/models/vqperceptual.py
inflating: unicolor/framework/chroma_vqgan/models/discriminator.py
creating: unicolor/framework/chroma_vqgan/models/__pycache__/
inflating: unicolor/framework/chroma_vqgan/models/__pycache__/discriminator.cpython-38.pyc
inflating: unicolor/framework/chroma_vqgan/models/__pycache__/vqgan.cpython-38.pyc
inflating: unicolor/framework/chroma_vqgan/models/__pycache__/util.cpython-310.pyc
inflating: unicolor/framework/chroma_vqgan/models/__pycache__/ops.cpython-310.pyc
inflating: unicolor/framework/chroma_vqgan/models/__pycache__/lpips.cpython-310.pyc
inflating: unicolor/framework/chroma_vqgan/models/__pycache__/module.cpython-38.pyc
inflating: unicolor/framework/chroma_vqgan/models/__pycache__/quantize.cpython-38.pyc
inflating: unicolor/framework/chroma_vqgan/models/__pycache__/util.cpython-38.pyc
inflating: unicolor/framework/chroma_vqgan/models/__pycache__/quantize.cpython-310.pyc
inflating: unicolor/framework/chroma_vqgan/models/__pycache__/ops.cpython-38.pyc
inflating: unicolor/framework/chroma_vqgan/models/__pycache__/vqgan.cpython-310.pyc
inflating: unicolor/framework/chroma_vqgan/models/__pycache__/lpips.cpython-38.pyc
inflating: unicolor/framework/chroma_vqgan/models/__pycache__/vqperceptual.cpython-38.pyc
inflating: unicolor/framework/chroma_vqgan/models/__pycache__/module.cpython-310.pyc
inflating: unicolor/framework/chroma_vqgan/models/__pycache__/vqperceptual.cpython-310.pyc
inflating: unicolor/framework/chroma_vqgan/models/__pycache__/discriminator.cpython-310.pyc
inflating: unicolor/framework/chroma_vqgan/models/vqgan.py
inflating: unicolor/framework/chroma_vqgan/models/ops.py
inflating: unicolor/framework/chroma_vqgan/models/quantize.py
creating: unicolor/framework/chroma_vqgan/configs/
inflating: unicolor/framework/chroma_vqgan/configs/coco.yaml
inflating: unicolor/framework/chroma_vqgan/configs/testing.yaml
inflating: unicolor/framework/chroma_vqgan/configs/imagenet.yaml
creating: unicolor/framework/checkpoints/
creating: unicolor/framework/checkpoints/unicolor_mscoco/
inflating: unicolor/framework/checkpoints/unicolor_mscoco/config.yaml
inflating: unicolor/framework/checkpoints/unicolor_mscoco/mscoco_step259999.ckpt
creating: unicolor/framework/checkpoints/unicolor_imagenet/
inflating: unicolor/framework/checkpoints/unicolor_imagenet/imagenet_step142124.ckpt
inflating: unicolor/framework/checkpoints/unicolor_imagenet/config.yaml
creating: unicolor/framework/hybrid_tran/
creating: unicolor/framework/hybrid_tran/models/
inflating: unicolor/framework/hybrid_tran/models/transformer.py
inflating: unicolor/framework/hybrid_tran/models/colorization.py
creating: unicolor/framework/hybrid_tran/models/__pycache__/
inflating: unicolor/framework/hybrid_tran/models/__pycache__/vqgan.cpython-38.pyc
inflating: unicolor/framework/hybrid_tran/models/__pycache__/transformer.cpython-38.pyc
inflating: unicolor/framework/hybrid_tran/models/__pycache__/colorization.cpython-310.pyc
inflating: unicolor/framework/hybrid_tran/models/__pycache__/vqgan.cpython-310.pyc
inflating: unicolor/framework/hybrid_tran/models/__pycache__/colorization.cpython-38.pyc
inflating: unicolor/framework/hybrid_tran/models/__pycache__/transformer.cpython-310.pyc
inflating: unicolor/framework/hybrid_tran/models/vqgan.py
creating: unicolor/framework/hybrid_tran/utils/
creating: unicolor/framework/hybrid_tran/utils/__pycache__/
inflating: unicolor/framework/hybrid_tran/utils/__pycache__/ops.cpython-310.pyc
inflating: unicolor/framework/hybrid_tran/utils/__pycache__/ops.cpython-38.pyc
inflating: unicolor/framework/hybrid_tran/utils/ops.py
creating: unicolor/framework/hybrid_tran/configs/
inflating: unicolor/framework/hybrid_tran/configs/coco.yaml
inflating: unicolor/framework/hybrid_tran/configs/testing.yaml
inflating: unicolor/framework/train_tran.py
......
Looking in indexes: https://mirrors.cloud.aliyuncs.com/pypi/simple
Collecting nltk
Downloading https://mirrors.cloud.aliyuncs.com/pypi/packages/a6/0a/0d20d2c0f16be91b9fa32a77b76c60f9baf6eba419e5ef5deca17af9c582/nltk-3.8.1-py3-none-any.whl (1.5 MB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 1.5/1.5 MB 16.7 MB/s eta 0:00:00a 0:00:01
Requirement already satisfied: regex>=2021.8.3 in /usr/local/lib/python3.10/dist-packages (from nltk) (2023.3.23)
Requirement already satisfied: click in /usr/local/lib/python3.10/dist-packages (from nltk) (8.1.3)
Collecting joblib
Downloading https://mirrors.cloud.aliyuncs.com/pypi/packages/10/40/d551139c85db202f1f384ba8bcf96aca2f329440a844f924c8a0040b6d02/joblib-1.3.2-py3-none-any.whl (302 kB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 302.2/302.2 kB 72.8 MB/s eta 0:00:00
Requirement already satisfied: tqdm in /usr/local/lib/python3.10/dist-packages (from nltk) (4.65.0)
Installing collected packages: joblib, nltk
Successfully installed joblib-1.3.2 nltk-3.8.1
WARNING: Running pip as the 'root' user can result in broken permissions and conflicting behaviour with the system package manager. It is recommended to use a virtual environment instead: https://pip.pypa.io/warnings/venv
[notice] A new release of pip is available: 23.0.1 -> 23.2.1
[notice] To update, run: python3 -m pip install --upgrade pipLooking in indexes: https://mirrors.cloud.aliyuncs.com/pypi/simple
Requirement already satisfied: modelscope in /usr/local/lib/python3.10/dist-packages (1.8.4)
Requirement already satisfied: simplejson>=3.3.0 in /usr/local/lib/python3.10/dist-packages (from modelscope) (3.19.1)
Requirement already satisfied: gast>=0.2.2 in /usr/local/lib/python3.10/dist-packages (from modelscope) (0.5.4)
Requirement already satisfied: requests>=2.25 in /usr/local/lib/python3.10/dist-packages (from modelscope) (2.25.1)
Requirement already satisfied: oss2 in /usr/local/lib/python3.10/dist-packages (from modelscope) (2.18.1)
Requirement already satisfied: yapf in /usr/local/lib/python3.10/dist-packages (from modelscope) (0.32.0)
Requirement already satisfied: tqdm>=4.64.0 in /usr/local/lib/python3.10/dist-packages (from modelscope) (4.65.0)
Requirement already satisfied: urllib3>=1.26 in /usr/local/lib/python3.10/dist-packages (from modelscope) (1.26.15)
Requirement already satisfied: pandas in /usr/local/lib/python3.10/dist-packages (from modelscope) (1.5.3)
Requirement already satisfied: sortedcontainers>=1.5.9 in /usr/local/lib/python3.10/dist-packages (from modelscope) (2.4.0)
Requirement already satisfied: filelock>=3.3.0 in /usr/local/lib/python3.10/dist-packages (from modelscope) (3.10.7)
Requirement already satisfied: addict in /usr/local/lib/python3.10/dist-packages (from modelscope) (2.4.0)
Requirement already satisfied: setuptools in /usr/lib/python3/dist-packages (from modelscope) (59.6.0)
Requirement already satisfied: einops in /usr/local/lib/python3.10/dist-packages (from modelscope) (0.4.1)
Requirement already satisfied: Pillow>=6.2.0 in /usr/local/lib/python3.10/dist-packages (from modelscope) (9.4.0)
Requirement already satisfied: attrs in /usr/local/lib/python3.10/dist-packages (from modelscope) (22.2.0)
Requirement already satisfied: pyarrow!=9.0.0,>=6.0.0 in /usr/local/lib/python3.10/dist-packages (from modelscope) (11.0.0)
Requirement already satisfied: python-dateutil>=2.1 in /usr/local/lib/python3.10/dist-packages (from modelscope) (2.8.2)
Requirement already satisfied: numpy in /usr/local/lib/python3.10/dist-packages (from modelscope) (1.23.3)
Requirement already satisfied: scipy in /usr/local/lib/python3.10/dist-packages (from modelscope) (1.10.1)
Requirement already satisfied: datasets<=2.13.0,>=2.8.0 in /usr/local/lib/python3.10/dist-packages (from modelscope) (2.11.0)
Requirement already satisfied: pyyaml in /usr/local/lib/python3.10/dist-packages (from modelscope) (6.0)
Requirement already satisfied: dill<0.3.7,>=0.3.0 in /usr/local/lib/python3.10/dist-packages (from datasets<=2.13.0,>=2.8.0->modelscope) (0.3.6)
Requirement already satisfied: aiohttp in /usr/local/lib/python3.10/dist-packages (from datasets<=2.13.0,>=2.8.0->modelscope) (3.8.4)
Requirement already satisfied: huggingface-hub<1.0.0,>=0.11.0 in /usr/local/lib/python3.10/dist-packages (from datasets<=2.13.0,>=2.8.0->modelscope) (0.13.3)
Requirement already satisfied: xxhash in /usr/local/lib/python3.10/dist-packages (from datasets<=2.13.0,>=2.8.0->modelscope) (3.2.0)
Requirement already satisfied: multiprocess in /usr/local/lib/python3.10/dist-packages (from datasets<=2.13.0,>=2.8.0->modelscope) (0.70.14)
Requirement already satisfied: fsspec[http]>=2021.11.1 in /usr/local/lib/python3.10/dist-packages (from datasets<=2.13.0,>=2.8.0->modelscope) (2023.3.0)
Requirement already satisfied: responses<0.19 in /usr/local/lib/python3.10/dist-packages (from datasets<=2.13.0,>=2.8.0->modelscope) (0.18.0)
Requirement already satisfied: packaging in /usr/local/lib/python3.10/dist-packages (from datasets<=2.13.0,>=2.8.0->modelscope) (23.0)
Requirement already satisfied: six>=1.5 in /usr/local/lib/python3.10/dist-packages (from python-dateutil>=2.1->modelscope) (1.16.0)
Requirement already satisfied: idna<3,>=2.5 in /usr/local/lib/python3.10/dist-packages (from requests>=2.25->modelscope) (2.10)
Requirement already satisfied: certifi>=2017.4.17 in /usr/local/lib/python3.10/dist-packages (from requests>=2.25->modelscope) (2022.12.7)
Requirement already satisfied: chardet<5,>=3.0.2 in /usr/local/lib/python3.10/dist-packages (from requests>=2.25->modelscope) (4.0.0)
Requirement already satisfied: pycryptodome>=3.4.7 in /usr/local/lib/python3.10/dist-packages (from oss2->modelscope) (3.17)
Requirement already satisfied: crcmod>=1.7 in /usr/local/lib/python3.10/dist-packages (from oss2->modelscope) (1.7)
Requirement already satisfied: aliyun-python-sdk-kms>=2.4.1 in /usr/local/lib/python3.10/dist-packages (from oss2->modelscope) (2.16.1)
Requirement already satisfied: aliyun-python-sdk-core>=2.13.12 in /usr/local/lib/python3.10/dist-packages (from oss2->modelscope) (2.13.36)
Requirement already satisfied: pytz>=2020.1 in /usr/local/lib/python3.10/dist-packages (from pandas->modelscope) (2023.3)
Requirement already satisfied: cryptography>=2.6.0 in /usr/local/lib/python3.10/dist-packages (from aliyun-python-sdk-core>=2.13.12->oss2->modelscope) (40.0.1)
Requirement already satisfied: jmespath<1.0.0,>=0.9.3 in /usr/local/lib/python3.10/dist-packages (from aliyun-python-sdk-core>=2.13.12->oss2->modelscope) (0.10.0)
Requirement already satisfied: multidict<7.0,>=4.5 in /usr/local/lib/python3.10/dist-packages (from aiohttp->datasets<=2.13.0,>=2.8.0->modelscope) (6.0.4)
Requirement already satisfied: charset-normalizer<4.0,>=2.0 in /usr/local/lib/python3.10/dist-packages (from aiohttp->datasets<=2.13.0,>=2.8.0->modelscope) (3.1.0)
Requirement already satisfied: aiosignal>=1.1.2 in /usr/local/lib/python3.10/dist-packages (from aiohttp->datasets<=2.13.0,>=2.8.0->modelscope) (1.3.1)
Requirement already satisfied: frozenlist>=1.1.1 in /usr/local/lib/python3.10/dist-packages (from aiohttp->datasets<=2.13.0,>=2.8.0->modelscope) (1.3.3)
Requirement already satisfied: async-timeout<5.0,>=4.0.0a3 in /usr/local/lib/python3.10/dist-packages (from aiohttp->datasets<=2.13.0,>=2.8.0->modelscope) (4.0.2)
Requirement already satisfied: yarl<2.0,>=1.0 in /usr/local/lib/python3.10/dist-packages (from aiohttp->datasets<=2.13.0,>=2.8.0->modelscope) (1.8.2)
Requirement already satisfied: typing-extensions>=3.7.4.3 in /usr/local/lib/python3.10/dist-packages (from huggingface-hub<1.0.0,>=0.11.0->datasets<=2.13.0,>=2.8.0->modelscope) (4.5.0)
Requirement already satisfied: cffi>=1.12 in /usr/local/lib/python3.10/dist-packages (from cryptography>=2.6.0->aliyun-python-sdk-core>=2.13.12->oss2->modelscope) (1.15.1)
Requirement already satisfied: pycparser in /usr/local/lib/python3.10/dist-packages (from cffi>=1.12->cryptography>=2.6.0->aliyun-python-sdk-core>=2.13.12->oss2->modelscope) (2.21)
WARNING: Running pip as the 'root' user can result in broken permissions and conflicting behaviour with the system package manager. It is recommended to use a virtual environment instead: https://pip.pypa.io/warnings/venv
[notice] A new release of pip is available: 23.0.1 -> 23.2.1
[notice] To update, run: python3 -m pip install --upgrade pip
http://pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com/aigc-data/restoration/repo/inpaint.zip
cn-hangzhou
--2023-09-05 03:10:46-- http://pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com/aigc-data/restoration/repo/inpaint.zip
Resolving pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com (pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com)... 100.118.28.45, 100.118.28.50, 100.118.28.49, ...
Connecting to pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com (pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com)|100.118.28.45|:80... connected.
HTTP request sent, awaiting response... 200 OK
Length: 603971848 (576M) [application/zip]
Saving to: ‘inpaint.zip’
inpaint.zip 100%[===================>] 575.99M 10.0MB/s in 56s
2023-09-05 03:11:42 (10.3 MB/s) - ‘inpaint.zip’ saved [603971848/603971848]
Archive: inpaint.zip
creating: inpaint/
creating: inpaint/.ipynb_checkpoints/
inflating: inpaint/.ipynb_checkpoints/demo-checkpoint.py
inflating: inpaint/demo.py
creating: inpaint/pretrain/
creating: inpaint/pretrain/cv_fft_inpainting_lama/
inflating: inpaint/pretrain/cv_fft_inpainting_lama/resnet50-imagenet.pth
inflating: inpaint/pretrain/cv_fft_inpainting_lama/.mdl
inflating: inpaint/pretrain/cv_fft_inpainting_lama/pytorch_model.pt
creating: inpaint/pretrain/cv_fft_inpainting_lama/ade20k/
creating: inpaint/pretrain/cv_fft_inpainting_lama/ade20k/ade20k-resnet50dilated-ppm_deepsup/
inflating: inpaint/pretrain/cv_fft_inpainting_lama/ade20k/ade20k-resnet50dilated-ppm_deepsup/encoder_epoch_20.pth
inflating: inpaint/pretrain/cv_fft_inpainting_lama/README.md
inflating: inpaint/pretrain/cv_fft_inpainting_lama/configuration.json
creating: inpaint/pretrain/cv_fft_inpainting_lama/data/
inflating: inpaint/pretrain/cv_fft_inpainting_lama/data/1.gif
inflating: inpaint/pretrain/cv_fft_inpainting_lama/data/2.gif
inflating: inpaint/pretrain/cv_fft_inpainting_lama/.msc 2023-09-05 03:12:31,086 - modelscope - INFO - PyTorch version 1.13.1+cu117 Found.
2023-09-05 03:12:31,087 - modelscope - INFO - Loading ast index from /root/.cache/modelscope/ast_indexer
2023-09-05 03:12:31,118 - modelscope - INFO - Loading done! Current index file version is 1.8.4, with md5 80fa9349fc3e7b04fcfad511918062b1 and a total number of 902 components indexed
2023-09-05 03:12:31,921 - modelscope - INFO - initiate model from /mnt/workspace/inpaint/pretrain/cv_fft_inpainting_lama
2023-09-05 03:12:31,921 - modelscope - INFO - initiate model from location /mnt/workspace/inpaint/pretrain/cv_fft_inpainting_lama.
2023-09-05 03:12:31,922 - modelscope - INFO - initialize model from /mnt/workspace/inpaint/pretrain/cv_fft_inpainting_lama
2023-09-05 03:12:32,123 - modelscope - INFO - BaseInpaintingTrainingModule init called, predict_only is False
Loading weights for net_encoder
2023-09-05 03:12:33,068 - modelscope - INFO - BaseInpaintingTrainingModule init done
2023-09-05 03:12:33,068 - modelscope - INFO - loading pretrained model from /mnt/workspace/inpaint/pretrain/cv_fft_inpainting_lama/pytorch_model.pt
2023-09-05 03:12:33,319 - modelscope - WARNING - No preprocessor field found in cfg.
2023-09-05 03:12:33,319 - modelscope - WARNING - No val key and type key found in preprocessor domain of configuration.json file.
2023-09-05 03:12:33,319 - modelscope - WARNING - Cannot find available config to build preprocessor at mode inference, current config: {'model_dir': '/mnt/workspace/inpaint/pretrain/cv_fft_inpainting_lama'}. trying to build by task and model information.
2023-09-05 03:12:33,319 - modelscope - WARNING - No preprocessor key ('FFTInpainting', 'image-inpainting') found in PREPROCESSOR_MAP, skip building preprocessor.
2023-09-05 03:12:33,320 - modelscope - INFO - loading model from dir /mnt/workspace/inpaint/pretrain/cv_fft_inpainting_lama
2023-09-05 03:12:33,320 - modelscope - INFO - BaseInpaintingTrainingModule init called, predict_only is True
2023-09-05 03:12:33,708 - modelscope - INFO - BaseInpaintingTrainingModule init done
2023-09-05 03:12:33,709 - modelscope - INFO - loading pretrained model from /mnt/workspace/inpaint/pretrain/cv_fft_inpainting_lama/pytorch_model.pt
2023-09-05 03:12:34,498 - modelscope - INFO - loading model done, refinement is set to False
/usr/local/lib/python3.10/dist-packages/gradio/layouts.py:75: UserWarning: mobile_collapse is no longer supported.
warnings.warn("mobile_collapse is no longer supported.")
/usr/local/lib/python3.10/dist-packages/gradio/components.py:122: UserWarning: 'rounded' styling is no longer supported. To round adjacent components together, place them in a Column(variant='box').
warnings.warn(
/usr/local/lib/python3.10/dist-packages/gradio/components.py:131: UserWarning: 'margin' styling is no longer supported. To place adjacent components together without margin, place them in a Column(variant='box').
warnings.warn(
Running on local URL: http://127.0.0.1:7860
To create a public link, set `share=True` in `launch()`.http://pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com/aigc-data/restoration/repo/sdwebui.zip
cn-hangzhou
--2023-09-05 07:22:40-- http://pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com/aigc-data/restoration/repo/sdwebui.zip
Resolving pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com (pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com)... 100.118.28.45, 100.118.28.50, 100.118.28.44, ...
Connecting to pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com (pai-vision-data-hz2.oss-cn-hangzhou-internal.aliyuncs.com)|100.118.28.45|:80... connected.
HTTP request sent, awaiting response... 200 OK
Length: 15945784501 (15G) [application/zip]
Saving to: ‘sdwebui.zip’
sdwebui.zip 99%[==================> ] 14.82G 20.8MB/s in 12m 22s
Cannot write to ‘sdwebui.zip’ (No space left on device).
Archive: sdwebui.zip
End-of-central-directory signature not found. Either this file is not
a zipfile, or it constitutes one disk of a multi-part archive. In the
latter case the central directory and zipfile comment will be found on
the last disk(s) of this archive.
unzip: cannot find zipfile directory in one of sdwebui.zip or
sdwebui.zip.zip, and cannot find sdwebui.zip.ZIP, period.--2023-09-05 07:35:39-- https://pai-vision-data-sh.oss-cn-shanghai.aliyuncs.com/aigc-data/restoration/models/resnet101-63fe2227.pth
Resolving pai-vision-data-sh.oss-cn-shanghai.aliyuncs.com (pai-vision-data-sh.oss-cn-shanghai.aliyuncs.com)... 106.14.228.10
Connecting to pai-vision-data-sh.oss-cn-shanghai.aliyuncs.com (pai-vision-data-sh.oss-cn-shanghai.aliyuncs.com)|106.14.228.10|:443... connected.
HTTP request sent, awaiting response... 200 OK
Length: 178793939 (171M) [application/octet-stream]
Saving to: ‘/root/.cache/torch/hub/checkpoints/resnet101-63fe2227.pth.1’
resnet101-63fe2227. 100%[===================>] 170.51M 14.8MB/s in 12s
2023-09-05 07:35:52 (14.0 MB/s) - ‘/root/.cache/torch/hub/checkpoints/resnet101-63fe2227.pth.1’ saved [178793939/178793939]
--2023-09-05 07:35:52-- https://pai-vision-data-sh.oss-cn-shanghai.aliyuncs.com/aigc-data/restoration/models/resnet34-b627a593.pth
Resolving pai-vision-data-sh.oss-cn-shanghai.aliyuncs.com (pai-vision-data-sh.oss-cn-shanghai.aliyuncs.com)... 106.14.228.10
Connecting to pai-vision-data-sh.oss-cn-shanghai.aliyuncs.com (pai-vision-data-sh.oss-cn-shanghai.aliyuncs.com)|106.14.228.10|:443... connected.
HTTP request sent, awaiting response... 200 OK
Length: 87319819 (83M) [application/octet-stream]
Saving to: ‘/root/.cache/torch/hub/checkpoints/resnet34-b627a593.pth’
resnet34-b627a593.p 100%[===================>] 83.27M 19.6MB/s in 4.4s
2023-09-05 07:35:56 (18.9 MB/s) - ‘/root/.cache/torch/hub/checkpoints/resnet34-b627a593.pth’ saved [87319819/87319819]
Unable to symlink '/usr/bin/python' to '/mnt/workspace/stable-diffusion-webui/venv/bin/python'
################################################################
Install script for stable-diffusion + Web UI
Tested on Debian 11 (Bullseye)
################################################################
################################################################
Running on root user
################################################################
################################################################
Repo already cloned, using it as install directory
################################################################
################################################################
Create and activate python venv
################################################################
################################################################
Launching launch.py...
################################################################
Cannot locate TCMalloc (improves CPU memory usage)
Python 3.10.6 (main, Mar 10 2023, 10:55:28) [GCC 11.3.0]
Version: v1.5.1
Commit hash: 68f336bd994bed5442ad95bad6b6ad5564a5409a
Installing fastai==1.0.60 for DeOldify extension
Installing ffmpeg-python for DeOldify extension
Installing yt-dlp for DeOldify extension
Installing opencv-python for DeOldify extension
Installing Pillow for DeOldify extension
Launching Web UI with arguments: --no-download-sd-model --xformers --gradio-queue --disable-safe-unpickle
No SDP backend available, likely because you are running in pytorch versions < 2.0. In fact, you are using PyTorch 1.13.1+cu117. You might want to consider upgrading.
==============================================================================
You are running torch 1.13.1+cu117.
The program is tested to work with torch 2.0.0.
To reinstall the desired version, run with commandline flag --reinstall-torch.
Beware that this will cause a lot of large files to be downloaded, as well as
there are reports of issues with training tab on the latest version.
Use --skip-version-check commandline argument to disable this check.
==============================================================================
=================================================================================
You are running xformers 0.0.16rc425.
The program is tested to work with xformers 0.0.20.
To reinstall the desired version, run with commandline flag --reinstall-xformers.
Use --skip-version-check commandline argument to disable this check.
=================================================================================
Tag Autocomplete: Could not locate model-keyword extension, Lora trigger word completion will be limited to those added through the extra networks menu.
2023-09-05 07:38:10,384 - ControlNet - INFO - ControlNet v1.1.238
ControlNet preprocessor location: /mnt/workspace/stable-diffusion-webui/extensions/sd-webui-controlnet/annotator/downloads
2023-09-05 07:38:10,554 - ControlNet - INFO - ControlNet v1.1.238
Loading weights [e9d3cedc4b] from /mnt/workspace/stable-diffusion-webui/models/Stable-diffusion/realisticVisionV40_v40VAE.safetensors
Running on local URL: http://127.0.0.1:7860
To create a public link, set `share=True` in `launch()`.
Startup time: 28.1s (launcher: 17.7s, import torch: 3.3s, import gradio: 1.6s, setup paths: 1.7s, other imports: 0.9s, setup codeformer: 0.1s, load scripts: 1.9s, create ui: 0.6s, gradio launch: 0.3s).
Creating model from config: /mnt/workspace/stable-diffusion-webui/configs/v1-inference.yaml
LatentDiffusion: Running in eps-prediction mode
DiffusionWrapper has 859.52 M params.
Applying attention optimization: xformers... done.
Model loaded in 13.3s (load weights from disk: 1.2s, create model: 0.7s, apply weights to model: 10.6s, apply half(): 0.3s, move model to device: 0.4s).步骤三:修复图片
完成基于源码的图片修复操作步骤后,您已经成功完成了图片的修复,您可以前往指定的目录查看修复后的图片结果。
完成基于SDWebUI的图片修复操作步骤后,您需要在SDWebUI页面上进行进一步的操作,以实现老照片的修复。PAI提供了最常见的三种操作方式供您选择:
./stable-diffusion-webui/outputs/extras-images./stable-diffusion-webui/outputs/img2img-images/dateraw photo,a chinese woman,happy,high quality,high detail./stable-diffusion-webui/outputs/img2img-images/date
