abdbarho / stable-diffusion-webui-docker Goto Github PK
View Code? Open in Web Editor NEWEasy Docker setup for Stable Diffusion with user-friendly UI
License: Other
Easy Docker setup for Stable Diffusion with user-friendly UI
License: Other
How do I delete the downloaded cache to start fresh to use custom model?
Specs:
CPU: Ryzen 3600
RAM : 16 GB
GPU: 1660 Super 6GB
Output for --precision full in both cases:
!!Runtime error (txt2img)!!
expected scalar type Half but found Float
exiting...calling os._exit(0)
Had trouble controlling Docker Compose remotely via SSH on Win/WSL2 while there were no errors with regular Windows desktop session. Seems like logging in via SSH messes up the Docker Desktop's credit store access with WSL somehow where docker compose --profile hlky up --build
, docker compose --profile auto restart
or just docker login
would end with:
Error saving credentials: error storing credentials - err: exit status 1, out: `A specified logon session does not exist. It may already have been terminated.`
Got it to work by forcing plain text auth creds storage by removing line with "credStore" from ~/.docker/config.json
(in WSL) and then docker login
and docker compose
commands worked. Used a read-only access token made on Docker Hub site for safety as a password. Might need to be redone after reboot as config.json file is controlled by Docker Desktop app where some seem to have set that file read-only to keep it from changing with some possible side effects.
There were some open issues on Docker and Docker Compose issue trackers with this workaround but leaving this as a note here, could maybe moved to wiki.
Has this issue been opened before? Check issues here and in this one as well
Describe the bug
running docker-compose build causes
Error response from daemon: dockerfile parse error line 33: unknown instruction: GIT
ERROR: Service 'model' failed to build : Build failed
Hardware / Software:
FixFaces is broken and returns «RuntimeError: Input type (torch.cuda.FloatTensor) and weight type (torch.FloatTensor) should be the same»
stable-diffusion-webui-docker-master-model-1 | Loaded GFPGAN
stable-diffusion-webui-docker-master-model-1 | Couldn't find metadata on image
stable-diffusion-webui-docker-master-model-1 | Couldn't find metadata on image
stable-diffusion-webui-docker-master-model-1 | Traceback (most recent call last):
stable-diffusion-webui-docker-master-model-1 | File "/opt/conda/lib/python3.8/site-packages/gradio/routes.py", line 247, in run_predict
stable-diffusion-webui-docker-master-model-1 | output = await app.blocks.process_api(
stable-diffusion-webui-docker-master-model-1 | File "/opt/conda/lib/python3.8/site-packages/gradio/blocks.py", line 641, in process_api
stable-diffusion-webui-docker-master-model-1 | predictions, duration = await self.call_function(fn_index, processed_input)
stable-diffusion-webui-docker-master-model-1 | File "/opt/conda/lib/python3.8/site-packages/gradio/blocks.py", line 556, in call_function
stable-diffusion-webui-docker-master-model-1 | prediction = await anyio.to_thread.run_sync(
stable-diffusion-webui-docker-master-model-1 | File "/opt/conda/lib/python3.8/site-packages/anyio/to_thread.py", line 31, in run_sync
stable-diffusion-webui-docker-master-model-1 | return await get_asynclib().run_sync_in_worker_thread(
stable-diffusion-webui-docker-master-model-1 | File "/opt/conda/lib/python3.8/site-packages/anyio/_backends/_asyncio.py", line 937, in run_sync_in_worker_thread
stable-diffusion-webui-docker-master-model-1 | return await future
stable-diffusion-webui-docker-master-model-1 | File "/opt/conda/lib/python3.8/site-packages/anyio/_backends/_asyncio.py", line 867, in run
stable-diffusion-webui-docker-master-model-1 | result = context.run(func, *args)
stable-diffusion-webui-docker-master-model-1 | File "scripts/webui.py", line 2027, in imgproc
stable-diffusion-webui-docker-master-model-1 | image = processGFPGAN(image,imgproc_gfpgan_strength)
stable-diffusion-webui-docker-master-model-1 | File "scripts/webui.py", line 1732, in processGFPGAN
stable-diffusion-webui-docker-master-model-1 | cropped_faces, restored_faces, restored_img = GFPGAN.enhance(np.array(image, dtype=np.uint8), has_aligned=False, only_center_face=False, paste_back=True)
stable-diffusion-webui-docker-master-model-1 | File "/opt/conda/lib/python3.8/site-packages/torch/autograd/grad_mode.py", line 27, in decorate_context
stable-diffusion-webui-docker-master-model-1 | return func(*args, **kwargs)
stable-diffusion-webui-docker-master-model-1 | File "/stable-diffusion/src/gfpgan/gfpgan/utils.py", line 108, in enhance
stable-diffusion-webui-docker-master-model-1 | self.face_helper.get_face_landmarks_5(only_center_face=only_center_face, eye_dist_threshold=5)
stable-diffusion-webui-docker-master-model-1 | File "/opt/conda/lib/python3.8/site-packages/facexlib/utils/face_restoration_helper.py", line 139, in get_face_landmarks_5
stable-diffusion-webui-docker-master-model-1 | bboxes = self.face_det.detect_faces(input_img, 0.97) * scale
stable-diffusion-webui-docker-master-model-1 | File "/opt/conda/lib/python3.8/site-packages/facexlib/detection/retinaface.py", line 205, in detect_faces
stable-diffusion-webui-docker-master-model-1 | loc, conf, landmarks, priors = self.__detect_faces(image)
stable-diffusion-webui-docker-master-model-1 | File "/opt/conda/lib/python3.8/site-packages/facexlib/detection/retinaface.py", line 156, in __detect_faces
stable-diffusion-webui-docker-master-model-1 | loc, conf, landmarks = self(inputs)
stable-diffusion-webui-docker-master-model-1 | File "/opt/conda/lib/python3.8/site-packages/torch/nn/modules/module.py", line 1110, in _call_impl
stable-diffusion-webui-docker-master-model-1 | return forward_call(*input, **kwargs)
stable-diffusion-webui-docker-master-model-1 | File "/opt/conda/lib/python3.8/site-packages/facexlib/detection/retinaface.py", line 121, in forward
stable-diffusion-webui-docker-master-model-1 | out = self.body(inputs)
stable-diffusion-webui-docker-master-model-1 | File "/opt/conda/lib/python3.8/site-packages/torch/nn/modules/module.py", line 1110, in _call_impl
stable-diffusion-webui-docker-master-model-1 | return forward_call(*input, **kwargs)
stable-diffusion-webui-docker-master-model-1 | File "/opt/conda/lib/python3.8/site-packages/torchvision/models/_utils.py", line 63, in forward
stable-diffusion-webui-docker-master-model-1 | x = module(x)
stable-diffusion-webui-docker-master-model-1 | File "/opt/conda/lib/python3.8/site-packages/torch/nn/modules/module.py", line 1110, in _call_impl
stable-diffusion-webui-docker-master-model-1 | return forward_call(*input, **kwargs)
stable-diffusion-webui-docker-master-model-1 | File "/opt/conda/lib/python3.8/site-packages/torch/nn/modules/conv.py", line 447, in forward
stable-diffusion-webui-docker-master-model-1 | return self._conv_forward(input, self.weight, self.bias)
stable-diffusion-webui-docker-master-model-1 | File "/opt/conda/lib/python3.8/site-packages/torch/nn/modules/conv.py", line 443, in _conv_forward
stable-diffusion-webui-docker-master-model-1 | return F.conv2d(input, weight, bias, self.stride,
stable-diffusion-webui-docker-master-model-1 | RuntimeError: Input type (torch.cuda.FloatTensor) and weight type (torch.FloatTensor) should be the same
Steps to Reproduce
Hardware / Software:
This was working with the version pulled on Friday and stopped working this morning
Building the container works. Starting the container fails (again) with the following message:
docker compose --profile download up
[+] Running 1/0
⠿ Container webui-docker-automatic1111-1 Created 0.0s
Attaching to webui-docker-automatic1111-1, webui-docker-download-1, webui-docker-hlky-1
webui-docker-download-1 | Downloading...
webui-docker-download-1 | - GFPGANv1.3.pth exists
webui-docker-download-1 | - RealESRGAN_x4plus.pth exists
webui-docker-download-1 | - LDSR.ckpt exists
webui-docker-download-1 | - LDSR.yaml exists
webui-docker-download-1 | - RealESRGAN_x4plus_anime_6B.pth exists
webui-docker-download-1 | - model.ckpt exists
webui-docker-download-1 | Checking SHAs...
Error response from daemon: failed to create shim task: OCI runtime create failed: runc create failed: unable to start container process: error during container init: error mounting "/run/desktop/mnt/host/wsl/docker-desktop-bind-mounts/Ubuntu/0c85a38b66f627aacc67ce658a895bc6a612c6eba2cb018c2750e2a2332febb0" to rootfs at "/stable-diffusion-webui/config.json": mount /run/desktop/mnt/host/wsl/docker-desktop-bind-mounts/Ubuntu/0c85a38b66f627aacc67ce658a895bc6a612c6eba2cb018c2750e2a2332febb0:/stable-diffusion-webui/config.json (via /proc/self/fd/14), flags: 0x5000: no such file or directory: unknown
Steps to Reproduce
docker compose --profile download up
[error...]
Hardware / Software:
OS: Windows 11/WSL2
Docker version
GPU: [Nvidia 3090]
Docker version: v20.10.17
This seems to have been functioningearlier in the week week but no longer. The expected behavior is to bring the image up in the image lab tab.
Describe the bug
stable-diffusion-webui-docker-model-1 | Traceback (most recent call last):
stable-diffusion-webui-docker-model-1 | File "/opt/conda/lib/python3.8/site-packages/gradio/routes.py", line 247, in run_predict
stable-diffusion-webui-docker-model-1 | output = await app.blocks.process_api(
stable-diffusion-webui-docker-model-1 | File "/opt/conda/lib/python3.8/site-packages/gradio/blocks.py", line 639, in process_api
stable-diffusion-webui-docker-model-1 | processed_input = self.preprocess_data(fn_index, raw_input, state)
stable-diffusion-webui-docker-model-1 | File "/opt/conda/lib/python3.8/site-packages/gradio/blocks.py", line 543, in preprocess_data
stable-diffusion-webui-docker-model-1 | processed_input.append(block.preprocess(raw_input[i]))
stable-diffusion-webui-docker-model-1 | File "/opt/conda/lib/python3.8/site-packages/gradio/components.py", line 1270, in preprocess
stable-diffusion-webui-docker-model-1 | return self.choices.index(x)
stable-diffusion-webui-docker-model-1 | ValueError: [['frankfurt in the rain moody lighting\n', None], ['seed:', '4146656096'], [' ', None], ['width:', '768'], [' ', None], ['height:', '384'], [' ', None], ['steps:', '50'], [' ', None], ['cfg_scale:', '7.5'], [' ', None], ['sampler:', 'k_lms'], ['', None]] is not in list
Steps to Reproduce
Hardware / Software:
OS: Windows 11/WSL2
Docker version: v20.10.17
GPU: Nvidia 3090
Docker version: v20.10.17
Additional context
Hello, and thank you for this amazing docker image!
Upsacling with GoBig or GoLatent fail with the fllowing exception:
Traceback (most recent call last):
File "/opt/conda/lib/python3.8/site-packages/gradio/routes.py", line 247, in run_predict
output = await app.blocks.process_api(
File "/opt/conda/lib/python3.8/site-packages/gradio/blocks.py", line 641, in process_api
predictions, duration = await self.call_function(fn_index, processed_input)
File "/opt/conda/lib/python3.8/site-packages/gradio/blocks.py", line 556, in call_function
prediction = await anyio.to_thread.run_sync(
File "/opt/conda/lib/python3.8/site-packages/anyio/to_thread.py", line 31, in run_sync
return await get_asynclib().run_sync_in_worker_thread(
File "/opt/conda/lib/python3.8/site-packages/anyio/_backends/_asyncio.py", line 937, in run_sync_in_worker_thread
return await future
File "/opt/conda/lib/python3.8/site-packages/anyio/_backends/_asyncio.py", line 867, in run
result = context.run(func, *args)
File "scripts/webui.py", line 1731, in imgproc
image = processGoBig(image)
File "scripts/webui.py", line 1665, in processGoBig
work_results.append(output_images[0])
IndexError: list index out of range
Steps to Reproduce
Hardware / Software:
Additional context
Any other context about the problem here. If applicable, add screenshots to help explain your problem.
Steps to Reproduce
There are no problems with the previous steps, but the last error is reported:
Use 'docker scan' to run Snyk tests against images to find vulnerabilities and learn how to fix them
[+] Running 2/2
⠿ Network automatic1111_default Created 0.0s
⠿ Container automatic1111-model-1 Created 0.1s
Attaching to automatic1111-model-1
Error response from daemon: could not select device driver "nvidia" with capabilities: [[gpu]]
Hardware / Software:
What do I need to do to fix this?
Describe the bug
Right after txt2img finishes I receive the error
Total progress: 100%|██████████| 20/20 [07:53<00:00, 23.66s/it]
webui-docker-automatic1111-cpu-1 | Traceback (most recent call last):
webui-docker-automatic1111-cpu-1 | File "/opt/conda/lib/python3.8/site-packages/gradio/routes.py", line 273, in run_predict
webui-docker-automatic1111-cpu-1 | output = await app.blocks.process_api(
webui-docker-automatic1111-cpu-1 | File "/opt/conda/lib/python3.8/site-packages/gradio/blocks.py", line 753, in process_api
webui-docker-automatic1111-cpu-1 | result = await self.call_function(fn_index, inputs, iterator)
webui-docker-automatic1111-cpu-1 | File "/opt/conda/lib/python3.8/site-packages/gradio/blocks.py", line 630, in call_function
webui-docker-automatic1111-cpu-1 | prediction = await anyio.to_thread.run_sync(
webui-docker-automatic1111-cpu-1 | File "/opt/conda/lib/python3.8/site-packages/anyio/to_thread.py", line 31, in run_sync
webui-docker-automatic1111-cpu-1 | return await get_asynclib().run_sync_in_worker_thread(
webui-docker-automatic1111-cpu-1 | File "/opt/conda/lib/python3.8/site-packages/anyio/_backends/_asyncio.py", line 937, in run_sync_in_worker_thread
webui-docker-automatic1111-cpu-1 | return await future
webui-docker-automatic1111-cpu-1 | File "/opt/conda/lib/python3.8/site-packages/anyio/_backends/_asyncio.py", line 867, in run
webui-docker-automatic1111-cpu-1 | result = context.run(func, *args)
webui-docker-automatic1111-cpu-1 | File "/stable-diffusion-webui/modules/ui.py", line 139, in f
webui-docker-automatic1111-cpu-1 | mem_stats = {k: -(v//-(1024*1024)) for k,v in shared.mem_mon.stop().items()}
webui-docker-automatic1111-cpu-1 | File "/stable-diffusion-webui/modules/memmon.py", line 77, in stop
webui-docker-automatic1111-cpu-1 | return self.read()
webui-docker-automatic1111-cpu-1 | File "/stable-diffusion-webui/modules/memmon.py", line 65, in read
webui-docker-automatic1111-cpu-1 | free, total = torch.cuda.mem_get_info()
webui-docker-automatic1111-cpu-1 | File "/opt/conda/lib/python3.8/site-packages/torch/cuda/memory.py", line 583, in mem_get_info
webui-docker-automatic1111-cpu-1 | device = torch.cuda.current_device()
webui-docker-automatic1111-cpu-1 | File "/opt/conda/lib/python3.8/site-packages/torch/cuda/__init__.py", line 481, in current_device
webui-docker-automatic1111-cpu-1 | _lazy_init()
webui-docker-automatic1111-cpu-1 | File "/opt/conda/lib/python3.8/site-packages/torch/cuda/__init__.py", line 216, in _lazy_init
webui-docker-automatic1111-cpu-1 | torch._C._cuda_init()
webui-docker-automatic1111-cpu-1 | RuntimeError: Found no NVIDIA driver on your system. Please check that you have an NVIDIA GPU and installed a driver from http://www.nvidia.com/Download/index.aspx
Which UI
auto-cpu
Steps to Reproduce
git clone
a few hours ago, I did docker compose --profile download up --build
cat cartoon handrawing
Hardware / Software:
auto-cpu
instead of gpu)Additional context
The WebUI only shows the message Error
and in browser console I have Uncaught (in promise) API Error
.
The final image is successfully saved to output
directory.
@AbdBarho Many thanks for your efforts on this docker setup! You're doing an awesome job
I get this error when trying to use GFPGAN:
Failed inference for GFPGAN: Input type (torch.FloatTensor) and weight type (torch.cuda.FloatTensor) should be the same or input should be a MKLDNN tensor and weight is a dense tensor
.
Should some flag be set to avoid using Torch's FloatTensor
?
Your first issue is a big thank you (and less than 24 hrs since you made the repo). 🎉 The webui is a bit of a mission to get up and running on e.g. Linux without the author's Windows scripts. This was an absolute pleasure! 🚀 So thanks for making and publishing.
Has this issue been opened before? Check the FAQ, the issues
Describe the bug
Each time new version of auto service has been build old files from cache/torch/hub/checkpoints are being ignored and running img2img interrogate option ends up with redownloading files.
Which UI
I play only with auto, so I can confirm it for auto.
Steps to Reproduce
Hardware / Software:
Does not matter
Additional context
N/A
Describe the bug
Running the web UI does not work. It hangs on making attention of type 'vanilla' with 512 in_channels
for a while and then exits with code 137.
Console output:
[+] Running 1/0
- Container webui-docker-automatic1111-1 Created 0.0s
Attaching to webui-docker-automatic1111-1
webui-docker-automatic1111-1 | + /docker/mount.sh
webui-docker-automatic1111-1 | Mounted model.ckpt
webui-docker-automatic1111-1 | Mounted GFPGANv1.3.pth
webui-docker-automatic1111-1 | + python3 -u ../../webui.py --listen --port 7860 --hide-ui-dir-config --medvram --opt-split-attention
webui-docker-automatic1111-1 | Loading model [7460a6fa] from /stable-diffusion-webui/repositories/stable-diffusion/models/ldm/stable-diffusion-v1/model.ckpt
webui-docker-automatic1111-1 | Global Step: 470000
webui-docker-automatic1111-1 | LatentDiffusion: Running in eps-prediction mode
webui-docker-automatic1111-1 | DiffusionWrapper has 859.52 M params.
webui-docker-automatic1111-1 | making attention of type 'vanilla' with 512 in_channels
webui-docker-automatic1111-1 | Working with z of shape (1, 4, 32, 32) = 4096 dimensions.
webui-docker-automatic1111-1 | making attention of type 'vanilla' with 512 in_channels
webui-docker-automatic1111-1 exited with code 137
Steps to Reproduce
docker compose --profile download up --build
docker compose --profile auto up --build
Hardware / Software:
Hello!
I'm trying to get this running on an M2 Mac, and am getting this error... does anyone know what it means or how to solve it?
failed to solve: executor failed running [/bin/bash -ceuxo pipefail conda install python=3.8.5 && conda clean -a -y]: exit code: 1
Thanks!!
Hello, how can I run this without GPU support?
I disabled the deploy resources section
# environment:
# - CLI_ARGS=--extra-models-cpu --optimized-turbo
# deploy:
# resources:
# reservations:
# devices:
# - driver: nvidia
# device_ids: ['0']
# capabilities: [gpu]
But I am still getting an error when trying to run:
stable-diffusion-webui-docker-model-1 | Traceback (most recent call last):
stable-diffusion-webui-docker-model-1 | File "scripts/webui.py", line 453, in
stable-diffusion-webui-docker-model-1 | model, device,config = load_SD_model()
stable-diffusion-webui-docker-model-1 | File "scripts/webui.py", line 444, in load_SD_model
stable-diffusion-webui-docker-model-1 | model = load_model_from_config(config, opt.ckpt)
stable-diffusion-webui-docker-model-1 | File "scripts/webui.py", line 159, in load_model_from_config
stable-diffusion-webui-docker-model-1 | model.cuda()
stable-diffusion-webui-docker-model-1 | File "/opt/conda/lib/python3.8/site-packages/pytorch_lightning/core/mixins/device_dtype_mixin.py", line 127, in cuda
stable-diffusion-webui-docker-model-1 | return super().cuda(device=device)
stable-diffusion-webui-docker-model-1 | File "/opt/conda/lib/python3.8/site-packages/torch/nn/modules/module.py", line 688, in cuda
stable-diffusion-webui-docker-model-1 | return self._apply(lambda t: t.cuda(device))
stable-diffusion-webui-docker-model-1 | File "/opt/conda/lib/python3.8/site-packages/torch/nn/modules/module.py", line 578, in _apply
stable-diffusion-webui-docker-model-1 | module._apply(fn)
stable-diffusion-webui-docker-model-1 | File "/opt/conda/lib/python3.8/site-packages/torch/nn/modules/module.py", line 578, in _apply
stable-diffusion-webui-docker-model-1 | module._apply(fn)
stable-diffusion-webui-docker-model-1 | File "/opt/conda/lib/python3.8/site-packages/torch/nn/modules/module.py", line 578, in _apply
stable-diffusion-webui-docker-model-1 | module._apply(fn)
stable-diffusion-webui-docker-model-1 | [Previous line repeated 1 more time]
stable-diffusion-webui-docker-model-1 | File "/opt/conda/lib/python3.8/site-packages/torch/nn/modules/module.py", line 601, in _apply
stable-diffusion-webui-docker-model-1 | param_applied = fn(param)
stable-diffusion-webui-docker-model-1 | File "/opt/conda/lib/python3.8/site-packages/torch/nn/modules/module.py", line 688, in
stable-diffusion-webui-docker-model-1 | return self._apply(lambda t: t.cuda(device))
stable-diffusion-webui-docker-model-1 | File "/opt/conda/lib/python3.8/site-packages/torch/cuda/init.py", line 216, in _lazy_init
stable-diffusion-webui-docker-model-1 | torch._C._cuda_init()
stable-diffusion-webui-docker-model-1 | RuntimeError: Found no NVIDIA driver on your system. Please check that you have an NVIDIA GPU and installed a driver from http://www.nvidia.com/Download/index.aspx
stable-diffusion-webui-docker-model-1 exited with code 1
Thank you
I haven't used Docker before, but I like how you've packaged the best repos together to try and switch between. I'm trying to get multiple repos running on my machine, and this could be the ticket. I have some questions:
How do you update the repos? Does docker update them automatically, or do you run a git pull?
Do the repos share package libraries, or are they all in their separate independent environments?
Is there any memory overhead running them with Docker?
Do the output files land in their regular output folders within the repos on the Windows file system? I read somewhere that WSL2 uses a separate Linux file system and doesn't interoperate with the Windows file system.
Thanks!
Has this issue been opened before? Check the FAQ, the issues
invoke-ai/InvokeAI#34
Running into this error, how would this be resolved in your setup?
Describe the bug
Which UI
hlky or auto or auto-cpu or lstein?
Steps to Reproduce
Hardware / Software:
Additional context
Any other context about the problem here. If applicable, add screenshots to help explain your problem.
Not sure where to ask and whether it's a known issue but the Web UI's don't seem to recover when increasing resolution enough to run out of VRAM. Having to interrupt Docker with ctrl+c and restart it currently to recover from out of memory situation with the RuntimeError: CUDA error: an illegal memory access was encountered
error. Same behavior with the earlier default UI a couple days ago, and the 'auto' UI-profile of the latest 1.0.0 release on WSL2/Docker Desktop/RTX3060.
Good job with this project, very convenient way to get experimenting with SD web UI's, thank you.
stable-diffusion-webui-docker-model-1 | /docker/mount.sh: line 25: /latentdiffusion/experiments/pretrained_models/project.yaml: No such file or directory
stable-diffusion-webui-docker-model-1 exited with code 1
EDIT: Fixed the issue in the hlky mount.sh file by removing the # from the project.yaml line but now facing this issue
stable-diffusion-webui-docker-model-1 | Error loading LDSR
stable-diffusion-webui-docker-model-1 | Loading model from /models/model.ckpt
stable-diffusion-webui-docker-model-1 | Traceback (most recent call last):
stable-diffusion-webui-docker-model-1 | File "scripts/webui.py", line 382, in try_loading_LDSR
stable-diffusion-webui-docker-model-1 | LDSR = load_LDSR(checking=True) # TODO: Should try to load both models before giving up
stable-diffusion-webui-docker-model-1 | File "scripts/webui.py", line 297, in load_LDSR
stable-diffusion-webui-docker-model-1 | raise Exception("LDSR model not found at path "+model_path)
stable-diffusion-webui-docker-model-1 | Exception: LDSR model not found at path /latent diffusion/experiments/pretrained_models/model.ckpt
stable-diffusion-webui-docker-model-1 | Global Step: 470000
stable-diffusion-webui-docker-model-1 | UNet: Running in eps-prediction mode
stable-diffusion-webui-docker-model-1 exited with code 137
Hi,
I've been trying a number of stable-diffusion installation methods that should work without GPU, but this one is the only one that I got working (out of the box!) without any breaking issue.
That's very good, but problem is that I was actually looking for an installation that would allow renderings to be run from the command line (scripted)...
So my question (which is definitely not a bug, as the repo says it's about offering a GUI): is it still possible somehow to render images without using the GUI?
Best regards,
Vic
Hello,
I really love what you're doing here. Unfortunately I don't have a system with a GPU to throw at this. Is it possible to do it only in CPU? I saw that there is the vino project for that however it doesn't have nearly the same features of this project.
Following error poped up after building image.
ln: failed to create symbolic link '/stable-diffusion/src/gfpgan/experiments/pretrained_models/GFPGANv1.3.pth': No such file or directory
(base) root@f6e4124011d3:/stable-diffusion# ls
LICENSE README.md assets data latent_diffusion.egg-info ldm.cmd models optimizedSD scripts txt2img.yaml webuildm.cmd
'Launch Waifu Diffusion.lnk' Stable_Diffusion_v1_Model_Card.md configs environment.yaml ldm main.py notebook_helpers.py run.cmd setup.py webui.cmd
I received this error when building the repo a few minutes ago. Another user has explored it in more detail here: Sygil-Dev/stable-diffusion#169
running this cammand: docker compose --profile auto up --build
#0 7.553 Solving environment: ...working... failed with initial frozen solve. Retrying with flexible solve.
#0 7.554
#0 7.554 PackagesNotFoundError: The following packages are not available from current channels:
#0 7.554
#0 7.554 - python=3.8.5
#0 7.554
#0 7.554 Current channels:
#0 7.554
#0 7.554 - https://repo.anaconda.com/pkgs/main/linux-aarch64
#0 7.554 - https://repo.anaconda.com/pkgs/main/noarch
#0 7.554 - https://repo.anaconda.com/pkgs/r/linux-aarch64
#0 7.554 - https://repo.anaconda.com/pkgs/r/noarch
#0 7.554
#0 7.554 To search for alternate channels that may provide the conda package you're
#0 7.554 looking for, navigate to
#0 7.554
#0 7.554 https://anaconda.org
#0 7.554
#0 7.554 and use the search bar at the top of the page.
Hi,
When I docker compose up -d
I get the following
root@stablediffusion:/opt/stable-diffusion-webui-docker# docker compose up -d
[+] Running 0/1
⠹ Container stable-diffusion-webui-docker-model-1 Starting 68.2s
Error response from daemon: failed to create shim task: OCI runtime create failed: runc create failed: unable to start container process: error during container init: error running hook #0: error running hook: exit status 1, stdout: , stderr: Auto-detected mode as 'legacy'
nvidia-container-cli: initialization error: load library failed: libnvidia-ml.so.1: cannot open shared object file: no such file or directory: unknown
My setup is as follows, I have a x64 computer running latest proxmox
On it I have created an LXC container with Ubuntu 22.04.1 LTS
This container is updated upgraded, docker-compose initially installed
The initial install failed due to old version 20.10.12 was replaced with 20.10.18
Installed with the following commands
apt remove docker-compose
curl -fsSL https://get.docker.com -o get-docker.sh
sh get-docker.sh
This didn't work, so I installed nvidia-docker2 as follows
distribution=$(. /etc/os-release;echo $ID$VERSION_ID) && curl -s -L https://nvidia.github.io/nvidia-docker/gpgkey | sudo apt-key add - && curl -s -L https://nvidia.github.io/nvidia-docker/$distribution/nvidia-docker.list | sudo tee /etc/apt/sources.list.d/nvidia-docker.list
apt-get update
apt-get install -y nvidia-docker2
Here are my model files
root@stablediffusion:/opt/stable-diffusion-webui-docker/models# sha1sum *
e5102721808b8f0e421e6c3a8b73248d3b08373f GFPGANv1.3.pth
60482fe909787328a7bf5a7edffb96e390465915 LDSR.ckpt
f7f8941c3117649042047220b762fe3d8b5f616c LDSR.yaml
a1b5c65842f69da7c857479b501fb7125022fabf RealESRGAN_x4plus.pth
d91e1688357e2d2238911253c57a94aeac4625a4 RealESRGAN_x4plus_anime_6B.pth
210783247af4f65a3d23d026490cc37a670964dd model.ckpt
While writing this issue, I have found one thing I have not yet tried, and I will after posting this.
Hopefully anyone walking in my footsteps will find this and not waste so much time !
apt-add-repository ppa:graphics-drivers/ppa
apt-get install nvidia-370 nvidia-prime
Maybe that will work, the version 370 is probably very old, looks like what we need here is "nvidia-smi" and that may be part of the drivers. I suspect this is going to install a whole X server ...
Also this LXC container is unprivileged so I suspect I will also run into some other problems later down the road and I don't know if I have the IOMMU stuff working at all, this is my first time using the graphics card in this computer (5 year old 1080ti)
ok installed the nvidia drivers with the following
apt install software-properties-common
apt-add-repository ppa:graphics-drivers/ppa
apt install -y nvidia-prime nvidia-driver-515-server
The nvidia repository recommends running
apt-get install phoronix-test-suite
phoronix-test-suite default-benchmark openarena xonotic tesseract gputest unigine-valley
However this package is nowhere to be found ( Unable to locate package phoronix-test-suite )
I tried installing only nvidia-prime, it changed nothing, I don't know if this package is required at all
root@stablediffusion:/opt/stable-diffusion-webui-docker# docker compose up -d
[+] Running 0/1
⠦ Container stable-diffusion-webui-docker-model-1 Starting 1.6s
Error response from daemon: failed to create shim task: OCI runtime create failed: runc create failed: unable to start container process: error during container init: error running hook #0: error running hook: exit status 1, stdout: , stderr: Auto-detected mode as 'legacy'
nvidia-container-cli: initialization error: nvml error: driver not loaded: unknown
Ok, well, on to the next error !
It looks like you set the webui repo to a specific stable version.
When I go look at that project it is frequently being updated.
How can I easily update my container? Do I need to rebuild it from main each time?
be a nice option, thanks
It's unclear to me whether or not this build has the NSFW filter option activated or not. Regardless of the default, it would be good if this could be toggled within the webUI so that users can decide whether or not they would like to generate those sorts of images. Is that an easy thing to do?
Building the container works. Starting the container fails with the following message:
stable-diffusion-webui-docker-model-1 | + /docker/mount.sh
stable-diffusion-webui-docker-model-1 | /bin/bash: line 1: /docker/mount.sh: Permission denied
Steps to Reproduce
Hardware / Software:
Additional context
Attempted to add a chmod +x /docker/mount.sh before the command is run in dockerfile, the command does execute but does not improve things. Docker seems to work otherwise.
update to latest windows 10 release 2021H2 (required for WSL to see the GPU)
update to WSL2 (download and switch to version 2)
install docker (duh, Docker Desktop)
wsl2 dos2unix dockerfile ( these were not so obvious, some characters in the files werent going well)
wsl2 dos2unix ./build/mount.sh
It would be nice to build and publish the docker image on commit, using e.g. https://github.com/marketplace/actions/build-and-push-docker-images
That way, the setup will become a lot easier as well:
curl -fsS https://github.com/AbdBarho/stable-diffusion-webui-docker/raw/master/docker-compose.yml - | docker-compose up -f -
Please add changes required to support upload of textual inversion embeddings as mentioned here: https://github.com/hlky/sd-enable-textual-inversion
Hello, firstly thank you for this great webui.
I've been playing about using auto-cpu on an Epyc CPU.
I had reached memory limitations (32gb) when trying to upscale images past 896x896 and found the improvements form Doggettx (basujindal/stable-diffusion#117) and applied this.
The improvements now let me scale images at higher resolutions but when using img2img and the "SD upscale" option anything over 1024x1024 gets ignored (although upon completion it reports the resolution chosen).
Using "Redraw whole image" is fine at higher image sizes until I run out of memory again.
Is there a limit set somewhere or am I using the wrong options somewhere?
#0 201.6 gcc -pthread -B /opt/conda/compiler_compat -Wl,--sysroot=/ -Wsign-compare -DNDEBUG -g -fwrapv -O3 -Wall -Wstrict-prototypes -fPIC -DPY_EXTENSION -DHAVE_C99_COMPLEX -Ipywt/_extensions/c -I/opt/conda/include/python3.8 -c pywt/_extensions/c/common.c -o build/temp.linux-x86_64-cpython-38/pywt/_extensions/c/common.o
#0 201.6 error: command 'gcc' failed: No such file or directory
#0 201.6 ----------------------------------------
#0 201.6 ERROR: Failed building wheel for PyWavelets
#0 201.6 ERROR: Could not build wheels for PyWavelets which use PEP 517 and cannot be installed directly
#0 201.6
#0 201.6
#0 201.6 CondaEnvException: Pip failed
#0 201.6
------
failed to solve: executor failed running [/bin/bash -ceuxo pipefail git clone https://github.com/lstein/stable-diffusion.git
cd stable-diffusion
git reset --hard 751283a2de81bee4bb571fbabe4adb19f1d85b97
conda env update --file environment.yaml -n base
conda clean -a -y
]: exit code: 1
Any ideas how to fix it?
Has this issue been opened before? Check the FAQ, the issues and in the issues in the WebUI repo
Nothing related found
Describe the bug
Python module error when trying to use Image Lab tab under hulk interface
ModuleNotFoundError: No module named 'gfpgan.archs.stylegan2_cleanonnx_arch'
Steps to Reproduce
docker compose --profile download up --build
docker compose --profile hlky up --build
webui-docker-hlky-1 | Traceback (most recent call last):
webui-docker-hlky-1 | File "/opt/conda/lib/python3.8/site-packages/gradio/routes.py", line 247, in run_predict
webui-docker-hlky-1 | output = await app.blocks.process_api(
webui-docker-hlky-1 | File "/opt/conda/lib/python3.8/site-packages/gradio/blocks.py", line 641, in process_api
webui-docker-hlky-1 | predictions, duration = await self.call_function(fn_index, processed_input)
webui-docker-hlky-1 | File "/opt/conda/lib/python3.8/site-packages/gradio/blocks.py", line 556, in call_function
webui-docker-hlky-1 | prediction = await anyio.to_thread.run_sync(
webui-docker-hlky-1 | File "/opt/conda/lib/python3.8/site-packages/anyio/to_thread.py", line 31, in run_sync
webui-docker-hlky-1 | return await get_asynclib().run_sync_in_worker_thread(
webui-docker-hlky-1 | File "/opt/conda/lib/python3.8/site-packages/anyio/_backends/_asyncio.py", line 937, in run_sync_in_worker_thread
webui-docker-hlky-1 | return await future
webui-docker-hlky-1 | File "/opt/conda/lib/python3.8/site-packages/anyio/_backends/_asyncio.py", line 867, in run
webui-docker-hlky-1 | result = context.run(func, *args)
webui-docker-hlky-1 | File "scripts/webui.py", line 1838, in imgproc
webui-docker-hlky-1 | ModelLoader(['GFPGAN'],True,False) # Load used models
webui-docker-hlky-1 | File "scripts/webui.py", line 1935, in ModelLoader
webui-docker-hlky-1 | global_vars[m] = load_GFPGAN()
webui-docker-hlky-1 | File "scripts/webui.py", line 317, in load_GFPGAN
webui-docker-hlky-1 | from gfpgan import GFPGANer
webui-docker-hlky-1 | File "/stable-diffusion/src/gfpgan/gfpgan/__init__.py", line 2, in <module>
webui-docker-hlky-1 | from .archs import *
webui-docker-hlky-1 | File "/stable-diffusion/src/gfpgan/gfpgan/archs/__init__.py", line 10, in <module>
webui-docker-hlky-1 | _arch_modules = [importlib.import_module(f'gfpgan.archs.{file_name}') for file_name in arch_filenames]
webui-docker-hlky-1 | File "/stable-diffusion/src/gfpgan/gfpgan/archs/__init__.py", line 10, in <listcomp>
webui-docker-hlky-1 | _arch_modules = [importlib.import_module(f'gfpgan.archs.{file_name}') for file_name in arch_filenames]
webui-docker-hlky-1 | File "/opt/conda/lib/python3.8/importlib/__init__.py", line 127, in import_module
webui-docker-hlky-1 | return _bootstrap._gcd_import(name[level:], package, level)
webui-docker-hlky-1 | File "/stable-diffusion/src/gfpgan/gfpgan/archs/gfpganv1_cleanonnx_arch.py", line 8, in <module>
webui-docker-hlky-1 | from .stylegan2_cleanonnx_arch import StyleGAN2GeneratorCleanONNX
webui-docker-hlky-1 | ModuleNotFoundError: No module named 'gfpgan.archs.stylegan2_cleanonnx_arch'
Hardware / Software:
Additional context
Text-to-image prompts work fine.
Thanks for this great use of docker! The --allow-code flag allows custom code to be run in auto's ui, and I don't see a reason why it shouldn't be enabled by default. I just changed the last line of the Dockerfile from
python3 -u ../../webui.py --listen --port 7860 --hide-ui-dir-config --ckpt-dir /cache/custom-models --ckpt /cache/models/model.ckpt ${CLI_ARGS}
to
python3 -u ../../webui.py --allow-code --listen --port 7860 --hide-ui-dir-config --ckpt-dir /cache/custom-models --ckpt /cache/models/model.ckpt ${CLI_ARGS}
and the feature becomes enabled.
Was working through the scripts, and found that there are explicit mountings in mount.sh
for the apps.
I tried adding a custom entry but couldn't get the model to show up in the UI.
Is there something beyond adding a new entry into mount.sh
that is required?
Example:
#!/bin/bash
set -e
declare -A MODELS
ROOT=/stable-diffusion/src
MODELS["${ROOT}/gfpgan/experiments/pretrained_models/GFPGANv1.3.pth"]=GFPGANv1.3.pth
MODELS["${ROOT}/realesrgan/experiments/pretrained_models/RealESRGAN_x4plus.pth"]=RealESRGAN_x4plus.pth
MODELS["${ROOT}/realesrgan/experiments/pretrained_models/RealESRGAN_x4plus_anime_6B.pth"]=RealESRGAN_x4plus_anime_6B.pth
MODELS["/latent-diffusion/experiments/pretrained_models/model.ckpt"]=LDSR.ckpt
MODELS["${ROOT}/waifu-diffusion/experiments/pretrained_models/waifu-diffusion.ckpt"]=waifu-diffusion.ckpt # 👈👈👈👈👈👈👈👈👈👈👈👈👈👈
MODELS_DIR=/cache/models
Building the container works. Starting the container fails (again) with the following message:
stable-diffusion-webui-docker-model-1 | + /docker/mount.sh
stable-diffusion-webui-docker-model-1 | /bin/bash: line 1: /docker/mount.sh: Permission denied
Steps to Reproduce
docker compose build
docker compose up --build
[error...]
Hardware / Software:
OS: Windows 11/WSL2
Docker version
GPU: [Nvidia 3090]
Docker version: v20.10.17
Additional context
Attempted to add a chmod +x /docker/mount.sh before the command is run in dockerfile, the command does execute but does not improve things. Docker seems to work otherwise. A different version seems to have been deployed.
Has this issue been opened before? Check the FAQ, the issues
Describe the bug
Build of hlky or auto fails in the final step with "Attaching to webui-docker-automatic1111-1
Error response from daemon: could not select device driver "nvidia" with capabilities: [[gpu]]" on an AWS g5.xlarge using an AWS Nvidia A10G Tensor Core GPU.
This occurs after step 12 in the build and web interface will not launch following the error. Error is still present after reboot and rebuild of images.
Behavior duplicated on Debian 11 and Ubuntu 22.04, both with and without adding Nvidia drivers per https://levelup.gitconnected.com/how-to-install-an-nvidia-gpu-driver-on-an-aws-ec2-instance-20185c1c578c (This shouldn't have been necessary in any case, since AWS g5 instances include the appropriate Nvidia drivers.)
Which UI
hlky or auto
Steps to Reproduce
Hardware / Software:
i checked that this issue has not been reported before.
Describe the bug
docker-compose up
produces this:
=> => transferring context: 4.20kB 0.0s
=> ERROR [ 2/12] RUN conda install python=3.8.5 && conda clean -a -y 3.0s
------
> [ 2/12] RUN conda install python=3.8.5 && conda clean -a -y:
#0 0.241 + conda install python=3.8.5
#0 0.430 Collecting package metadata (current_repodata.json): ...working... done
#0 1.315 Solving environment: ...working... failed with initial frozen solve. Retrying with flexible solve.
#0 1.316 Collecting package metadata (repodata.json): ...working... done
#0 2.806 Solving environment: ...working... failed with initial frozen solve. Retrying with flexible solve.
#0 2.807
#0 2.807 PackagesNotFoundError: The following packages are not available from current channels:
#0 2.807
#0 2.807 - python=3.8.5
#0 2.807
#0 2.807 Current channels:
#0 2.807
#0 2.807 - https://repo.anaconda.com/pkgs/main/linux-aarch64
#0 2.807 - https://repo.anaconda.com/pkgs/main/noarch
#0 2.807 - https://repo.anaconda.com/pkgs/r/linux-aarch64
#0 2.807 - https://repo.anaconda.com/pkgs/r/noarch
#0 2.807
#0 2.807 To search for alternate channels that may provide the conda package you're
#0 2.807 looking for, navigate to
#0 2.807
#0 2.807 https://anaconda.org
#0 2.807
#0 2.807 and use the search bar at the top of the page.
#0 2.807
#0 2.807
------
failed to solve: executor failed running [/bin/bash -ceuxo pipefail conda install python=3.8.5 && conda clean -a -y]: exit code: 1
Steps to Reproduce
Hardware / Software:
Hello,
First of all, brilliant idea to make a docker image for this, this is very cool 👍
When I started up the docker image, I got this error
[+] Building 0.9s (20/20) FINISHED
=> [internal] load build definition from Dockerfile 0.0s
=> => transferring dockerfile: 32B 0.0s
=> [internal] load .dockerignore 0.0s
=> => transferring context: 2B 0.0s
=> resolve image config for docker.io/docker/dockerfile:1 0.3s
=> CACHED docker-image://docker.io/docker/dockerfile:1@sha256:443aab4ca21183e069e7d8b2dc6800659 0.0s
=> [internal] load .dockerignore 0.0s
=> [internal] load build definition from Dockerfile 0.0s
=> [internal] load metadata for docker.io/continuumio/miniconda3:4.12.0 0.3s
=> [internal] load build context 0.0s
=> => transferring context: 29B 0.0s
=> [ 1/11] FROM docker.io/continuumio/miniconda3:4.12.0@sha256:977263e8d1e476972fddab1c75fe050d 0.0s
=> CACHED [ 2/11] RUN conda install python=3.8.5 && conda clean -a -y 0.0s
=> CACHED [ 3/11] RUN conda install pytorch==1.11.0 torchvision==0.12.0 cudatoolkit=11.3 -c pyt 0.0s
=> CACHED [ 4/11] RUN git clone https://github.com/hlky/stable-diffusion.git && cd stable-diffu 0.0s
=> CACHED [ 5/11] RUN conda env update --file stable-diffusion/environment.yaml --name base && 0.0s
=> CACHED [ 6/11] RUN apt-get update && apt install fonts-dejavu-core && apt-get clean 0.0s
=> CACHED [ 7/11] RUN cd stable-diffusion && git pull && git reset --hard f3ab556a0c25389cf76f8 0.0s
=> CACHED [ 8/11] RUN git clone https://github.com/hlky/stable-diffusion-webui.git && cd stable 0.0s
=> CACHED [ 9/11] COPY info.py /info.py 0.0s
=> CACHED [10/11] RUN python /info.py /stable-diffusion/frontend/frontend.py 0.0s
=> CACHED [11/11] WORKDIR /stable-diffusion 0.0s
=> exporting to image 0.0s
=> => exporting layers 0.0s
=> => writing image sha256:3e58109d2fb9a6751d41b1bf3b6b0a8cd16de27bfdf2b1ba7d68964c7a1efea9 0.0s
=> => naming to docker.io/library/stable-diffusion-webui-docker_model 0.0s
Use 'docker scan' to run Snyk tests against images to find vulnerabilities and learn how to fix them
[+] Running 2/1
- Network stable-diffusion-webui-docker_default Created 0.8s
- Container stable-diffusion-webui-docker-model-1 Created 0.1s
Attaching to stable-diffusion-webui-docker-model-1
Error response from daemon: OCI runtime create failed: container_linux.go:380: starting container process caused: process_linux.go:545: container init caused: Running hook #0:: error running hook: signal: segmentation fault, stdout: , stderr:: unknown
Obviously that doesn't tell us very much 😁 However, I have a theory that it may be the result of me using an AMD graphics card. The other day when I was trying to get SD working on my machine, I ran into a lot of problems around making SD work with AMD, so I was wondering if this has been a consideration at all... no problem at all if not, of course!! As I said it was a MASSVIE pain to try to get running when I tried it the other day (in fact, I didn't even get it!)
My graphics card is a Radeon RX 570, if that helps, also, I am running on Windows.
Thanks!!
I had a previous working build. I get this same error in main folder and the automatic subfolder:
automatic1111-model-1 | f"It looks like the config file at '{resolved_config_file}' is not a valid JSON file."
automatic1111-model-1 | OSError: It looks like the config file at '/cache/transformers/9c24e6cd9f499d02c4f21a033736dabd365962dc80fe3aeb57a8f85ea45a20a3.26fead7ea4f0f843f6eb4055dfd25693f1a71f3c6871b184042d4b126244e142' is not a valid JSON file.
Getting following error after running docker-compose up --build
model_1 | Loading model from /models/model.ckpt
model_1 | Traceback (most recent call last):
model_1 | File "scripts/webui.py", line 308, in <module>
model_1 | sd = load_sd_from_config(opt.ckpt)
model_1 | File "scripts/webui.py", line 140, in load_sd_from_config
model_1 | pl_sd = torch.load(ckpt, map_location="cpu")
model_1 | File "/opt/conda/lib/python3.8/site-packages/torch/serialization.py", line 699, in load
model_1 | with _open_file_like(f, 'rb') as opened_file:
model_1 | File "/opt/conda/lib/python3.8/site-packages/torch/serialization.py", line 231, in _open_file_like
model_1 | return _open_file(name_or_buffer, mode)
model_1 | File "/opt/conda/lib/python3.8/site-packages/torch/serialization.py", line 212, in __init__
model_1 | super(_open_file, self).__init__(open(name, mode))
model_1 | FileNotFoundError: [Errno 2] No such file or directory: '/models/model.ckpt'
Has this issue been opened before? Check the FAQ, the issues and in the issues in the WebUI repo
Describe the bug
Img 2 img generates an error.
Traceback (most recent call last):
stable-diffusion-webui-docker-model-1 | File "/opt/conda/lib/python3.8/site-packages/gradio/routes.py", line 247, in run_predict
stable-diffusion-webui-docker-model-1 | output = await app.blocks.process_api(
stable-diffusion-webui-docker-model-1 | File "/opt/conda/lib/python3.8/site-packages/gradio/blocks.py", line 639, in process_api
stable-diffusion-webui-docker-model-1 | processed_input = self.preprocess_data(fn_index, raw_input, state)
stable-diffusion-webui-docker-model-1 | File "/opt/conda/lib/python3.8/site-packages/gradio/blocks.py", line 543, in preprocess_data
stable-diffusion-webui-docker-model-1 | processed_input.append(block.preprocess(raw_input[i]))
stable-diffusion-webui-docker-model-1 | File "/opt/conda/lib/python3.8/site-packages/gradio/components.py", line 1546, in preprocess
stable-diffusion-webui-docker-model-1 | x, mask = x["image"], x["mask"]
stable-diffusion-webui-docker-model-1 | TypeError: string indices must be integers
Steps to Reproduce
Hardware / Software:
Additional context
Any other context about the problem here. If applicable, add screenshots to help explain your problem.
Describe the bug
I have a AMD Radeon RX 6800 XT. Stable diffusion supports this GPU.
After building this image, it fails to run:
=> => naming to docker.io/library/webui-docker-automatic1111 0.0s
[+] Running 1/1
⠿ Container webui-docker-automatic1111-1 Created 0.2s
Attaching to webui-docker-automatic1111-1
Error response from daemon: could not select device driver "nvidia" with capabilities: [[gpu]]
Steps to Reproduce
docker compose --profile auto up --build
(after download)Hardware / Software:
Just posting here in case someone gets the same issue. I had this issue while building the services.
Installing NVIDIA Container Toolkit solved it:
Thanks for this amazing web ui! In the settings tab, I see a box for images filename pattern, but I have no idea how to use it. Could you include an explanation of how to use it or a link to a website that does?
Does this version also expose the REST api on port 5000? (Just like the normal https://replicate.com/docs/reference/http ?)
A declarative, efficient, and flexible JavaScript library for building user interfaces.
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
An Open Source Machine Learning Framework for Everyone
The Web framework for perfectionists with deadlines.
A PHP framework for web artisans
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
Some thing interesting about web. New door for the world.
A server is a program made to process requests and deliver data to clients.
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
Some thing interesting about visualization, use data art
Some thing interesting about game, make everyone happy.
We are working to build community through open source technology. NB: members must have two-factor auth.
Open source projects and samples from Microsoft.
Google ❤️ Open Source for everyone.
Alibaba Open Source for everyone
Data-Driven Documents codes.
China tencent open source team.