Error message about Langchain project startup (Python startup. py - a)

2024-02-08 15:05:47 | ERROR | stderr | INFO: Started server process [7926]
2024-02-08 15:05:47 | ERROR | stderr | INFO: Waiting for application startup.
2024-02-08 15:05:47 | ERROR | stderr | INFO: Application startup complete.
2024-02-08 15:05:47 | ERROR | stderr | INFO: Uvicorn running on http://127.0.0.1:20000 (Press CTRL+C to quit)
2024-02-08 15:05:48 | INFO | model_worker | Loading the model [‘chatglm3-6b’] on worker da1dc11a …
Loading checkpoint shards: 0%| | 0/7 [00:00<?, ?it/s]
2024-02-08 15:05:48 | ERROR | stderr | /home/david/anaconda3/envs/chat/lib/python3.11/site-packages/torch/_utils.py:831: UserWarning: TypedStor age is deprecated. It will be removed in the future and UntypedStorage will be the only storage class. This should only matter to you if you ar e using storages directly. To access UntypedStorage directly, use tensor.untyped_storage() instead of tensor.storage()
2024-02-08 15:05:48 | ERROR | stderr | return self.fget.get(instance, owner)()
Loading checkpoint shards: 14%|███████████▌ | 1/7 [00:01<00:09, 1.60s/it]
Loading checkpoint shards: 29%|███████████████████████▏ | 2/7 [00:02<00:07, 1.47s/it]
Loading checkpoint shards: 43%|██████████████████████████████████▋ | 3/7 [00:04<00:06, 1.60s/it]
Loading checkpoint shards: 57%|██████████████████████████████████████████████▎ | 4/7 [00:06<00:05, 1.86s/it]
Loading checkpoint shards: 71%|█████████████████████████████████████████████████████████▊ | 5/7 [00:09<00:03, 1.93s/it]
Loading checkpoint shards: 86%|█████████████████████████████████████████████████████████████████████▍ | 6/7 [00:10<00:01, 1.77s/it]
Loading checkpoint shards: 100%|█████████████████████████████████████████████████████████████████████████████████| 7/7 [00:11<00:00, 1.44s/it]
Loading checkpoint shards: 100%|█████████████████████████████████████████████████████████████████████████████████| 7/7 [00:11<00:00, 1.61s/it]
2024-02-08 15:06:00 | ERROR | stderr |
2024-02-08 15:06:00 | ERROR | stderr | Process model_worker - chatglm3-6b:
2024-02-08 15:06:00 | ERROR | stderr | Traceback (most recent call last):
2024-02-08 15:06:00 | ERROR | stderr | File “/home/david/anaconda3/envs/chat/lib/python3.11/site-packages/torch/cuda/init.py”, line 311, in _lazy_init
2024-02-08 15:06:00 | ERROR | stderr | queued_call()
2024-02-08 15:06:00 | ERROR | stderr | File “/home/david/anaconda3/envs/chat/lib/python3.11/site-packages/torch/cuda/init.py”, line 180, in _check_capability
2024-02-08 15:06:00 | ERROR | stderr | capability = get_device_capability(d)
2024-02-08 15:06:00 | ERROR | stderr | ^^^^^^^^^^^^^^^^^^^^^^^^
2024-02-08 15:06:00 | ERROR | stderr | File “/home/david/anaconda3/envs/chat/lib/python3.11/site-packages/torch/cuda/init.py”, line 435, in get_device_capability
2024-02-08 15:06:00 | ERROR | stderr | prop = get_device_properties(device)
2024-02-08 15:06:00 | ERROR | stderr | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
2024-02-08 15:06:00 | ERROR | stderr | File “/home/david/anaconda3/envs/chat/lib/python3.11/site-packages/torch/cuda/init.py”, line 453, in get_device_properties
2024-02-08 15:06:00 | ERROR | stderr | return _get_device_properties(device) # type: ignore[name-defined]
2024-02-08 15:06:00 | ERROR | stderr | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
2024-02-08 15:06:00 | ERROR | stderr | RuntimeError: device >= 0 && device < num_gpus INTERNAL ASSERT FAILED at “…/aten/src/ATen/cuda/CUDACont ext.cpp”:50, please report a bug to PyTorch. device=1, num_gpus=
······
2024-02-08 15:06:00 | ERROR | stderr | raise DeferredCudaCallError(msg) from e
2024-02-08 15:06:00 | ERROR | stderr | torch.cuda.DeferredCudaCallError: CUDA call failed lazily at initialization with error: device >= 0 && d evice < num_gpus INTERNAL ASSERT FAILED at “…/aten/src/ATen/cuda/CUDAContext.cpp”:50, please report a bug to PyTorch. device=1, num_gpus=

nvidia-smi
Thu Feb 8 15:10:12 2024
±----------------------------------------------------------------------------------------+
| NVIDIA-SMI 550.40.06 Driver Version: 551.23 CUDA Version: 12.4 |
|-----------------------------------------±-----------------------±---------------------+
| GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC |
| Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. |
| | | MIG M. |
|=========================================+========================+======================|
| 0 NVIDIA GeForce RTX 3090 On | 00000000:AF:00.0 Off | N/A |
| 32% 22C P8 6W / 350W | 17MiB / 24576MiB | 0% Default |
| | | N/A |
±----------------------------------------±-----------------------±---------------------+
| 1 NVIDIA GeForce RTX 3090 On | 00000000:D8:00.0 Off | N/A |
| 32% 20C P8 11W / 350W | 162MiB / 24576MiB | 0% Default |
| | | N/A |
±----------------------------------------±-----------------------±---------------------+

±----------------------------------------------------------------------------------------+
| Processes: |
| GPU GI CI PID Type Process name GPU Memory |
| ID ID Usage |
|=========================================================================================|
| 0 N/A N/A 35 G /Xwayland N/A |
| 1 N/A N/A 35 G /Xwayland N/A |
±----------------------------------------------------------------------------------------+

import torch
torch.version
‘2.1.2+cu121’
torch.version.cuda
‘12.1’
torch.cuda.is_available()
True
torch.cuda.current_device()
0