[Doc] Improve OOM troubleshooting#16704
Conversation
|
👋 Hi! Thank you for contributing to the vLLM project. 💬 Join our developer Slack at https://slack.vllm.ai to discuss your PR in #pr-reviews, coordinate on features in #feat- channels, or join special interest groups in #sig- channels. Just a reminder: PRs would not trigger full CI run by default. Instead, it would only run Once the PR is approved and ready to go, your PR reviewer(s) can run CI to test the changes comprehensively before merging. To run CI, PR reviewers can either: Add 🚀 |
hmellor
left a comment
There was a problem hiding this comment.
LGTM, only issue is the DCO check
|
when I tried the latest vllm install with llama 3.1 8b now I got: @DarkLight1337 in <cell line: 0>() 10 frames /usr/local/lib/python3.11/dist-packages/vllm/entrypoints/llm.py in init(self, model, tokenizer, tokenizer_mode, skip_tokenizer_init, trust_remote_code, allowed_local_media_path, tensor_parallel_size, dtype, quantization, revision, tokenizer_revision, seed, gpu_memory_utilization, swap_space, cpu_offload_gb, enforce_eager, max_seq_len_to_capture, disable_custom_all_reduce, disable_async_output_proc, hf_token, hf_overrides, mm_processor_kwargs, task, override_pooler_config, compilation_config, **kwargs) /usr/local/lib/python3.11/dist-packages/vllm/engine/llm_engine.py in from_engine_args(cls, engine_args, usage_context, stat_loggers) /usr/local/lib/python3.11/dist-packages/vllm/v1/engine/llm_engine.py in from_vllm_config(cls, vllm_config, usage_context, stat_loggers, disable_log_stats) /usr/local/lib/python3.11/dist-packages/vllm/v1/engine/llm_engine.py in init(self, vllm_config, executor_class, log_stats, usage_context, stat_loggers, mm_registry, use_cached_outputs, multiprocess_mode) /usr/local/lib/python3.11/dist-packages/vllm/v1/engine/core_client.py in make_client(multiprocess_mode, asyncio_mode, vllm_config, executor_class, log_stats) /usr/local/lib/python3.11/dist-packages/vllm/v1/engine/core_client.py in init(self, vllm_config, executor_class, log_stats) /usr/local/lib/python3.11/dist-packages/vllm/v1/engine/core_client.py in init(self, asyncio_mode, vllm_config, executor_class, log_stats) /usr/local/lib/python3.11/dist-packages/vllm/v1/engine/core_client.py in _wait_for_engine_startup(self) /usr/local/lib/python3.11/dist-packages/zmq/sugar/socket.py in shadow(cls, address) /usr/local/lib/python3.11/dist-packages/zmq/utils/interop.py in cast_int_addr(n) ValueError: Cannot cast <zmq.Socket(zmq.ROUTER) at 0x783b0f864360> to int |
|
@darkness8i8 you need to upgrade your pyzmq version to >= 25.0 (preferably latest 26.4) |
|
@njhill sorry I'm a noob lol but shouldn't that be a dependency VLLM installs if it needs that? All I did was %pip install vllm. I did not have a no deps flag or anything. |
|
@darkness8i8 yes it should have been, and it is now, just the version constraint was missed in the last release. It also wouldn't be a problem if you didn't already have pyzmq in your env since pip would install the latest. So you must have had a preexisting older version installed (which didn't get upgraded automatically as it should have been). |
Signed-off-by: DarkLight1337 <tlleungac@connect.ust.hk>
Signed-off-by: DarkLight1337 <tlleungac@connect.ust.hk> Signed-off-by: Yang Wang <elainewy@meta.com>
Signed-off-by: DarkLight1337 <tlleungac@connect.ust.hk>
Signed-off-by: DarkLight1337 <tlleungac@connect.ust.hk>
Signed-off-by: DarkLight1337 <tlleungac@connect.ust.hk> Signed-off-by: Agata Dobrzyniewicz <adobrzyniewicz@habana.ai>
Signed-off-by: DarkLight1337 <tlleungac@connect.ust.hk> Signed-off-by: Mu Huai <tianbowen.tbw@antgroup.com>
FIX #15664
FIX #16551
FIX #16570