randomath commited on
Commit
5fb6526
·
verified ·
1 Parent(s): 08e60b0

Add files using upload-large-folder tool

Browse files
Files changed (50) hide show
  1. dashboard.out +0 -0
  2. dashboard_EventHead.err +0 -0
  3. dashboard_JobHead.err +0 -0
  4. dashboard_JobHead.log +5 -0
  5. dashboard_MetricsHead.err +0 -0
  6. dashboard_MetricsHead.log +5 -0
  7. dashboard_MetricsHead.out +0 -0
  8. dashboard_NodeHead.err +2 -0
  9. dashboard_ReportHead.err +0 -0
  10. dashboard_ServeHead.err +0 -0
  11. dashboard_ServeHead.out +0 -0
  12. dashboard_TrainHead.out +0 -0
  13. dashboard_agent.log +38 -0
  14. dashboard_agent.out +0 -0
  15. debug_state.txt +217 -0
  16. gcs_server.err +0 -0
  17. log_monitor.err +0 -0
  18. monitor.out +0 -0
  19. python-core-driver-01000000ffffffffffffffffffffffffffffffffffffffffffffffff_10593.log +159 -0
  20. python-core-worker-15c410d5d6a75625cb50c80927d18090e899b8edc49402fe08e50ee6_12567.log +71 -0
  21. python-core-worker-33b9d0a21a51ca22dda2aa2142cb264d1ee4f9d53a55dc567b49496c_12500.log +71 -0
  22. python-core-worker-389d4ca43c5eadc5290ba2907f911210cffe11839a5cfe9496d636c1_12110.log +142 -0
  23. python-core-worker-5ad3b871f9c47a0419d1c26aa73c88d3ae2d40ede3aeceeef3079ef2_11379.log +139 -0
  24. python-core-worker-772052e39bd349442253d65d82fc94825e9e58c75098b1b473bedce2_12586.log +72 -0
  25. python-core-worker-8d27d27b6a5c820150d6a54cd27fe296fd0409567d2b4685b9a84fc8_11896.log +195 -0
  26. python-core-worker-9224fcd6abcfd04deeca6990e3ac522c58f6eec637ba09c0e927aaef_12481.log +72 -0
  27. python-core-worker-af6e4d2eae80c226c783dd6717832e015ec8fc0144d801649c12abfe_12563.log +76 -0
  28. python-core-worker-bb0f50c5405699ae07f957ec3f7c03f2bdf40be03f6e39b39232dc16_11378.log +82 -0
  29. python-core-worker-e98598eeddae739fb0211beef22a201ab9028016b2b64fe185d8c813_12584.log +72 -0
  30. python-core-worker-f46103e29121f0b748164b47d1653310da7f304c2c8c8df73871f0e5_12507.log +72 -0
  31. python-core-worker-fc14e0d4e4b6acb4ecead813c2d960587eefa7859aac6d8e19aeec98_11374.log +84 -0
  32. runtime_env_agent.err +0 -0
  33. runtime_env_agent.log +38 -0
  34. runtime_env_agent.out +2 -0
  35. worker-100e3eeb4a57ce034285a311628f834885904c1e1ea9caa911a3c4da-ffffffff-11375.err +0 -0
  36. worker-100e3eeb4a57ce034285a311628f834885904c1e1ea9caa911a3c4da-ffffffff-11375.out +0 -0
  37. worker-1535028fd440028216a02042c55e0b58baec34df171b54a8306f4bc8-ffffffff-11376.out +0 -0
  38. worker-15c410d5d6a75625cb50c80927d18090e899b8edc49402fe08e50ee6-01000000-12567.err +2 -0
  39. worker-2cda7ffb1fdfeaaf98e6be62760ae2627c565d43cb10409f83c0a748-ffffffff-11372.err +0 -0
  40. worker-33b9d0a21a51ca22dda2aa2142cb264d1ee4f9d53a55dc567b49496c-01000000-12500.err +2 -0
  41. worker-658f00c930b44d143152233262d8e94af875b52448898614a4b579ba-ffffffff-11377.err +0 -0
  42. worker-772052e39bd349442253d65d82fc94825e9e58c75098b1b473bedce2-01000000-12586.out +2 -0
  43. worker-8d27d27b6a5c820150d6a54cd27fe296fd0409567d2b4685b9a84fc8-01000000-11896.err +8 -0
  44. worker-9224fcd6abcfd04deeca6990e3ac522c58f6eec637ba09c0e927aaef-01000000-12481.out +2 -0
  45. worker-af6e4d2eae80c226c783dd6717832e015ec8fc0144d801649c12abfe-01000000-12563.err +2 -0
  46. worker-b809b75ac50a13f3d02e083041fe7ba32c1445fa33db5801d3c6cfe5-01000000-12477.err +2 -0
  47. worker-bb0f50c5405699ae07f957ec3f7c03f2bdf40be03f6e39b39232dc16-ffffffff-11378.err +0 -0
  48. worker-f46103e29121f0b748164b47d1653310da7f304c2c8c8df73871f0e5-01000000-12507.err +2 -0
  49. worker-f46103e29121f0b748164b47d1653310da7f304c2c8c8df73871f0e5-01000000-12507.out +2 -0
  50. worker-fc14e0d4e4b6acb4ecead813c2d960587eefa7859aac6d8e19aeec98-ffffffff-11374.out +0 -0
dashboard.out ADDED
File without changes
dashboard_EventHead.err ADDED
File without changes
dashboard_JobHead.err ADDED
File without changes
dashboard_JobHead.log ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ 2026-02-27 00:30:29,626 INFO module.py:210 -- Starting module JobHead with incarnation 0 and config SubprocessModuleConfig(cluster_id_hex='1055e483f4dc49122b1241989fd976e87d4b63a2cfd37b9f5e0a28de', gcs_address='10.128.0.163:54299', session_name='session_2026-02-27_00-30-26_175126_10593', temp_dir='/tmp/ray', session_dir='/tmp/ray/session_2026-02-27_00-30-26_175126_10593', logging_level=20, logging_format='%(asctime)s\t%(levelname)s %(filename)s:%(lineno)s -- %(message)s', log_dir='/tmp/ray/session_2026-02-27_00-30-26_175126_10593/logs', logging_filename='dashboard.log', logging_rotate_bytes=536870912, logging_rotate_backup_count=5, socket_dir='/tmp/ray/session_2026-02-27_00-30-26_175126_10593/sockets')
2
+ 2026-02-27 00:30:29,631 WARNING __init__.py:161 -- DeprecationWarning: `ray.ray_constants.DASHBOARD_CLIENT_MAX_SIZE` is a private attribute and access will be removed in a future Ray version.
3
+ 2026-02-27 00:30:29,649 INFO module.py:142 -- Started aiohttp server over /tmp/ray/session_2026-02-27_00-30-26_175126_10593/sockets/dash_JobHead.
4
+ 2026-02-27 00:30:29,649 INFO module.py:225 -- Module JobHead initialized, receiving messages...
5
+ 2026-02-27 00:32:14,785 WARNING module.py:82 -- Parent process 10931 died. Exiting...
dashboard_MetricsHead.err ADDED
File without changes
dashboard_MetricsHead.log ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ 2026-02-27 00:30:29,432 INFO module.py:210 -- Starting module MetricsHead with incarnation 0 and config SubprocessModuleConfig(cluster_id_hex='1055e483f4dc49122b1241989fd976e87d4b63a2cfd37b9f5e0a28de', gcs_address='10.128.0.163:54299', session_name='session_2026-02-27_00-30-26_175126_10593', temp_dir='/tmp/ray', session_dir='/tmp/ray/session_2026-02-27_00-30-26_175126_10593', logging_level=20, logging_format='%(asctime)s\t%(levelname)s %(filename)s:%(lineno)s -- %(message)s', log_dir='/tmp/ray/session_2026-02-27_00-30-26_175126_10593/logs', logging_filename='dashboard.log', logging_rotate_bytes=536870912, logging_rotate_backup_count=5, socket_dir='/tmp/ray/session_2026-02-27_00-30-26_175126_10593/sockets')
2
+ 2026-02-27 00:30:29,433 WARNING __init__.py:161 -- DeprecationWarning: `ray.ray_constants.DASHBOARD_CLIENT_MAX_SIZE` is a private attribute and access will be removed in a future Ray version.
3
+ 2026-02-27 00:30:29,458 INFO module.py:142 -- Started aiohttp server over /tmp/ray/session_2026-02-27_00-30-26_175126_10593/sockets/dash_MetricsHead.
4
+ 2026-02-27 00:30:29,659 INFO module.py:225 -- Module MetricsHead initialized, receiving messages...
5
+ 2026-02-27 00:32:14,599 WARNING module.py:82 -- Parent process 10931 died. Exiting...
dashboard_MetricsHead.out ADDED
File without changes
dashboard_NodeHead.err ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ WARNING: All log messages before absl::InitializeLog() is called are written to STDERR
2
+ I0000 00:00:1772152332.808143 11109 chttp2_transport.cc:1341] ipv4:10.128.0.163:54299: Got goaway [2] err=UNAVAILABLE:GOAWAY received; Error code: 2; Debug Text: Cancelling all calls {http2_error:2, grpc_status:14}
dashboard_ReportHead.err ADDED
File without changes
dashboard_ServeHead.err ADDED
File without changes
dashboard_ServeHead.out ADDED
File without changes
dashboard_TrainHead.out ADDED
File without changes
dashboard_agent.log ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2026-02-27 00:30:32,687 INFO agent.py:141 -- Dashboard agent grpc address: 10.128.0.163:55678
2
+ 2026-02-27 00:30:32,688 INFO utils.py:307 -- Get all modules by type: DashboardAgentModule
3
+ 2026-02-27 00:30:33,016 INFO utils.py:340 -- Available modules: [<class 'ray.dashboard.modules.aggregator.aggregator_agent.AggregatorAgent'>, <class 'ray.dashboard.modules.event.event_agent.EventAgent'>, <class 'ray.dashboard.modules.job.job_agent.JobAgent'>, <class 'ray.dashboard.modules.log.log_agent.LogAgent'>, <class 'ray.dashboard.modules.log.log_agent.LogAgentV1Grpc'>, <class 'ray.dashboard.modules.reporter.healthz_agent.HealthzAgent'>, <class 'ray.dashboard.modules.reporter.reporter_agent.ReporterAgent'>]
4
+ 2026-02-27 00:30:33,016 INFO agent.py:160 -- Loading DashboardAgentModule: <class 'ray.dashboard.modules.aggregator.aggregator_agent.AggregatorAgent'>
5
+ 2026-02-27 00:30:33,017 WARNING __init__.py:864 -- Overriding of current MeterProvider is not allowed
6
+ 2026-02-27 00:30:33,018 INFO aggregator_agent.py:139 -- Event HTTP target is not enabled or publishing events to external HTTP service is disabled. Skipping sending events to external HTTP service. events_export_addr:
7
+ 2026-02-27 00:30:33,019 WARNING __init__.py:864 -- Overriding of current MeterProvider is not allowed
8
+ 2026-02-27 00:30:33,019 INFO agent.py:160 -- Loading DashboardAgentModule: <class 'ray.dashboard.modules.event.event_agent.EventAgent'>
9
+ 2026-02-27 00:30:33,019 INFO event_agent.py:48 -- Event agent cache buffer size: 10240
10
+ 2026-02-27 00:30:33,019 INFO agent.py:160 -- Loading DashboardAgentModule: <class 'ray.dashboard.modules.job.job_agent.JobAgent'>
11
+ 2026-02-27 00:30:33,020 INFO agent.py:160 -- Loading DashboardAgentModule: <class 'ray.dashboard.modules.log.log_agent.LogAgent'>
12
+ 2026-02-27 00:30:33,020 INFO agent.py:160 -- Loading DashboardAgentModule: <class 'ray.dashboard.modules.log.log_agent.LogAgentV1Grpc'>
13
+ 2026-02-27 00:30:33,020 INFO agent.py:160 -- Loading DashboardAgentModule: <class 'ray.dashboard.modules.reporter.healthz_agent.HealthzAgent'>
14
+ 2026-02-27 00:30:33,020 INFO agent.py:160 -- Loading DashboardAgentModule: <class 'ray.dashboard.modules.reporter.reporter_agent.ReporterAgent'>
15
+ 2026-02-27 00:30:33,026 WARNING __init__.py:864 -- Overriding of current MeterProvider is not allowed
16
+ 2026-02-27 00:30:33,166 WARNING gpu_profile_manager.py:82 -- [GpuProfilingManager] `dynolog` is not installed, GPU profiling will not be available.
17
+ 2026-02-27 00:30:33,167 WARNING gpu_profile_manager.py:125 -- [GpuProfilingManager] GPU profiling is disabled, skipping daemon setup.
18
+ 2026-02-27 00:30:33,167 INFO agent.py:165 -- Loaded 7 modules.
19
+ 2026-02-27 00:30:33,171 INFO http_server_agent.py:123 -- Dashboard agent http address: 10.128.0.163:52365
20
+ 2026-02-27 00:30:33,171 INFO http_server_agent.py:131 -- <ResourceRoute [POST] <PlainResource /api/job_agent/jobs/> -> <function JobAgent.submit_job at 0x71447bb01760>
21
+ 2026-02-27 00:30:33,171 INFO http_server_agent.py:131 -- <ResourceRoute [OPTIONS] <PlainResource /api/job_agent/jobs/> -> <bound method _PreflightHandler._preflight_handler of <aiohttp_cors.cors_config._CorsConfigImpl object at 0x71447bbfbc50>>
22
+ 2026-02-27 00:30:33,171 INFO http_server_agent.py:131 -- <ResourceRoute [POST] <DynamicResource /api/job_agent/jobs/{job_or_submission_id}/stop> -> <function JobAgent.stop_job at 0x71447bb019e0>
23
+ 2026-02-27 00:30:33,171 INFO http_server_agent.py:131 -- <ResourceRoute [OPTIONS] <DynamicResource /api/job_agent/jobs/{job_or_submission_id}/stop> -> <bound method _PreflightHandler._preflight_handler of <aiohttp_cors.cors_config._CorsConfigImpl object at 0x71447bbfbc50>>
24
+ 2026-02-27 00:30:33,172 INFO http_server_agent.py:131 -- <ResourceRoute [DELETE] <DynamicResource /api/job_agent/jobs/{job_or_submission_id}> -> <function JobAgent.delete_job at 0x71447bb01c60>
25
+ 2026-02-27 00:30:33,172 INFO http_server_agent.py:131 -- <ResourceRoute [OPTIONS] <DynamicResource /api/job_agent/jobs/{job_or_submission_id}> -> <bound method _PreflightHandler._preflight_handler of <aiohttp_cors.cors_config._CorsConfigImpl object at 0x71447bbfbc50>>
26
+ 2026-02-27 00:30:33,172 INFO http_server_agent.py:131 -- <ResourceRoute [GET] <DynamicResource /api/job_agent/jobs/{job_or_submission_id}/logs> -> <function JobAgent.get_job_logs at 0x71447bb01e40>
27
+ 2026-02-27 00:30:33,172 INFO http_server_agent.py:131 -- <ResourceRoute [OPTIONS] <DynamicResource /api/job_agent/jobs/{job_or_submission_id}/logs> -> <bound method _PreflightHandler._preflight_handler of <aiohttp_cors.cors_config._CorsConfigImpl object at 0x71447bbfbc50>>
28
+ 2026-02-27 00:30:33,172 INFO http_server_agent.py:131 -- <ResourceRoute [GET] <DynamicResource /api/job_agent/jobs/{job_or_submission_id}/logs/tail> -> <function JobAgent.tail_job_logs at 0x71447bb02020>
29
+ 2026-02-27 00:30:33,172 INFO http_server_agent.py:131 -- <ResourceRoute [OPTIONS] <DynamicResource /api/job_agent/jobs/{job_or_submission_id}/logs/tail> -> <bound method _PreflightHandler._preflight_handler of <aiohttp_cors.cors_config._CorsConfigImpl object at 0x71447bbfbc50>>
30
+ 2026-02-27 00:30:33,172 INFO http_server_agent.py:131 -- <ResourceRoute [GET] <PlainResource /api/local_raylet_healthz> -> <function HealthzAgent.health_check at 0x71447bbec9a0>
31
+ 2026-02-27 00:30:33,172 INFO http_server_agent.py:131 -- <ResourceRoute [OPTIONS] <PlainResource /api/local_raylet_healthz> -> <bound method _PreflightHandler._preflight_handler of <aiohttp_cors.cors_config._CorsConfigImpl object at 0x71447bbfbc50>>
32
+ 2026-02-27 00:30:33,172 INFO http_server_agent.py:131 -- <ResourceRoute [GET] <StaticResource /logs -> PosixPath('/tmp/ray/session_2026-02-27_00-30-26_175126_10593/logs')> -> <bound method StaticResource._handle of <StaticResource /logs -> PosixPath('/tmp/ray/session_2026-02-27_00-30-26_175126_10593/logs')>>
33
+ 2026-02-27 00:30:33,172 INFO http_server_agent.py:131 -- <ResourceRoute [OPTIONS] <StaticResource /logs -> PosixPath('/tmp/ray/session_2026-02-27_00-30-26_175126_10593/logs')> -> <bound method _PreflightHandler._preflight_handler of <aiohttp_cors.cors_config._CorsConfigImpl object at 0x71447bbfbc50>>
34
+ 2026-02-27 00:30:33,172 INFO http_server_agent.py:132 -- Registered 14 routes.
35
+ 2026-02-27 00:30:33,175 INFO process_watcher.py:45 -- raylet pid is 11302
36
+ 2026-02-27 00:30:33,175 INFO process_watcher.py:65 -- check_parent_via_pipe
37
+ 2026-02-27 00:30:33,175 INFO event_utils.py:130 -- Monitor events logs modified after 1772150432.7741392 on /tmp/ray/session_2026-02-27_00-30-26_175126_10593/logs/events, the source types are all.
38
+ 2026-02-27 00:30:33,200 INFO gpu_providers.py:500 -- Using GPU Provider: NvidiaGpuProvider
dashboard_agent.out ADDED
File without changes
debug_state.txt ADDED
@@ -0,0 +1,217 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ NodeManager:
2
+ Node ID: d9160dc27b026e787f1f81b4465a36f139cf00f32ff2820e42f20120
3
+ Node name: 10.128.0.163
4
+ InitialConfigResources: {node:__internal_head__: 1, GPU: 1, accelerator_type:L4: 1, memory: 2.07935e+10, object_store_memory: 8.91152e+09, node:10.128.0.163: 1, CPU: 8}
5
+ ClusterLeaseManager:
6
+ ========== Node: d9160dc27b026e787f1f81b4465a36f139cf00f32ff2820e42f20120 =================
7
+ Infeasible queue length: 0
8
+ Schedule queue length: 0
9
+ Grant queue length: 0
10
+ num_waiting_for_resource: 0
11
+ num_waiting_for_plasma_memory: 0
12
+ num_waiting_for_remote_node_resources: 0
13
+ num_worker_not_started_by_job_config_not_exist: 0
14
+ num_worker_not_started_by_registration_timeout: 0
15
+ num_tasks_waiting_for_workers: 0
16
+ num_cancelled_leases: 0
17
+ cluster_resource_scheduler state:
18
+ Local id: 9054801897395801548 Local resources: {"total":{memory: [207935393800000], accelerator_type:L4: [10000], CPU_group_6834375140417d94aa5cc2a5c3d701000000: [30000], node:__internal_head__: [10000], CPU: [80000], object_store_memory: [89115168760000], CPU_group_0_6834375140417d94aa5cc2a5c3d701000000: [30000], GPU_group_0_6834375140417d94aa5cc2a5c3d701000000: [10000], GPU_group_6834375140417d94aa5cc2a5c3d701000000: [10000], GPU: [10000], bundle_group_0_6834375140417d94aa5cc2a5c3d701000000: [10000000], bundle_group_6834375140417d94aa5cc2a5c3d701000000: [10000000], node:10.128.0.163: [10000]}}, "available": {memory: [207935393800000], accelerator_type:L4: [10000], CPU_group_6834375140417d94aa5cc2a5c3d701000000: [20000], node:__internal_head__: [10000], CPU: [40000], object_store_memory: [89115168760000], CPU_group_0_6834375140417d94aa5cc2a5c3d701000000: [20000], GPU_group_0_6834375140417d94aa5cc2a5c3d701000000: [6667], GPU_group_6834375140417d94aa5cc2a5c3d701000000: [6667], GPU: [0], bundle_group_0_6834375140417d94aa5cc2a5c3d701000000: [9999990], bundle_group_6834375140417d94aa5cc2a5c3d701000000: [9999990], node:10.128.0.163: [10000]}}, "labels":{"ray.io/accelerator-type":"L4","ray.io/node-id":"d9160dc27b026e787f1f81b4465a36f139cf00f32ff2820e42f20120",} is_draining: 0 is_idle: 0 Cluster resources (at most 20 nodes are shown): node id: 9054801897395801548{"total":{node:__internal_head__: 10000, GPU_group_6834375140417d94aa5cc2a5c3d701000000: 10000, object_store_memory: 89115168760000, CPU: 80000, CPU_group_0_6834375140417d94aa5cc2a5c3d701000000: 30000, accelerator_type:L4: 10000, memory: 207935393800000, bundle_group_6834375140417d94aa5cc2a5c3d701000000: 10000000, bundle_group_0_6834375140417d94aa5cc2a5c3d701000000: 10000000, CPU_group_6834375140417d94aa5cc2a5c3d701000000: 30000, node:10.128.0.163: 10000, GPU_group_0_6834375140417d94aa5cc2a5c3d701000000: 10000, GPU: 10000}}, "available": {node:__internal_head__: 10000, GPU_group_6834375140417d94aa5cc2a5c3d701000000: 6667, object_store_memory: 89115168760000, CPU_group_0_6834375140417d94aa5cc2a5c3d701000000: 20000, accelerator_type:L4: 10000, memory: 207935393800000, CPU: 40000, bundle_group_6834375140417d94aa5cc2a5c3d701000000: 9999990, bundle_group_0_6834375140417d94aa5cc2a5c3d701000000: 9999990, CPU_group_6834375140417d94aa5cc2a5c3d701000000: 20000, node:10.128.0.163: 10000, GPU_group_0_6834375140417d94aa5cc2a5c3d701000000: 6667}}, "labels":{"ray.io/node-id":"d9160dc27b026e787f1f81b4465a36f139cf00f32ff2820e42f20120","ray.io/accelerator-type":"L4",}, "is_draining": 0, "draining_deadline_timestamp_ms": -1} { "placement group locations": [], "node to bundles": []}
19
+ Waiting leases size: 0
20
+ Number of granted lease arguments: 0
21
+ Number of pinned lease arguments: 0
22
+ Number of total spilled leases: 0
23
+ Number of spilled waiting leases: 0
24
+ Number of spilled unschedulable leases: 0
25
+ Resource usage {
26
+ - (language=PYTHON actor_or_taskvLLMHttpServer.__init__ pid=13106 worker_id=a99abd04b70eed71bbc2b85849964e1f45cdec8a7b96f35e101ab940): {}
27
+ - (language=PYTHON actor_or_taskRewardLoopWorker.__init__ pid=12567 worker_id=15c410d5d6a75625cb50c80927d18090e899b8edc49402fe08e50ee6): {}
28
+ - (language=PYTHON actor_or_taskRewardLoopWorker.__init__ pid=12500 worker_id=33b9d0a21a51ca22dda2aa2142cb264d1ee4f9d53a55dc567b49496c): {}
29
+ - (language=PYTHON actor_or_taskTaskRunner.__init__ pid=11896 worker_id=8d27d27b6a5c820150d6a54cd27fe296fd0409567d2b4685b9a84fc8): {CPU: 1}
30
+ - (language=PYTHON actor_or_taskRewardLoopWorker.__init__ pid=12563 worker_id=af6e4d2eae80c226c783dd6717832e015ec8fc0144d801649c12abfe): {}
31
+ - (language=PYTHON actor_or_taskRewardLoopWorker.__init__ pid=12586 worker_id=772052e39bd349442253d65d82fc94825e9e58c75098b1b473bedce2): {}
32
+ - (language=PYTHON actor_or_taskRewardLoopWorker.__init__ pid=12584 worker_id=e98598eeddae739fb0211beef22a201ab9028016b2b64fe185d8c813): {}
33
+ - (language=PYTHON actor_or_taskWorkerDict.__init__ pid=12223 worker_id=a711ab381f1202e338fc2083afa6dd5133aebf91969e1e83b35a9610): {GPU_group_6834375140417d94aa5cc2a5c3d701000000: 0.3333, CPU_group_6834375140417d94aa5cc2a5c3d701000000: 1, bundle_group_6834375140417d94aa5cc2a5c3d701000000: 0.001, bundle_group_0_6834375140417d94aa5cc2a5c3d701000000: 0.001, CPU_group_0_6834375140417d94aa5cc2a5c3d701000000: 1, GPU_group_0_6834375140417d94aa5cc2a5c3d701000000: 0.3333}
34
+ - (language=PYTHON actor_or_taskRewardLoopWorker.__init__ pid=12507 worker_id=f46103e29121f0b748164b47d1653310da7f304c2c8c8df73871f0e5): {}
35
+ - (language=PYTHON actor_or_taskRewardLoopWorker.__init__ pid=12481 worker_id=9224fcd6abcfd04deeca6990e3ac522c58f6eec637ba09c0e927aaef): {}
36
+ - (language=PYTHON actor_or_taskRewardLoopWorker.__init__ pid=12477 worker_id=b809b75ac50a13f3d02e083041fe7ba32c1445fa33db5801d3c6cfe5): {}
37
+ }
38
+ Backlog Size per scheduling descriptor :{workerId: num backlogs}:
39
+
40
+ Granted leases by scheduling class:
41
+ ==================================================
42
+
43
+ ClusterResources:
44
+ LocalObjectManager:
45
+ - num pinned objects: 0
46
+ - pinned objects size: 0
47
+ - num objects pending restore: 0
48
+ - num objects pending spill: 0
49
+ - num bytes pending spill: 0
50
+ - num bytes currently spilled: 0
51
+ - cumulative spill requests: 0
52
+ - cumulative restore requests: 0
53
+ - spilled objects pending delete: 0
54
+
55
+ ObjectManager:
56
+ - num local objects: 0
57
+ - num unfulfilled push requests: 0
58
+ - num object pull requests: 0
59
+ - num chunks received total: 0
60
+ - num chunks received failed (all): 0
61
+ - num chunks received failed / cancelled: 0
62
+ - num chunks received failed / plasma error: 0
63
+ Event stats:
64
+ Global stats: 0 total (0 active)
65
+ Queueing time: mean = -nanms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
66
+ Execution time: mean = -nanms, total = 0.00ms
67
+ Event stats:
68
+ PushManager:
69
+ - num pushes remaining: 0
70
+ - num chunks in flight: 0
71
+ - num chunks remaining: 0
72
+ - max chunks allowed: 409
73
+ OwnershipBasedObjectDirectory:
74
+ - num listeners: 0
75
+ - cumulative location updates: 0
76
+ - num location updates per second: 0.000
77
+ - num location lookups per second: 0.000
78
+ - num locations added per second: 0.000
79
+ - num locations removed per second: 0.000
80
+ BufferPool:
81
+ - create buffer state map size: 0
82
+ PullManager:
83
+ - num bytes available for pulled objects: 8911516876
84
+ - num bytes being pulled (all): 0
85
+ - num bytes being pulled / pinned: 0
86
+ - get request bundles: BundlePullRequestQueue{0 total, 0 active, 0 inactive, 0 unpullable}
87
+ - wait request bundles: BundlePullRequestQueue{0 total, 0 active, 0 inactive, 0 unpullable}
88
+ - task request bundles: BundlePullRequestQueue{0 total, 0 active, 0 inactive, 0 unpullable}
89
+ - first get request bundle: N/A
90
+ - first wait request bundle: N/A
91
+ - first task request bundle: N/A
92
+ - num objects queued: 0
93
+ - num objects actively pulled (all): 0
94
+ - num objects actively pulled / pinned: 0
95
+ - num bundles being pulled: 0
96
+ - num pull retries: 0
97
+ - max timeout seconds: 0
98
+ - max timeout request is already processed. No entry.
99
+
100
+ WorkerPool:
101
+ - registered jobs: 1
102
+ - process_failed_job_config_missing: 0
103
+ - process_failed_rate_limited: 0
104
+ - process_failed_pending_registration: 0
105
+ - process_failed_runtime_env_setup_failed: 0
106
+ - num PYTHON workers: 15
107
+ - num PYTHON drivers: 1
108
+ - num PYTHON pending start requests: 0
109
+ - num PYTHON pending registration requests: 0
110
+ - num object spill callbacks queued: 0
111
+ - num object restore queued: 0
112
+ - num util functions queued: 0
113
+ - num idle workers: 4
114
+ LeaseDependencyManager:
115
+ - lease deps map size: 0
116
+ - get req map size: 0
117
+ - wait req map size: 0
118
+ - local objects map size: 0
119
+ WaitManager:
120
+ - num active wait requests: 0
121
+ Subscriber:
122
+ Channel WORKER_OBJECT_EVICTION
123
+ - cumulative subscribe requests: 0
124
+ - cumulative unsubscribe requests: 0
125
+ - active subscribed publishers: 0
126
+ - cumulative published messages: 0
127
+ - cumulative processed messages: 0
128
+ Channel WORKER_REF_REMOVED_CHANNEL
129
+ - cumulative subscribe requests: 0
130
+ - cumulative unsubscribe requests: 0
131
+ - active subscribed publishers: 0
132
+ - cumulative published messages: 0
133
+ - cumulative processed messages: 0
134
+ Channel WORKER_OBJECT_LOCATIONS_CHANNEL
135
+ - cumulative subscribe requests: 0
136
+ - cumulative unsubscribe requests: 0
137
+ - active subscribed publishers: 0
138
+ - cumulative published messages: 0
139
+ - cumulative processed messages: 0
140
+ num async plasma notifications: 0
141
+ Event stats:
142
+ Global stats: 7340 total (31 active)
143
+ Queueing time: mean = 33.34ms, max = 27904.78ms, min = -0.02ms, total = 244748.69ms
144
+ Execution time: mean = 10.17ms, total = 74659.76ms
145
+ Event stats:
146
+ NodeManagerService.grpc_server.ReportWorkerBacklog - 1077 total (0 active), Execution time: mean = 0.30ms, total = 328.03ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
147
+ NodeManagerService.grpc_server.ReportWorkerBacklog.HandleRequestImpl - 1077 total (0 active), Execution time: mean = 0.07ms, total = 75.18ms, Queueing time: mean = 0.04ms, max = 3.58ms, min = 0.01ms, total = 46.22ms
148
+ NodeManager.CheckGC - 1000 total (1 active), Execution time: mean = 0.00ms, total = 2.35ms, Queueing time: mean = 0.09ms, max = 3.81ms, min = 0.01ms, total = 85.07ms
149
+ RaySyncer.OnDemandBroadcasting - 1000 total (1 active), Execution time: mean = 0.01ms, total = 11.10ms, Queueing time: mean = 0.08ms, max = 3.81ms, min = 0.00ms, total = 76.83ms
150
+ ObjectManager.UpdateAvailableMemory - 1000 total (0 active), Execution time: mean = 0.00ms, total = 3.36ms, Queueing time: mean = 0.03ms, max = 2.65ms, min = 0.00ms, total = 27.45ms
151
+ RayletWorkerPool.deadline_timer.kill_idle_workers - 500 total (1 active), Execution time: mean = 0.01ms, total = 7.18ms, Queueing time: mean = 0.08ms, max = 3.95ms, min = -0.02ms, total = 40.53ms
152
+ MemoryMonitor.CheckIsMemoryUsageAboveThreshold - 400 total (1 active), Execution time: mean = 0.27ms, total = 107.49ms, Queueing time: mean = 0.08ms, max = 4.72ms, min = 0.02ms, total = 31.70ms
153
+ NodeManager.CheckForUnexpectedWorkerDisconnects - 101 total (1 active), Execution time: mean = 0.02ms, total = 2.18ms, Queueing time: mean = 0.07ms, max = 3.43ms, min = 0.01ms, total = 7.02ms
154
+ NodeManager.ScheduleAndGrantLeases - 101 total (1 active), Execution time: mean = 0.01ms, total = 1.45ms, Queueing time: mean = 0.08ms, max = 3.44ms, min = 0.02ms, total = 7.88ms
155
+ ClientConnection.async_read.ProcessMessageHeader - 100 total (16 active), Execution time: mean = 0.01ms, total = 0.67ms, Queueing time: mean = 2411.77ms, max = 27904.78ms, min = 0.02ms, total = 241176.80ms
156
+ NodeManager.deadline_timer.spill_objects_when_over_threshold - 100 total (1 active), Execution time: mean = 0.00ms, total = 0.24ms, Queueing time: mean = 0.16ms, max = 4.03ms, min = 0.02ms, total = 15.68ms
157
+ NodeManagerService.grpc_server.GetResourceLoad - 100 total (0 active), Execution time: mean = 0.37ms, total = 36.90ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
158
+ NodeManagerService.grpc_server.GetResourceLoad.HandleRequestImpl - 100 total (0 active), Execution time: mean = 0.13ms, total = 12.61ms, Queueing time: mean = 0.05ms, max = 2.00ms, min = 0.01ms, total = 4.78ms
159
+ NodeManager.deadline_timer.flush_free_objects - 100 total (1 active), Execution time: mean = 0.01ms, total = 0.61ms, Queueing time: mean = 0.15ms, max = 4.03ms, min = 0.02ms, total = 15.42ms
160
+ ClientConnection.async_read.ProcessMessage - 84 total (0 active), Execution time: mean = 0.57ms, total = 48.29ms, Queueing time: mean = 0.02ms, max = 0.62ms, min = 0.00ms, total = 2.03ms
161
+ CoreWorkerService.grpc_client.GetCoreWorkerStats - 62 total (0 active), Execution time: mean = 2.60ms, total = 161.03ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
162
+ CoreWorkerService.grpc_client.GetCoreWorkerStats.OnReplyReceived - 62 total (0 active), Execution time: mean = 0.02ms, total = 1.20ms, Queueing time: mean = 0.82ms, max = 5.35ms, min = 0.01ms, total = 50.80ms
163
+ ClusterResourceManager.ResetRemoteNodeView - 34 total (1 active), Execution time: mean = 0.01ms, total = 0.29ms, Queueing time: mean = 0.08ms, max = 1.25ms, min = 0.02ms, total = 2.64ms
164
+ ClientConnection.async_write.DoAsyncWrites - 22 total (0 active), Execution time: mean = 0.00ms, total = 0.02ms, Queueing time: mean = 0.30ms, max = 4.39ms, min = 0.01ms, total = 6.65ms
165
+ NodeManagerService.grpc_server.GetSystemConfig.HandleRequestImpl - 21 total (0 active), Execution time: mean = 0.12ms, total = 2.61ms, Queueing time: mean = 0.03ms, max = 0.16ms, min = 0.01ms, total = 0.70ms
166
+ ObjectManager.ObjectAdded - 21 total (0 active), Execution time: mean = 0.08ms, total = 1.69ms, Queueing time: mean = 0.10ms, max = 1.31ms, min = 0.01ms, total = 2.15ms
167
+ ObjectManager.ObjectDeleted - 21 total (0 active), Execution time: mean = 0.02ms, total = 0.48ms, Queueing time: mean = 0.11ms, max = 0.56ms, min = 0.03ms, total = 2.25ms
168
+ NodeManagerService.grpc_server.GetSystemConfig - 21 total (0 active), Execution time: mean = 0.54ms, total = 11.32ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
169
+ NodeManager.deadline_timer.record_metrics - 20 total (1 active), Execution time: mean = 0.26ms, total = 5.22ms, Queueing time: mean = 0.53ms, max = 3.92ms, min = 0.03ms, total = 10.53ms
170
+ NodeManagerService.grpc_server.GetWorkerPIDs - 19 total (0 active), Execution time: mean = 0.32ms, total = 6.07ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
171
+ NodeManagerService.grpc_server.GetWorkerPIDs.HandleRequestImpl - 19 total (0 active), Execution time: mean = 0.10ms, total = 1.85ms, Queueing time: mean = 0.03ms, max = 0.06ms, min = 0.02ms, total = 0.65ms
172
+ PeriodicalRunner.RunFnPeriodically - 14 total (0 active), Execution time: mean = 0.25ms, total = 3.56ms, Queueing time: mean = 7.37ms, max = 21.50ms, min = 0.08ms, total = 103.24ms
173
+ NodeManagerService.grpc_server.RequestWorkerLease.HandleRequestImpl - 13 total (0 active), Execution time: mean = 0.41ms, total = 5.37ms, Queueing time: mean = 0.06ms, max = 0.49ms, min = 0.02ms, total = 0.84ms
174
+ NodeManagerService.grpc_server.RequestWorkerLease - 13 total (0 active), Execution time: mean = 3354.45ms, total = 43607.90ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
175
+ event_loop_lag_probe - 12 total (0 active), Execution time: mean = 0.02ms, total = 0.19ms, Queueing time: mean = 1.15ms, max = 13.76ms, min = 0.00ms, total = 13.82ms
176
+ RaySyncer.BroadcastMessage - 10 total (0 active), Execution time: mean = 0.21ms, total = 2.05ms, Queueing time: mean = 0.00ms, max = 0.00ms, min = 0.00ms, total = 0.00ms
177
+ NodeManager.deadline_timer.debug_state_dump - 10 total (1 active, 1 running), Execution time: mean = 1.22ms, total = 12.15ms, Queueing time: mean = 0.33ms, max = 1.70ms, min = 0.02ms, total = 3.35ms
178
+ - 10 total (0 active), Execution time: mean = 0.00ms, total = 0.01ms, Queueing time: mean = 0.31ms, max = 2.76ms, min = 0.01ms, total = 3.06ms
179
+ ray::rpc::InternalPubSubGcsService.grpc_client.GcsSubscriberPoll - 7 total (1 active), Execution time: mean = 4145.00ms, total = 29015.02ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
180
+ NodeManagerService.grpc_server.GetNodeStats.HandleRequestImpl - 6 total (0 active), Execution time: mean = 2.44ms, total = 14.67ms, Queueing time: mean = 0.03ms, max = 0.04ms, min = 0.03ms, total = 0.19ms
181
+ ray::rpc::InternalPubSubGcsService.grpc_client.GcsSubscriberPoll.OnReplyReceived - 6 total (0 active), Execution time: mean = 0.16ms, total = 0.98ms, Queueing time: mean = 0.03ms, max = 0.03ms, min = 0.02ms, total = 0.18ms
182
+ NodeManagerService.grpc_server.GetNodeStats - 6 total (0 active), Execution time: mean = 16.21ms, total = 97.27ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
183
+ CoreWorkerService.grpc_client.Exit - 5 total (0 active), Execution time: mean = 1.01ms, total = 5.04ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
184
+ Subscriber.HandlePublishedMessage_GCS_WORKER_DELTA_CHANNEL - 5 total (0 active), Execution time: mean = 0.01ms, total = 0.04ms, Queueing time: mean = 0.19ms, max = 0.24ms, min = 0.13ms, total = 0.97ms
185
+ CoreWorkerService.grpc_client.Exit.OnReplyReceived - 5 total (0 active), Execution time: mean = 0.06ms, total = 0.32ms, Queueing time: mean = 0.04ms, max = 0.06ms, min = 0.02ms, total = 0.18ms
186
+ ray::rpc::WorkerInfoGcsService.grpc_client.ReportWorkerFailure - 5 total (0 active), Execution time: mean = 1.56ms, total = 7.81ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
187
+ ray::rpc::WorkerInfoGcsService.grpc_client.ReportWorkerFailure.OnReplyReceived - 5 total (0 active), Execution time: mean = 0.02ms, total = 0.11ms, Queueing time: mean = 0.23ms, max = 0.46ms, min = 0.01ms, total = 1.13ms
188
+ ReporterService.grpc_client.HealthCheck.OnReplyReceived - 4 total (0 active), Execution time: mean = 0.12ms, total = 0.50ms, Queueing time: mean = 0.36ms, max = 1.22ms, min = 0.02ms, total = 1.43ms
189
+ ReporterService.grpc_client.HealthCheck - 4 total (0 active), Execution time: mean = 0.71ms, total = 2.85ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
190
+ MetricsAgentClient.WaitForServerReadyWithRetry - 3 total (0 active), Execution time: mean = 0.19ms, total = 0.58ms, Queueing time: mean = 1000.04ms, max = 1000.06ms, min = 1000.02ms, total = 3000.13ms
191
+ ray::rpc::NodeInfoGcsService.grpc_client.CheckAlive - 2 total (0 active), Execution time: mean = 1.48ms, total = 2.95ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
192
+ ray::rpc::NodeInfoGcsService.grpc_client.CheckAlive.OnReplyReceived - 2 total (0 active), Execution time: mean = 0.06ms, total = 0.12ms, Queueing time: mean = 0.79ms, max = 1.56ms, min = 0.02ms, total = 1.57ms
193
+ ray::rpc::InternalPubSubGcsService.grpc_client.GcsSubscriberCommandBatch - 2 total (0 active), Execution time: mean = 1.59ms, total = 3.18ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
194
+ NodeManager.GcsCheckAlive - 2 total (1 active), Execution time: mean = 0.15ms, total = 0.29ms, Queueing time: mean = 1.02ms, max = 2.05ms, min = 2.05ms, total = 2.05ms
195
+ NodeManager.deadline_timer.print_event_loop_stats - 2 total (1 active), Execution time: mean = 1.14ms, total = 2.29ms, Queueing time: mean = 0.02ms, max = 0.03ms, min = 0.03ms, total = 0.03ms
196
+ NodeManagerService.grpc_server.ReturnWorkerLease - 2 total (0 active), Execution time: mean = 0.35ms, total = 0.69ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
197
+ ray::rpc::InternalPubSubGcsService.grpc_client.GcsSubscriberCommandBatch.OnReplyReceived - 2 total (0 active), Execution time: mean = 0.28ms, total = 0.57ms, Queueing time: mean = 0.96ms, max = 1.83ms, min = 0.08ms, total = 1.91ms
198
+ NodeManagerService.grpc_server.ReturnWorkerLease.HandleRequestImpl - 2 total (0 active), Execution time: mean = 0.13ms, total = 0.27ms, Queueing time: mean = 0.03ms, max = 0.03ms, min = 0.02ms, total = 0.05ms
199
+ RaySyncerRegister - 2 total (0 active), Execution time: mean = 0.00ms, total = 0.01ms, Queueing time: mean = 0.00ms, max = 0.00ms, min = 0.00ms, total = 0.00ms
200
+ ray::rpc::NodeInfoGcsService.grpc_client.GetAllNodeAddressAndLiveness - 1 total (0 active), Execution time: mean = 1.28ms, total = 1.28ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
201
+ ray::rpc::JobInfoGcsService.grpc_client.GetAllJobInfo.OnReplyReceived - 1 total (0 active), Execution time: mean = 0.02ms, total = 0.02ms, Queueing time: mean = 0.16ms, max = 0.16ms, min = 0.16ms, total = 0.16ms
202
+ ray::rpc::InternalKVGcsService.grpc_client.GetInternalConfig - 1 total (0 active), Execution time: mean = 1.38ms, total = 1.38ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
203
+ ray::rpc::InternalKVGcsService.grpc_client.GetInternalConfig.OnReplyReceived - 1 total (0 active), Execution time: mean = 1032.07ms, total = 1032.07ms, Queueing time: mean = 0.02ms, max = 0.02ms, min = 0.02ms, total = 0.02ms
204
+ ray::rpc::NodeInfoGcsService.grpc_client.RegisterNode - 1 total (0 active), Execution time: mean = 3.05ms, total = 3.05ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
205
+ WorkerPool.PopWorkerCallback - 1 total (0 active), Execution time: mean = 0.02ms, total = 0.02ms, Queueing time: mean = 0.01ms, max = 0.01ms, min = 0.01ms, total = 0.01ms
206
+ ray::rpc::JobInfoGcsService.grpc_client.AddJob.OnReplyReceived - 1 total (0 active), Execution time: mean = 0.07ms, total = 0.07ms, Queueing time: mean = 0.03ms, max = 0.03ms, min = 0.03ms, total = 0.03ms
207
+ ray::rpc::NodeInfoGcsService.grpc_client.RegisterNode.OnReplyReceived - 1 total (0 active), Execution time: mean = 0.91ms, total = 0.91ms, Queueing time: mean = 0.09ms, max = 0.09ms, min = 0.09ms, total = 0.09ms
208
+ NodeManager.GCTaskFailureReason - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
209
+ NodeManagerService.grpc_server.CommitBundleResources.HandleRequestImpl - 1 total (0 active), Execution time: mean = 0.26ms, total = 0.26ms, Queueing time: mean = 0.03ms, max = 0.03ms, min = 0.03ms, total = 0.03ms
210
+ ray::rpc::NodeInfoGcsService.grpc_client.GetAllNodeAddressAndLiveness.OnReplyReceived - 1 total (0 active), Execution time: mean = 0.86ms, total = 0.86ms, Queueing time: mean = 0.33ms, max = 0.33ms, min = 0.33ms, total = 0.33ms
211
+ NodeManagerService.grpc_server.CommitBundleResources - 1 total (0 active), Execution time: mean = 0.48ms, total = 0.48ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
212
+ NodeManagerService.grpc_server.PrepareBundleResources.HandleRequestImpl - 1 total (0 active), Execution time: mean = 0.26ms, total = 0.26ms, Queueing time: mean = 0.03ms, max = 0.03ms, min = 0.03ms, total = 0.03ms
213
+ Subscriber.HandlePublishedMessage_GCS_JOB_CHANNEL - 1 total (0 active), Execution time: mean = 0.26ms, total = 0.26ms, Queueing time: mean = 0.10ms, max = 0.10ms, min = 0.10ms, total = 0.10ms
214
+ NodeManagerService.grpc_server.PrepareBundleResources - 1 total (0 active), Execution time: mean = 0.51ms, total = 0.51ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
215
+ ray::rpc::JobInfoGcsService.grpc_client.AddJob - 1 total (0 active), Execution time: mean = 1.09ms, total = 1.09ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
216
+ ray::rpc::JobInfoGcsService.grpc_client.GetAllJobInfo - 1 total (0 active), Execution time: mean = 1.01ms, total = 1.01ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
217
+ DebugString() time ms: 1
gcs_server.err ADDED
File without changes
log_monitor.err ADDED
File without changes
monitor.out ADDED
File without changes
python-core-driver-01000000ffffffffffffffffffffffffffffffffffffffffffffffff_10593.log ADDED
@@ -0,0 +1,159 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2026-02-27 00:30:31,542 I 10593 10593] core_worker_process.cc:773: Constructing CoreWorkerProcess. pid: 10593
2
+ [2026-02-27 00:30:31,548 I 10593 10593] event.cc:499: Ray Event initialized for CORE_WORKER
3
+ [2026-02-27 00:30:31,549 I 10593 10593] event.cc:499: Ray Event initialized for EXPORT_TASK
4
+ [2026-02-27 00:30:31,549 I 10593 10593] event.cc:332: Set ray event level to warning
5
+ [2026-02-27 00:30:31,549 I 10593 10593] event_aggregator_client.h:50: Initiating the local event aggregator client with port: 55678
6
+ [2026-02-27 00:30:32,738 I 10593 10593] grpc_server.cc:143: driver server started, listening on port 50383.
7
+ [2026-02-27 00:30:32,748 I 10593 10593] core_worker_process.cc:261: Initializing worker at address: 10.128.0.163:50383 worker_id=01000000ffffffffffffffffffffffffffffffffffffffffffffffff node_id=d9160dc27b026e787f1f81b4465a36f139cf00f32ff2820e42f20120
8
+ [2026-02-27 00:30:32,749 I 10593 10593] task_event_buffer.cc:480: Reporting task events to GCS every 1000ms.
9
+ [2026-02-27 00:30:32,751 I 10593 11370] core_worker.cc:455: Event stats:
10
+
11
+
12
+ Global stats: 6 total (5 active)
13
+ Queueing time: mean = 0.01ms, max = 0.09ms, min = 0.09ms, total = 0.09ms
14
+ Execution time: mean = 0.00ms, total = 0.02ms
15
+ Event stats:
16
+ PeriodicalRunner.RunFnPeriodically - 2 total (1 active, 1 running), Execution time: mean = 0.01ms, total = 0.02ms, Queueing time: mean = 0.04ms, max = 0.09ms, min = 0.09ms, total = 0.09ms
17
+ ray::rpc::InternalPubSubGcsService.grpc_client.GcsSubscriberPoll - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
18
+ Publisher.CheckDeadSubscribers - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
19
+ ray::rpc::InternalPubSubGcsService.grpc_client.GcsSubscriberCommandBatch - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
20
+ ray::rpc::WorkerInfoGcsService.grpc_client.AddWorkerInfo - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
21
+
22
+ -----------------
23
+ Task execution event stats:
24
+
25
+ Global stats: 0 total (0 active)
26
+ Queueing time: mean = -nanms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
27
+ Execution time: mean = -nanms, total = 0.00ms
28
+ Event stats:
29
+
30
+ -----------------
31
+ Task Event stats:
32
+
33
+ IO Service Stats:
34
+
35
+ Global stats: 4 total (1 active)
36
+ Queueing time: mean = 0.01ms, max = 0.02ms, min = 0.01ms, total = 0.03ms
37
+ Execution time: mean = 0.27ms, total = 1.07ms
38
+ Event stats:
39
+ ray::rpc::TaskInfoGcsService.grpc_client.AddTaskEventData.OnReplyReceived - 1 total (0 active), Execution time: mean = 0.02ms, total = 0.02ms, Queueing time: mean = 0.02ms, max = 0.02ms, min = 0.02ms, total = 0.02ms
40
+ ray::rpc::TaskInfoGcsService.grpc_client.AddTaskEventData - 1 total (0 active), Execution time: mean = 0.71ms, total = 0.71ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
41
+ PeriodicalRunner.RunFnPeriodically - 1 total (0 active), Execution time: mean = 0.33ms, total = 0.33ms, Queueing time: mean = 0.01ms, max = 0.01ms, min = 0.01ms, total = 0.01ms
42
+ CoreWorker.deadline_timer.flush_task_events - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
43
+ Other Stats:
44
+ gcs_grpc_in_progress:0
45
+ event_aggregator_grpc_in_progress:0
46
+ current number of task status events in buffer: 1
47
+ current number of profile events in buffer: 0
48
+ current number of dropped task attempts tracked: 0
49
+ total task events sent: 0 MiB
50
+ total number of task attempts sent: 0
51
+ total number of task attempts dropped reported: 0
52
+ total number of sent failure: 0
53
+ num status task events dropped: 0
54
+ num profile task events dropped: 0
55
+ num ray task events reported to aggregator: 0
56
+ num ray task events failed to report to aggregator: 0
57
+ num of task attempts dropped reported to aggregator: 0
58
+ num of failed requests to aggregator: 0
59
+
60
+ [2026-02-27 00:30:32,752 I 10593 11370] accessor.cc:540: Received address and liveness notification for node, IsAlive = 1 node_id=d9160dc27b026e787f1f81b4465a36f139cf00f32ff2820e42f20120
61
+ [2026-02-27 00:30:32,752 I 10593 11370] normal_task_submitter.cc:824: Number of alive nodes:1
62
+ [2026-02-27 00:30:32,752 I 10593 10593] metrics_agent_client.cc:42: Initializing exporter ...
63
+ [2026-02-27 00:30:32,866 I 10593 10593] actor_task_submitter.cc:74: Set actor max pending calls to -1 actor_id=0f47a291c83181c09abd27b101000000
64
+ [2026-02-27 00:30:32,882 I 10593 11370] actor_manager.cc:236: received notification on actor, state: PENDING_CREATION, ip address: , port: 0, num_restarts: 0, death context type=CONTEXT_NOT_SET actor_id=0f47a291c83181c09abd27b101000000 worker_id=NIL_ID node_id=d9160dc27b026e787f1f81b4465a36f139cf00f32ff2820e42f20120
65
+ [2026-02-27 00:30:34,175 I 10593 11370] metrics_agent_client.cc:54: Exporter initialized.
66
+ [2026-02-27 00:30:50,695 I 10593 11370] actor_manager.cc:236: received notification on actor, state: ALIVE, ip address: 10.128.0.163, port: 50157, num_restarts: 0, death context type=CONTEXT_NOT_SET actor_id=0f47a291c83181c09abd27b101000000 worker_id=8d27d27b6a5c820150d6a54cd27fe296fd0409567d2b4685b9a84fc8 node_id=d9160dc27b026e787f1f81b4465a36f139cf00f32ff2820e42f20120
67
+ [2026-02-27 00:31:32,751 I 10593 11370] core_worker.cc:455: Event stats:
68
+
69
+
70
+ Global stats: 842 total (9 active)
71
+ Queueing time: mean = 1.23ms, max = 1000.04ms, min = 0.01ms, total = 1039.49ms
72
+ Execution time: mean = 43.07ms, total = 36264.22ms
73
+ Event stats:
74
+ CoreWorker.RecoverObjects - 600 total (1 active), Execution time: mean = 0.01ms, total = 4.89ms, Queueing time: mean = 0.05ms, max = 1.80ms, min = 0.02ms, total = 30.30ms
75
+ NodeManagerService.grpc_client.ReportWorkerBacklog - 60 total (0 active), Execution time: mean = 0.70ms, total = 41.88ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
76
+ CoreWorker.InternalHeartbeat - 60 total (1 active), Execution time: mean = 0.16ms, total = 9.68ms, Queueing time: mean = 0.05ms, max = 0.43ms, min = 0.02ms, total = 2.86ms
77
+ NodeManagerService.grpc_client.ReportWorkerBacklog.OnReplyReceived - 60 total (0 active), Execution time: mean = 0.02ms, total = 1.30ms, Queueing time: mean = 0.03ms, max = 0.08ms, min = 0.01ms, total = 1.73ms
78
+ CoreWorker.RecordMetrics - 12 total (1 active), Execution time: mean = 0.13ms, total = 1.58ms, Queueing time: mean = 0.04ms, max = 0.07ms, min = 0.02ms, total = 0.48ms
79
+ CoreWorker.TryDelPendingObjectRefStreams - 6 total (1 active), Execution time: mean = 0.01ms, total = 0.04ms, Queueing time: mean = 0.04ms, max = 0.07ms, min = 0.03ms, total = 0.25ms
80
+ PeriodicalRunner.RunFnPeriodically - 6 total (0 active), Execution time: mean = 0.13ms, total = 0.76ms, Queueing time: mean = 0.31ms, max = 0.68ms, min = 0.02ms, total = 1.86ms
81
+ CoreWorkerService.grpc_server.GetCoreWorkerStats.HandleRequestImpl - 4 total (0 active), Execution time: mean = 0.10ms, total = 0.40ms, Queueing time: mean = 0.08ms, max = 0.13ms, min = 0.02ms, total = 0.31ms
82
+ CoreWorkerService.grpc_server.GetCoreWorkerStats - 4 total (0 active), Execution time: mean = 0.35ms, total = 1.42ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
83
+ ray::rpc::InternalPubSubGcsService.grpc_client.GcsSubscriberCommandBatch - 2 total (0 active), Execution time: mean = 0.88ms, total = 1.75ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
84
+ ray::rpc::InternalPubSubGcsService.grpc_client.GcsSubscriberPoll - 2 total (1 active), Execution time: mean = 8971.98ms, total = 17943.96ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
85
+ ReporterService.grpc_client.HealthCheck.OnReplyReceived - 2 total (0 active), Execution time: mean = 0.17ms, total = 0.34ms, Queueing time: mean = 0.02ms, max = 0.02ms, min = 0.02ms, total = 0.04ms
86
+ ReporterService.grpc_client.HealthCheck - 2 total (0 active), Execution time: mean = 211.28ms, total = 422.55ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
87
+ ray::rpc::InternalPubSubGcsService.grpc_client.GcsSubscriberCommandBatch.OnReplyReceived - 2 total (0 active), Execution time: mean = 0.16ms, total = 0.32ms, Queueing time: mean = 0.18ms, max = 0.32ms, min = 0.04ms, total = 0.36ms
88
+ ray::rpc::WorkerInfoGcsService.grpc_client.AddWorkerInfo - 1 total (0 active), Execution time: mean = 1.08ms, total = 1.08ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
89
+ ActorTaskSubmitter::SubmitTask - 1 total (0 active), Execution time: mean = 0.01ms, total = 0.01ms, Queueing time: mean = 0.16ms, max = 0.16ms, min = 0.16ms, total = 0.16ms
90
+ ray::rpc::ActorInfoGcsService.grpc_client.GetActorInfo.OnReplyReceived - 1 total (0 active), Execution time: mean = 1.01ms, total = 1.01ms, Queueing time: mean = 0.04ms, max = 0.04ms, min = 0.04ms, total = 0.04ms
91
+ ray::rpc::ActorInfoGcsService.grpc_client.RegisterActor.OnReplyReceived - 1 total (0 active), Execution time: mean = 0.27ms, total = 0.27ms, Queueing time: mean = 0.07ms, max = 0.07ms, min = 0.07ms, total = 0.07ms
92
+ CoreWorker.PrintEventStats - 1 total (1 active, 1 running), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
93
+ ray::rpc::ActorInfoGcsService.grpc_client.CreateActor - 1 total (0 active), Execution time: mean = 17825.20ms, total = 17825.20ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
94
+ ActorCreator.AsyncRegisterActor - 1 total (0 active), Execution time: mean = 0.55ms, total = 0.55ms, Queueing time: mean = 0.03ms, max = 0.03ms, min = 0.03ms, total = 0.03ms
95
+ ray::rpc::WorkerInfoGcsService.grpc_client.AddWorkerInfo.OnReplyReceived - 1 total (0 active), Execution time: mean = 0.02ms, total = 0.02ms, Queueing time: mean = 0.28ms, max = 0.28ms, min = 0.28ms, total = 0.28ms
96
+ ray::rpc::ActorInfoGcsService.grpc_client.GetActorInfo - 1 total (0 active), Execution time: mean = 0.67ms, total = 0.67ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
97
+ ray::rpc::InternalPubSubGcsService.grpc_client.GcsSubscriberPoll.OnReplyReceived - 1 total (0 active), Execution time: mean = 0.47ms, total = 0.47ms, Queueing time: mean = 0.03ms, max = 0.03ms, min = 0.03ms, total = 0.03ms
98
+ ray::rpc::ActorInfoGcsService.grpc_client.CreateActor.OnReplyReceived - 1 total (0 active), Execution time: mean = 0.08ms, total = 0.08ms, Queueing time: mean = 0.34ms, max = 0.34ms, min = 0.34ms, total = 0.34ms
99
+ CoreWorkerService.grpc_server.WaitForActorRefDeleted - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
100
+ ray::rpc::NodeInfoGcsService.grpc_client.GetAllNodeAddressAndLiveness.OnReplyReceived - 1 total (0 active), Execution time: mean = 0.08ms, total = 0.08ms, Queueing time: mean = 0.02ms, max = 0.02ms, min = 0.02ms, total = 0.02ms
101
+ Publisher.CheckDeadSubscribers - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
102
+ Subscriber.HandlePublishedMessage_GCS_ACTOR_CHANNEL - 1 total (0 active), Execution time: mean = 0.79ms, total = 0.79ms, Queueing time: mean = 0.27ms, max = 0.27ms, min = 0.27ms, total = 0.27ms
103
+ ray::rpc::ActorInfoGcsService.grpc_client.RegisterActor - 1 total (0 active), Execution time: mean = 1.97ms, total = 1.97ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
104
+ CoreWorkerService.grpc_client.PushTask - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
105
+ ray::rpc::NodeInfoGcsService.grpc_client.GetAllNodeAddressAndLiveness - 1 total (0 active), Execution time: mean = 0.94ms, total = 0.94ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
106
+ CoreWorkerService.grpc_server.WaitForActorRefDeleted.HandleRequestImpl - 1 total (0 active), Execution time: mean = 0.06ms, total = 0.06ms, Queueing time: mean = 0.04ms, max = 0.04ms, min = 0.04ms, total = 0.04ms
107
+ MetricsAgentClient.WaitForServerReadyWithRetry - 1 total (0 active), Execution time: mean = 0.15ms, total = 0.15ms, Queueing time: mean = 1000.04ms, max = 1000.04ms, min = 1000.04ms, total = 1000.04ms
108
+
109
+ -----------------
110
+ Task execution event stats:
111
+
112
+ Global stats: 0 total (0 active)
113
+ Queueing time: mean = -nanms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
114
+ Execution time: mean = -nanms, total = 0.00ms
115
+ Event stats:
116
+
117
+ -----------------
118
+ Task Event stats:
119
+
120
+ IO Service Stats:
121
+
122
+ Global stats: 181 total (1 active)
123
+ Queueing time: mean = 0.03ms, max = 1.57ms, min = 0.01ms, total = 5.85ms
124
+ Execution time: mean = 0.30ms, total = 54.50ms
125
+ Event stats:
126
+ ray::rpc::TaskInfoGcsService.grpc_client.AddTaskEventData.OnReplyReceived - 60 total (0 active), Execution time: mean = 0.04ms, total = 2.19ms, Queueing time: mean = 0.03ms, max = 0.05ms, min = 0.02ms, total = 1.65ms
127
+ ray::rpc::TaskInfoGcsService.grpc_client.AddTaskEventData - 60 total (0 active), Execution time: mean = 0.68ms, total = 40.52ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
128
+ CoreWorker.deadline_timer.flush_task_events - 60 total (1 active), Execution time: mean = 0.19ms, total = 11.46ms, Queueing time: mean = 0.07ms, max = 1.57ms, min = 0.03ms, total = 4.20ms
129
+ PeriodicalRunner.RunFnPeriodically - 1 total (0 active), Execution time: mean = 0.33ms, total = 0.33ms, Queueing time: mean = 0.01ms, max = 0.01ms, min = 0.01ms, total = 0.01ms
130
+ Other Stats:
131
+ gcs_grpc_in_progress:0
132
+ event_aggregator_grpc_in_progress:0
133
+ current number of task status events in buffer: 0
134
+ current number of profile events in buffer: 0
135
+ current number of dropped task attempts tracked: 0
136
+ total task events sent: 0.00146484 MiB
137
+ total number of task attempts sent: 5
138
+ total number of task attempts dropped reported: 0
139
+ total number of sent failure: 0
140
+ num status task events dropped: 0
141
+ num profile task events dropped: 0
142
+ num ray task events reported to aggregator: 0
143
+ num ray task events failed to report to aggregator: 0
144
+ num of task attempts dropped reported to aggregator: 0
145
+ num of failed requests to aggregator: 0
146
+
147
+ [2026-02-27 00:32:12,445 I 10593 10593] core_worker.cc:584: Sending disconnect message to the local raylet.
148
+ [2026-02-27 00:32:12,445 I 10593 10593] raylet_ipc_client.cc:135: RayletIpcClient::Disconnect, exit_type=INTENDED_USER_EXIT, exit_detail=Shutdown by ray.shutdown()., has creation_task_exception_pb_bytes=0
149
+ [2026-02-27 00:32:12,446 I 10593 10593] core_worker.cc:589: Disconnected from the local raylet.
150
+ [2026-02-27 00:32:12,446 I 10593 10593] task_event_buffer.cc:491: Shutting down TaskEventBuffer.
151
+ [2026-02-27 00:32:12,446 I 10593 11747] task_event_buffer.cc:459: Task event buffer io service stopped.
152
+ [2026-02-27 00:32:12,446 I 10593 10593] core_worker_shutdown_executor.cc:54: Waiting for joining a core worker io thread. If it hangs here, there might be deadlock or a high load in the core worker io service.
153
+ [2026-02-27 00:32:12,446 I 10593 11370] core_worker_process.cc:194: Core worker main io service stopped.
154
+ [2026-02-27 00:32:12,464 I 10593 10593] core_worker_shutdown_executor.cc:72: Disconnecting a GCS client.
155
+ [2026-02-27 00:32:12,464 I 10593 10593] core_worker_shutdown_executor.cc:79: Core worker ready to be deallocated.
156
+ [2026-02-27 00:32:12,464 I 10593 10593] core_worker.cc:539: Core worker is destructed
157
+ [2026-02-27 00:32:12,464 I 10593 10593] task_event_buffer.cc:491: Shutting down TaskEventBuffer.
158
+ [2026-02-27 00:32:12,478 I 10593 10593] core_worker_process.cc:846: Destructing CoreWorkerProcessImpl. pid: 10593
159
+ [2026-02-27 00:32:12,488 I 10593 10593] stats.h:149: Stats module has shutdown.
python-core-worker-15c410d5d6a75625cb50c80927d18090e899b8edc49402fe08e50ee6_12567.log ADDED
@@ -0,0 +1,71 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2026-02-27 00:31:33,371 I 12567 12567] core_worker_process.cc:773: Constructing CoreWorkerProcess. pid: 12567
2
+ [2026-02-27 00:31:33,383 I 12567 12567] event.cc:499: Ray Event initialized for CORE_WORKER
3
+ [2026-02-27 00:31:33,383 I 12567 12567] event.cc:499: Ray Event initialized for EXPORT_TASK
4
+ [2026-02-27 00:31:33,383 I 12567 12567] event.cc:332: Set ray event level to warning
5
+ [2026-02-27 00:31:33,383 I 12567 12567] event_aggregator_client.h:50: Initiating the local event aggregator client with port: 55678
6
+ [2026-02-27 00:31:33,385 I 12567 12567] grpc_server.cc:143: worker server started, listening on port 50147.
7
+ [2026-02-27 00:31:33,405 I 12567 12567] core_worker_process.cc:261: Initializing worker at address: 10.128.0.163:50147 worker_id=15c410d5d6a75625cb50c80927d18090e899b8edc49402fe08e50ee6 node_id=d9160dc27b026e787f1f81b4465a36f139cf00f32ff2820e42f20120
8
+ [2026-02-27 00:31:33,406 I 12567 12567] task_event_buffer.cc:480: Reporting task events to GCS every 1000ms.
9
+ [2026-02-27 00:31:33,407 I 12567 12913] core_worker.cc:455: Event stats:
10
+
11
+
12
+ Global stats: 8 total (6 active)
13
+ Queueing time: mean = 0.01ms, max = 0.04ms, min = 0.02ms, total = 0.06ms
14
+ Execution time: mean = 0.00ms, total = 0.04ms
15
+ Event stats:
16
+ PeriodicalRunner.RunFnPeriodically - 3 total (1 active, 1 running), Execution time: mean = 0.01ms, total = 0.04ms, Queueing time: mean = 0.02ms, max = 0.04ms, min = 0.02ms, total = 0.06ms
17
+ Publisher.CheckDeadSubscribers - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
18
+ ray::rpc::InternalPubSubGcsService.grpc_client.GcsSubscriberCommandBatch - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
19
+ CoreWorker.ExitIfParentRayletDies - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
20
+ ray::rpc::InternalPubSubGcsService.grpc_client.GcsSubscriberPoll - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
21
+ ray::rpc::WorkerInfoGcsService.grpc_client.AddWorkerInfo - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
22
+
23
+ -----------------
24
+ Task execution event stats:
25
+
26
+ Global stats: 0 total (0 active)
27
+ Queueing time: mean = -nanms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
28
+ Execution time: mean = -nanms, total = 0.00ms
29
+ Event stats:
30
+
31
+ -----------------
32
+ Task Event stats:
33
+
34
+ IO Service Stats:
35
+
36
+ Global stats: 2 total (2 active)
37
+ Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
38
+ Execution time: mean = 0.00ms, total = 0.00ms
39
+ Event stats:
40
+ PeriodicalRunner.RunFnPeriodically - 1 total (1 active, 1 running), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
41
+ ray::rpc::TaskInfoGcsService.grpc_client.AddTaskEventData - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
42
+ Other Stats:
43
+ gcs_grpc_in_progress:1
44
+ event_aggregator_grpc_in_progress:0
45
+ current number of task status events in buffer: 0
46
+ current number of profile events in buffer: 0
47
+ current number of dropped task attempts tracked: 0
48
+ total task events sent: 0 MiB
49
+ total number of task attempts sent: 0
50
+ total number of task attempts dropped reported: 0
51
+ total number of sent failure: 0
52
+ num status task events dropped: 0
53
+ num profile task events dropped: 0
54
+ num ray task events reported to aggregator: 0
55
+ num ray task events failed to report to aggregator: 0
56
+ num of task attempts dropped reported to aggregator: 0
57
+ num of failed requests to aggregator: 0
58
+
59
+ [2026-02-27 00:31:33,409 I 12567 12567] core_worker.cc:515: Adjusted worker niceness to 15
60
+ [2026-02-27 00:31:33,410 I 12567 12913] accessor.cc:540: Received address and liveness notification for node, IsAlive = 1 node_id=d9160dc27b026e787f1f81b4465a36f139cf00f32ff2820e42f20120
61
+ [2026-02-27 00:31:33,411 I 12567 12913] normal_task_submitter.cc:824: Number of alive nodes:1
62
+ [2026-02-27 00:31:33,411 I 12567 12567] metrics_agent_client.cc:42: Initializing exporter ...
63
+ [2026-02-27 00:31:33,415 I 12567 12913] metrics_agent_client.cc:54: Exporter initialized.
64
+ [2026-02-27 00:31:33,417 I 12567 12567] actor_task_submitter.cc:74: Set actor max pending calls to -1 actor_id=ae31a85bdd5e21bfaec5bc9e01000000
65
+ [2026-02-27 00:31:33,417 I 12567 12567] core_worker.cc:2903: Creating actor actor_id=ae31a85bdd5e21bfaec5bc9e01000000
66
+ [2026-02-27 00:31:44,749 I 12567 12567] task_receiver.cc:142: Actor creation task finished, task_id: ffffffffffffffffae31a85bdd5e21bfaec5bc9e01000000, actor_id: ae31a85bdd5e21bfaec5bc9e01000000, actor_repr_name:
67
+ [2026-02-27 00:32:12,483 I 12567 12913] core_worker_shutdown_executor.cc:217: Try killing all child processes of this worker as it exits. Child process pids:
68
+ [2026-02-27 00:32:12,483 I 12567 12913] core_worker_shutdown_executor.cc:262: Sending disconnect message to the local raylet.
69
+ [2026-02-27 00:32:12,483 I 12567 12913] raylet_ipc_client.cc:135: RayletIpcClient::Disconnect, exit_type=INTENDED_USER_EXIT, exit_detail=Worker force exited because its job has finished, has creation_task_exception_pb_bytes=0
70
+ [2026-02-27 00:32:12,490 I 12567 12913] core_worker_shutdown_executor.cc:279: Disconnected from the local raylet.
71
+ [2026-02-27 00:32:12,490 W 12567 12913] core_worker_shutdown_executor.cc:288: Quick exit - terminating process immediately
python-core-worker-33b9d0a21a51ca22dda2aa2142cb264d1ee4f9d53a55dc567b49496c_12500.log ADDED
@@ -0,0 +1,71 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2026-02-27 00:31:33,542 I 12500 12500] core_worker_process.cc:773: Constructing CoreWorkerProcess. pid: 12500
2
+ [2026-02-27 00:31:33,548 I 12500 12500] event.cc:499: Ray Event initialized for CORE_WORKER
3
+ [2026-02-27 00:31:33,548 I 12500 12500] event.cc:499: Ray Event initialized for EXPORT_TASK
4
+ [2026-02-27 00:31:33,548 I 12500 12500] event.cc:332: Set ray event level to warning
5
+ [2026-02-27 00:31:33,548 I 12500 12500] event_aggregator_client.h:50: Initiating the local event aggregator client with port: 55678
6
+ [2026-02-27 00:31:33,550 I 12500 12500] grpc_server.cc:143: worker server started, listening on port 50457.
7
+ [2026-02-27 00:31:33,565 I 12500 12500] core_worker_process.cc:261: Initializing worker at address: 10.128.0.163:50457 worker_id=33b9d0a21a51ca22dda2aa2142cb264d1ee4f9d53a55dc567b49496c node_id=d9160dc27b026e787f1f81b4465a36f139cf00f32ff2820e42f20120
8
+ [2026-02-27 00:31:33,566 I 12500 12500] task_event_buffer.cc:480: Reporting task events to GCS every 1000ms.
9
+ [2026-02-27 00:31:33,567 I 12500 12500] core_worker.cc:515: Adjusted worker niceness to 15
10
+ [2026-02-27 00:31:33,568 I 12500 12500] metrics_agent_client.cc:42: Initializing exporter ...
11
+ [2026-02-27 00:31:33,567 I 12500 12987] core_worker.cc:455: Event stats:
12
+
13
+
14
+ Global stats: 12 total (10 active)
15
+ Queueing time: mean = 0.00ms, max = 0.03ms, min = 0.02ms, total = 0.06ms
16
+ Execution time: mean = 0.00ms, total = 0.04ms
17
+ Event stats:
18
+ PeriodicalRunner.RunFnPeriodically - 7 total (5 active, 1 running), Execution time: mean = 0.01ms, total = 0.04ms, Queueing time: mean = 0.01ms, max = 0.03ms, min = 0.02ms, total = 0.06ms
19
+ Publisher.CheckDeadSubscribers - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
20
+ ray::rpc::InternalPubSubGcsService.grpc_client.GcsSubscriberCommandBatch - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
21
+ ray::rpc::InternalPubSubGcsService.grpc_client.GcsSubscriberPoll - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
22
+ CoreWorker.ExitIfParentRayletDies - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
23
+ ray::rpc::WorkerInfoGcsService.grpc_client.AddWorkerInfo - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
24
+
25
+ -----------------
26
+ Task execution event stats:
27
+
28
+ Global stats: 0 total (0 active)
29
+ Queueing time: mean = -nanms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
30
+ Execution time: mean = -nanms, total = 0.00ms
31
+ Event stats:
32
+
33
+ -----------------
34
+ Task Event stats:
35
+
36
+ IO Service Stats:
37
+
38
+ Global stats: 2 total (2 active)
39
+ Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
40
+ Execution time: mean = 0.00ms, total = 0.00ms
41
+ Event stats:
42
+ PeriodicalRunner.RunFnPeriodically - 1 total (1 active, 1 running), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
43
+ ray::rpc::TaskInfoGcsService.grpc_client.AddTaskEventData - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
44
+ Other Stats:
45
+ gcs_grpc_in_progress:1
46
+ event_aggregator_grpc_in_progress:0
47
+ current number of task status events in buffer: 0
48
+ current number of profile events in buffer: 0
49
+ current number of dropped task attempts tracked: 0
50
+ total task events sent: 0 MiB
51
+ total number of task attempts sent: 0
52
+ total number of task attempts dropped reported: 0
53
+ total number of sent failure: 0
54
+ num status task events dropped: 0
55
+ num profile task events dropped: 0
56
+ num ray task events reported to aggregator: 0
57
+ num ray task events failed to report to aggregator: 0
58
+ num of task attempts dropped reported to aggregator: 0
59
+ num of failed requests to aggregator: 0
60
+
61
+ [2026-02-27 00:31:33,572 I 12500 12987] accessor.cc:540: Received address and liveness notification for node, IsAlive = 1 node_id=d9160dc27b026e787f1f81b4465a36f139cf00f32ff2820e42f20120
62
+ [2026-02-27 00:31:33,573 I 12500 12987] normal_task_submitter.cc:824: Number of alive nodes:1
63
+ [2026-02-27 00:31:33,574 I 12500 12987] metrics_agent_client.cc:54: Exporter initialized.
64
+ [2026-02-27 00:31:33,578 I 12500 12500] actor_task_submitter.cc:74: Set actor max pending calls to -1 actor_id=1b6a74cdbf74d457007f760101000000
65
+ [2026-02-27 00:31:33,578 I 12500 12500] core_worker.cc:2903: Creating actor actor_id=1b6a74cdbf74d457007f760101000000
66
+ [2026-02-27 00:31:45,005 I 12500 12500] task_receiver.cc:142: Actor creation task finished, task_id: ffffffffffffffff1b6a74cdbf74d457007f760101000000, actor_id: 1b6a74cdbf74d457007f760101000000, actor_repr_name:
67
+ [2026-02-27 00:32:12,479 I 12500 12987] core_worker_shutdown_executor.cc:217: Try killing all child processes of this worker as it exits. Child process pids:
68
+ [2026-02-27 00:32:12,479 I 12500 12987] core_worker_shutdown_executor.cc:262: Sending disconnect message to the local raylet.
69
+ [2026-02-27 00:32:12,479 I 12500 12987] raylet_ipc_client.cc:135: RayletIpcClient::Disconnect, exit_type=INTENDED_USER_EXIT, exit_detail=Worker force exited because its job has finished, has creation_task_exception_pb_bytes=0
70
+ [2026-02-27 00:32:12,485 I 12500 12987] core_worker_shutdown_executor.cc:279: Disconnected from the local raylet.
71
+ [2026-02-27 00:32:12,485 W 12500 12987] core_worker_shutdown_executor.cc:288: Quick exit - terminating process immediately
python-core-worker-389d4ca43c5eadc5290ba2907f911210cffe11839a5cfe9496d636c1_12110.log ADDED
@@ -0,0 +1,142 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2026-02-27 00:30:59,145 I 12110 12110] core_worker_process.cc:773: Constructing CoreWorkerProcess. pid: 12110
2
+ [2026-02-27 00:30:59,148 I 12110 12110] event.cc:499: Ray Event initialized for CORE_WORKER
3
+ [2026-02-27 00:30:59,148 I 12110 12110] event.cc:499: Ray Event initialized for EXPORT_TASK
4
+ [2026-02-27 00:30:59,148 I 12110 12110] event.cc:332: Set ray event level to warning
5
+ [2026-02-27 00:30:59,148 I 12110 12110] event_aggregator_client.h:50: Initiating the local event aggregator client with port: 55678
6
+ [2026-02-27 00:30:59,150 I 12110 12110] grpc_server.cc:143: worker server started, listening on port 50103.
7
+ [2026-02-27 00:30:59,164 I 12110 12110] core_worker_process.cc:261: Initializing worker at address: 10.128.0.163:50103 worker_id=389d4ca43c5eadc5290ba2907f911210cffe11839a5cfe9496d636c1 node_id=d9160dc27b026e787f1f81b4465a36f139cf00f32ff2820e42f20120
8
+ [2026-02-27 00:30:59,165 I 12110 12110] task_event_buffer.cc:480: Reporting task events to GCS every 1000ms.
9
+ [2026-02-27 00:30:59,166 I 12110 12110] core_worker.cc:515: Adjusted worker niceness to 15
10
+ [2026-02-27 00:30:59,166 I 12110 12110] metrics_agent_client.cc:42: Initializing exporter ...
11
+ [2026-02-27 00:30:59,166 I 12110 12152] core_worker.cc:455: Event stats:
12
+
13
+
14
+ Global stats: 12 total (10 active)
15
+ Queueing time: mean = 0.00ms, max = 0.03ms, min = 0.02ms, total = 0.06ms
16
+ Execution time: mean = 0.00ms, total = 0.04ms
17
+ Event stats:
18
+ PeriodicalRunner.RunFnPeriodically - 7 total (5 active, 1 running), Execution time: mean = 0.01ms, total = 0.04ms, Queueing time: mean = 0.01ms, max = 0.03ms, min = 0.02ms, total = 0.06ms
19
+ CoreWorker.ExitIfParentRayletDies - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
20
+ ray::rpc::InternalPubSubGcsService.grpc_client.GcsSubscriberCommandBatch - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
21
+ ray::rpc::InternalPubSubGcsService.grpc_client.GcsSubscriberPoll - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
22
+ ray::rpc::WorkerInfoGcsService.grpc_client.AddWorkerInfo - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
23
+ Publisher.CheckDeadSubscribers - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
24
+
25
+ -----------------
26
+ Task execution event stats:
27
+
28
+ Global stats: 0 total (0 active)
29
+ Queueing time: mean = -nanms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
30
+ Execution time: mean = -nanms, total = 0.00ms
31
+ Event stats:
32
+
33
+ -----------------
34
+ Task Event stats:
35
+
36
+ IO Service Stats:
37
+
38
+ Global stats: 4 total (1 active)
39
+ Queueing time: mean = 0.01ms, max = 0.02ms, min = 0.02ms, total = 0.04ms
40
+ Execution time: mean = 0.20ms, total = 0.79ms
41
+ Event stats:
42
+ PeriodicalRunner.RunFnPeriodically - 1 total (0 active), Execution time: mean = 0.22ms, total = 0.22ms, Queueing time: mean = 0.02ms, max = 0.02ms, min = 0.02ms, total = 0.02ms
43
+ ray::rpc::TaskInfoGcsService.grpc_client.AddTaskEventData.OnReplyReceived - 1 total (0 active), Execution time: mean = 0.03ms, total = 0.03ms, Queueing time: mean = 0.02ms, max = 0.02ms, min = 0.02ms, total = 0.02ms
44
+ ray::rpc::TaskInfoGcsService.grpc_client.AddTaskEventData - 1 total (0 active), Execution time: mean = 0.54ms, total = 0.54ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
45
+ CoreWorker.deadline_timer.flush_task_events - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
46
+ Other Stats:
47
+ gcs_grpc_in_progress:0
48
+ event_aggregator_grpc_in_progress:0
49
+ current number of task status events in buffer: 0
50
+ current number of profile events in buffer: 0
51
+ current number of dropped task attempts tracked: 0
52
+ total task events sent: 0 MiB
53
+ total number of task attempts sent: 0
54
+ total number of task attempts dropped reported: 0
55
+ total number of sent failure: 0
56
+ num status task events dropped: 0
57
+ num profile task events dropped: 0
58
+ num ray task events reported to aggregator: 0
59
+ num ray task events failed to report to aggregator: 0
60
+ num of task attempts dropped reported to aggregator: 0
61
+ num of failed requests to aggregator: 0
62
+
63
+ [2026-02-27 00:30:59,167 I 12110 12152] accessor.cc:540: Received address and liveness notification for node, IsAlive = 1 node_id=d9160dc27b026e787f1f81b4465a36f139cf00f32ff2820e42f20120
64
+ [2026-02-27 00:30:59,167 I 12110 12152] normal_task_submitter.cc:824: Number of alive nodes:1
65
+ [2026-02-27 00:30:59,168 I 12110 12152] metrics_agent_client.cc:54: Exporter initialized.
66
+ [2026-02-27 00:31:59,166 I 12110 12152] core_worker.cc:455: Event stats:
67
+
68
+
69
+ Global stats: 888 total (8 active)
70
+ Queueing time: mean = 0.07ms, max = 3.72ms, min = -0.02ms, total = 65.62ms
71
+ Execution time: mean = 0.37ms, total = 330.46ms
72
+ Event stats:
73
+ CoreWorker.RecoverObjects - 600 total (1 active), Execution time: mean = 0.01ms, total = 4.38ms, Queueing time: mean = 0.09ms, max = 3.72ms, min = -0.02ms, total = 53.86ms
74
+ CoreWorker.InternalHeartbeat - 60 total (1 active), Execution time: mean = 0.16ms, total = 9.34ms, Queueing time: mean = 0.05ms, max = 0.63ms, min = 0.02ms, total = 2.89ms
75
+ NodeManagerService.grpc_client.ReportWorkerBacklog - 60 total (0 active), Execution time: mean = 0.86ms, total = 51.35ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
76
+ CoreWorker.ExitIfParentRayletDies - 60 total (1 active), Execution time: mean = 0.01ms, total = 0.59ms, Queueing time: mean = 0.04ms, max = 0.12ms, min = 0.02ms, total = 2.56ms
77
+ NodeManagerService.grpc_client.ReportWorkerBacklog.OnReplyReceived - 60 total (0 active), Execution time: mean = 0.02ms, total = 1.37ms, Queueing time: mean = 0.04ms, max = 0.37ms, min = 0.02ms, total = 2.30ms
78
+ CoreWorker.RecordMetrics - 12 total (1 active), Execution time: mean = 0.15ms, total = 1.80ms, Queueing time: mean = 0.03ms, max = 0.07ms, min = 0.02ms, total = 0.40ms
79
+ PeriodicalRunner.RunFnPeriodically - 7 total (0 active), Execution time: mean = 0.10ms, total = 0.69ms, Queueing time: mean = 0.29ms, max = 0.69ms, min = 0.02ms, total = 2.03ms
80
+ CoreWorker.TryDelPendingObjectRefStreams - 6 total (1 active), Execution time: mean = 0.01ms, total = 0.04ms, Queueing time: mean = 0.03ms, max = 0.05ms, min = 0.02ms, total = 0.19ms
81
+ CoreWorkerService.grpc_server.GetCoreWorkerStats - 4 total (0 active), Execution time: mean = 0.50ms, total = 2.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
82
+ CoreWorkerService.grpc_server.GetCoreWorkerStats.HandleRequestImpl - 4 total (0 active), Execution time: mean = 0.06ms, total = 0.25ms, Queueing time: mean = 0.15ms, max = 0.33ms, min = 0.04ms, total = 0.59ms
83
+ CoreWorkerService.grpc_server.PushTask - 2 total (0 active), Execution time: mean = 126.87ms, total = 253.74ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
84
+ CoreWorkerService.grpc_server.PushTask.HandleRequestImpl - 2 total (0 active), Execution time: mean = 0.14ms, total = 0.27ms, Queueing time: mean = 0.03ms, max = 0.04ms, min = 0.03ms, total = 0.07ms
85
+ ray::rpc::WorkerInfoGcsService.grpc_client.AddWorkerInfo - 1 total (0 active), Execution time: mean = 0.67ms, total = 0.67ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
86
+ ReporterService.grpc_client.HealthCheck - 1 total (0 active), Execution time: mean = 2.16ms, total = 2.16ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
87
+ ReporterService.grpc_client.HealthCheck.OnReplyReceived - 1 total (0 active), Execution time: mean = 0.27ms, total = 0.27ms, Queueing time: mean = 0.03ms, max = 0.03ms, min = 0.03ms, total = 0.03ms
88
+ ray::rpc::WorkerInfoGcsService.grpc_client.AddWorkerInfo.OnReplyReceived - 1 total (0 active), Execution time: mean = 0.02ms, total = 0.02ms, Queueing time: mean = 0.43ms, max = 0.43ms, min = 0.43ms, total = 0.43ms
89
+ CoreWorker.PrintEventStats - 1 total (1 active, 1 running), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
90
+ ray::rpc::InternalPubSubGcsService.grpc_client.GcsSubscriberCommandBatch.OnReplyReceived - 1 total (0 active), Execution time: mean = 0.16ms, total = 0.16ms, Queueing time: mean = 0.25ms, max = 0.25ms, min = 0.25ms, total = 0.25ms
91
+ ray::rpc::NodeInfoGcsService.grpc_client.GetAllNodeAddressAndLiveness - 1 total (0 active), Execution time: mean = 0.55ms, total = 0.55ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
92
+ Publisher.CheckDeadSubscribers - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
93
+ ray::rpc::InternalPubSubGcsService.grpc_client.GcsSubscriberPoll - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
94
+ ray::rpc::NodeInfoGcsService.grpc_client.GetAllNodeAddressAndLiveness.OnReplyReceived - 1 total (0 active), Execution time: mean = 0.08ms, total = 0.08ms, Queueing time: mean = 0.02ms, max = 0.02ms, min = 0.02ms, total = 0.02ms
95
+ ray::rpc::InternalPubSubGcsService.grpc_client.GcsSubscriberCommandBatch - 1 total (0 active), Execution time: mean = 0.73ms, total = 0.73ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
96
+
97
+ -----------------
98
+ Task execution event stats:
99
+
100
+ Global stats: 5876 total (1 active)
101
+ Queueing time: mean = 0.21ms, max = 250.72ms, min = -0.04ms, total = 1231.40ms
102
+ Execution time: mean = 0.05ms, total = 314.23ms
103
+ Event stats:
104
+ CoreWorker.CheckSignal - 5873 total (1 active), Execution time: mean = 0.01ms, total = 61.20ms, Queueing time: mean = 0.17ms, max = 41.01ms, min = -0.04ms, total = 980.63ms
105
+ CoreWorker.HandlePushTask - 2 total (0 active), Execution time: mean = 126.51ms, total = 253.01ms, Queueing time: mean = 0.03ms, max = 0.05ms, min = 0.01ms, total = 0.06ms
106
+ PeriodicalRunner.RunFnPeriodically - 1 total (0 active), Execution time: mean = 0.01ms, total = 0.01ms, Queueing time: mean = 250.72ms, max = 250.72ms, min = 250.72ms, total = 250.72ms
107
+
108
+ -----------------
109
+ Task Event stats:
110
+
111
+ IO Service Stats:
112
+
113
+ Global stats: 181 total (1 active)
114
+ Queueing time: mean = 0.07ms, max = 2.96ms, min = 0.02ms, total = 13.31ms
115
+ Execution time: mean = 0.31ms, total = 57.01ms
116
+ Event stats:
117
+ ray::rpc::TaskInfoGcsService.grpc_client.AddTaskEventData.OnReplyReceived - 60 total (0 active), Execution time: mean = 0.03ms, total = 2.03ms, Queueing time: mean = 0.10ms, max = 2.07ms, min = 0.02ms, total = 5.96ms
118
+ ray::rpc::TaskInfoGcsService.grpc_client.AddTaskEventData - 60 total (0 active), Execution time: mean = 0.73ms, total = 43.83ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
119
+ CoreWorker.deadline_timer.flush_task_events - 60 total (1 active), Execution time: mean = 0.18ms, total = 10.93ms, Queueing time: mean = 0.12ms, max = 2.96ms, min = 0.03ms, total = 7.33ms
120
+ PeriodicalRunner.RunFnPeriodically - 1 total (0 active), Execution time: mean = 0.22ms, total = 0.22ms, Queueing time: mean = 0.02ms, max = 0.02ms, min = 0.02ms, total = 0.02ms
121
+ Other Stats:
122
+ gcs_grpc_in_progress:0
123
+ event_aggregator_grpc_in_progress:0
124
+ current number of task status events in buffer: 0
125
+ current number of profile events in buffer: 0
126
+ current number of dropped task attempts tracked: 0
127
+ total task events sent: 0.00138664 MiB
128
+ total number of task attempts sent: 2
129
+ total number of task attempts dropped reported: 0
130
+ total number of sent failure: 0
131
+ num status task events dropped: 0
132
+ num profile task events dropped: 0
133
+ num ray task events reported to aggregator: 0
134
+ num ray task events failed to report to aggregator: 0
135
+ num of task attempts dropped reported to aggregator: 0
136
+ num of failed requests to aggregator: 0
137
+
138
+ [2026-02-27 00:32:12,514 I 12110 12152] accessor.cc:540: Received address and liveness notification for node, IsAlive = 0 node_id=d9160dc27b026e787f1f81b4465a36f139cf00f32ff2820e42f20120
139
+ [2026-02-27 00:32:12,514 I 12110 12152] core_worker.cc:740: Node failure. All objects pinned on that node will be lost if object reconstruction is not enabled. node_id=d9160dc27b026e787f1f81b4465a36f139cf00f32ff2820e42f20120
140
+ [2026-02-27 00:32:12,514 I 12110 12152] normal_task_submitter.cc:824: Number of alive nodes:0
141
+ [2026-02-27 00:32:13,170 W 12110 12152] core_worker.cc:771: Shutting down the core worker because the local raylet failed. Check out the raylet.out log file. Raylet pid: 11302
142
+ [2026-02-27 00:32:13,175 I 12110 12152] core_worker.cc:628: Try killing all child processes of this worker as it exits. Child process pids:
python-core-worker-5ad3b871f9c47a0419d1c26aa73c88d3ae2d40ede3aeceeef3079ef2_11379.log ADDED
@@ -0,0 +1,139 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2026-02-27 00:30:32,515 I 11379 11379] core_worker_process.cc:773: Constructing CoreWorkerProcess. pid: 11379
2
+ [2026-02-27 00:30:32,521 I 11379 11379] event.cc:499: Ray Event initialized for CORE_WORKER
3
+ [2026-02-27 00:30:32,522 I 11379 11379] event.cc:499: Ray Event initialized for EXPORT_TASK
4
+ [2026-02-27 00:30:32,523 I 11379 11379] event.cc:332: Set ray event level to warning
5
+ [2026-02-27 00:30:32,523 I 11379 11379] event_aggregator_client.h:50: Initiating the local event aggregator client with port: 55678
6
+ [2026-02-27 00:30:32,525 I 11379 11379] grpc_server.cc:143: worker server started, listening on port 50199.
7
+ [2026-02-27 00:30:32,562 I 11379 11379] core_worker_process.cc:261: Initializing worker at address: 10.128.0.163:50199 worker_id=5ad3b871f9c47a0419d1c26aa73c88d3ae2d40ede3aeceeef3079ef2 node_id=d9160dc27b026e787f1f81b4465a36f139cf00f32ff2820e42f20120
8
+ [2026-02-27 00:30:32,568 I 11379 11379] task_event_buffer.cc:480: Reporting task events to GCS every 1000ms.
9
+ [2026-02-27 00:30:32,576 I 11379 11510] core_worker.cc:455: Event stats:
10
+
11
+
12
+ Global stats: 9 total (5 active)
13
+ Queueing time: mean = 0.84ms, max = 7.50ms, min = 0.02ms, total = 7.54ms
14
+ Execution time: mean = 0.07ms, total = 0.66ms
15
+ Event stats:
16
+ PeriodicalRunner.RunFnPeriodically - 3 total (1 active, 1 running), Execution time: mean = 0.01ms, total = 0.03ms, Queueing time: mean = 2.51ms, max = 7.50ms, min = 0.02ms, total = 7.52ms
17
+ ray::rpc::InternalPubSubGcsService.grpc_client.GcsSubscriberCommandBatch - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
18
+ ray::rpc::WorkerInfoGcsService.grpc_client.AddWorkerInfo - 1 total (0 active), Execution time: mean = 0.61ms, total = 0.61ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
19
+ Publisher.CheckDeadSubscribers - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
20
+ ray::rpc::WorkerInfoGcsService.grpc_client.AddWorkerInfo.OnReplyReceived - 1 total (0 active), Execution time: mean = 0.02ms, total = 0.02ms, Queueing time: mean = 0.02ms, max = 0.02ms, min = 0.02ms, total = 0.02ms
21
+ ray::rpc::InternalPubSubGcsService.grpc_client.GcsSubscriberPoll - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
22
+ CoreWorker.ExitIfParentRayletDies - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
23
+
24
+ -----------------
25
+ Task execution event stats:
26
+
27
+ Global stats: 0 total (0 active)
28
+ Queueing time: mean = -nanms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
29
+ Execution time: mean = -nanms, total = 0.00ms
30
+ Event stats:
31
+
32
+ -----------------
33
+ Task Event stats:
34
+
35
+ IO Service Stats:
36
+
37
+ Global stats: 4 total (1 active)
38
+ Queueing time: mean = 0.01ms, max = 0.02ms, min = 0.02ms, total = 0.04ms
39
+ Execution time: mean = 0.36ms, total = 1.44ms
40
+ Event stats:
41
+ CoreWorker.deadline_timer.flush_task_events - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
42
+ ray::rpc::TaskInfoGcsService.grpc_client.AddTaskEventData.OnReplyReceived - 1 total (0 active), Execution time: mean = 0.03ms, total = 0.03ms, Queueing time: mean = 0.02ms, max = 0.02ms, min = 0.02ms, total = 0.02ms
43
+ PeriodicalRunner.RunFnPeriodically - 1 total (0 active), Execution time: mean = 0.29ms, total = 0.29ms, Queueing time: mean = 0.02ms, max = 0.02ms, min = 0.02ms, total = 0.02ms
44
+ ray::rpc::TaskInfoGcsService.grpc_client.AddTaskEventData - 1 total (0 active), Execution time: mean = 1.11ms, total = 1.11ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
45
+ Other Stats:
46
+ gcs_grpc_in_progress:0
47
+ event_aggregator_grpc_in_progress:0
48
+ current number of task status events in buffer: 0
49
+ current number of profile events in buffer: 0
50
+ current number of dropped task attempts tracked: 0
51
+ total task events sent: 0 MiB
52
+ total number of task attempts sent: 0
53
+ total number of task attempts dropped reported: 0
54
+ total number of sent failure: 0
55
+ num status task events dropped: 0
56
+ num profile task events dropped: 0
57
+ num ray task events reported to aggregator: 0
58
+ num ray task events failed to report to aggregator: 0
59
+ num of task attempts dropped reported to aggregator: 0
60
+ num of failed requests to aggregator: 0
61
+
62
+ [2026-02-27 00:30:32,579 I 11379 11510] accessor.cc:540: Received address and liveness notification for node, IsAlive = 1 node_id=d9160dc27b026e787f1f81b4465a36f139cf00f32ff2820e42f20120
63
+ [2026-02-27 00:30:32,579 I 11379 11510] normal_task_submitter.cc:824: Number of alive nodes:1
64
+ [2026-02-27 00:30:32,580 I 11379 11379] core_worker.cc:515: Adjusted worker niceness to 15
65
+ [2026-02-27 00:30:32,581 I 11379 11379] metrics_agent_client.cc:42: Initializing exporter ...
66
+ [2026-02-27 00:30:33,586 I 11379 11510] metrics_agent_client.cc:54: Exporter initialized.
67
+ [2026-02-27 00:31:32,577 I 11379 11510] core_worker.cc:455: Event stats:
68
+
69
+
70
+ Global stats: 887 total (8 active)
71
+ Queueing time: mean = 1.19ms, max = 1000.03ms, min = 0.01ms, total = 1052.01ms
72
+ Execution time: mean = 0.09ms, total = 84.25ms
73
+ Event stats:
74
+ CoreWorker.RecoverObjects - 600 total (1 active), Execution time: mean = 0.01ms, total = 4.64ms, Queueing time: mean = 0.06ms, max = 5.10ms, min = 0.02ms, total = 36.75ms
75
+ CoreWorker.InternalHeartbeat - 60 total (1 active), Execution time: mean = 0.14ms, total = 8.28ms, Queueing time: mean = 0.04ms, max = 0.06ms, min = 0.02ms, total = 2.20ms
76
+ CoreWorker.ExitIfParentRayletDies - 60 total (1 active), Execution time: mean = 0.01ms, total = 0.54ms, Queueing time: mean = 0.04ms, max = 0.26ms, min = 0.03ms, total = 2.64ms
77
+ NodeManagerService.grpc_client.ReportWorkerBacklog - 60 total (0 active), Execution time: mean = 0.94ms, total = 56.48ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
78
+ NodeManagerService.grpc_client.ReportWorkerBacklog.OnReplyReceived - 60 total (0 active), Execution time: mean = 0.02ms, total = 1.22ms, Queueing time: mean = 0.02ms, max = 0.04ms, min = 0.01ms, total = 1.48ms
79
+ CoreWorker.RecordMetrics - 12 total (1 active), Execution time: mean = 0.04ms, total = 0.49ms, Queueing time: mean = 0.03ms, max = 0.04ms, min = 0.02ms, total = 0.30ms
80
+ PeriodicalRunner.RunFnPeriodically - 7 total (0 active), Execution time: mean = 0.10ms, total = 0.70ms, Queueing time: mean = 1.18ms, max = 7.50ms, min = 0.01ms, total = 8.26ms
81
+ CoreWorker.TryDelPendingObjectRefStreams - 6 total (1 active), Execution time: mean = 0.00ms, total = 0.02ms, Queueing time: mean = 0.03ms, max = 0.04ms, min = 0.02ms, total = 0.17ms
82
+ CoreWorkerService.grpc_server.GetCoreWorkerStats - 4 total (0 active), Execution time: mean = 0.49ms, total = 1.96ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
83
+ CoreWorkerService.grpc_server.GetCoreWorkerStats.HandleRequestImpl - 4 total (0 active), Execution time: mean = 0.08ms, total = 0.33ms, Queueing time: mean = 0.02ms, max = 0.03ms, min = 0.02ms, total = 0.10ms
84
+ ReporterService.grpc_client.HealthCheck - 2 total (0 active), Execution time: mean = 2.36ms, total = 4.72ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
85
+ ReporterService.grpc_client.HealthCheck.OnReplyReceived - 2 total (0 active), Execution time: mean = 0.24ms, total = 0.47ms, Queueing time: mean = 0.02ms, max = 0.02ms, min = 0.02ms, total = 0.04ms
86
+ ray::rpc::WorkerInfoGcsService.grpc_client.AddWorkerInfo.OnReplyReceived - 1 total (0 active), Execution time: mean = 0.02ms, total = 0.02ms, Queueing time: mean = 0.02ms, max = 0.02ms, min = 0.02ms, total = 0.02ms
87
+ ray::rpc::InternalPubSubGcsService.grpc_client.GcsSubscriberPoll - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
88
+ Publisher.CheckDeadSubscribers - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
89
+ ray::rpc::NodeInfoGcsService.grpc_client.GetAllNodeAddressAndLiveness.OnReplyReceived - 1 total (0 active), Execution time: mean = 0.09ms, total = 0.09ms, Queueing time: mean = 0.01ms, max = 0.01ms, min = 0.01ms, total = 0.01ms
90
+ ray::rpc::InternalPubSubGcsService.grpc_client.GcsSubscriberCommandBatch - 1 total (0 active), Execution time: mean = 2.57ms, total = 2.57ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
91
+ ray::rpc::WorkerInfoGcsService.grpc_client.AddWorkerInfo - 1 total (0 active), Execution time: mean = 0.61ms, total = 0.61ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
92
+ MetricsAgentClient.WaitForServerReadyWithRetry - 1 total (0 active), Execution time: mean = 0.12ms, total = 0.12ms, Queueing time: mean = 1000.03ms, max = 1000.03ms, min = 1000.03ms, total = 1000.03ms
93
+ ray::rpc::InternalPubSubGcsService.grpc_client.GcsSubscriberCommandBatch.OnReplyReceived - 1 total (0 active), Execution time: mean = 0.15ms, total = 0.15ms, Queueing time: mean = 0.01ms, max = 0.01ms, min = 0.01ms, total = 0.01ms
94
+ ray::rpc::NodeInfoGcsService.grpc_client.GetAllNodeAddressAndLiveness - 1 total (0 active), Execution time: mean = 0.83ms, total = 0.83ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
95
+ CoreWorker.PrintEventStats - 1 total (1 active, 1 running), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
96
+
97
+ -----------------
98
+ Task execution event stats:
99
+
100
+ Global stats: 5945 total (1 active)
101
+ Queueing time: mean = 0.09ms, max = 40.99ms, min = -0.00ms, total = 509.24ms
102
+ Execution time: mean = 0.01ms, total = 63.59ms
103
+ Event stats:
104
+ CoreWorker.CheckSignal - 5944 total (1 active), Execution time: mean = 0.01ms, total = 63.58ms, Queueing time: mean = 0.09ms, max = 40.99ms, min = -0.00ms, total = 509.24ms
105
+ PeriodicalRunner.RunFnPeriodically - 1 total (0 active), Execution time: mean = 0.01ms, total = 0.01ms, Queueing time: mean = 0.00ms, max = 0.00ms, min = 0.00ms, total = 0.00ms
106
+
107
+ -----------------
108
+ Task Event stats:
109
+
110
+ IO Service Stats:
111
+
112
+ Global stats: 181 total (1 active)
113
+ Queueing time: mean = 0.06ms, max = 2.58ms, min = 0.02ms, total = 10.15ms
114
+ Execution time: mean = 0.42ms, total = 76.83ms
115
+ Event stats:
116
+ CoreWorker.deadline_timer.flush_task_events - 60 total (1 active), Execution time: mean = 0.31ms, total = 18.44ms, Queueing time: mean = 0.07ms, max = 0.63ms, min = 0.03ms, total = 4.10ms
117
+ ray::rpc::TaskInfoGcsService.grpc_client.AddTaskEventData.OnReplyReceived - 60 total (0 active), Execution time: mean = 0.04ms, total = 2.34ms, Queueing time: mean = 0.10ms, max = 2.58ms, min = 0.02ms, total = 6.04ms
118
+ ray::rpc::TaskInfoGcsService.grpc_client.AddTaskEventData - 60 total (0 active), Execution time: mean = 0.93ms, total = 55.76ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
119
+ PeriodicalRunner.RunFnPeriodically - 1 total (0 active), Execution time: mean = 0.29ms, total = 0.29ms, Queueing time: mean = 0.02ms, max = 0.02ms, min = 0.02ms, total = 0.02ms
120
+ Other Stats:
121
+ gcs_grpc_in_progress:0
122
+ event_aggregator_grpc_in_progress:0
123
+ current number of task status events in buffer: 0
124
+ current number of profile events in buffer: 0
125
+ current number of dropped task attempts tracked: 0
126
+ total task events sent: 0 MiB
127
+ total number of task attempts sent: 0
128
+ total number of task attempts dropped reported: 0
129
+ total number of sent failure: 0
130
+ num status task events dropped: 0
131
+ num profile task events dropped: 0
132
+ num ray task events reported to aggregator: 0
133
+ num ray task events failed to report to aggregator: 0
134
+ num of task attempts dropped reported to aggregator: 0
135
+ num of failed requests to aggregator: 0
136
+
137
+ [2026-02-27 00:32:12,517 I 11379 11510] accessor.cc:540: Received address and liveness notification for node, IsAlive = 0 node_id=d9160dc27b026e787f1f81b4465a36f139cf00f32ff2820e42f20120
138
+ [2026-02-27 00:32:12,517 I 11379 11510] core_worker.cc:740: Node failure. All objects pinned on that node will be lost if object reconstruction is not enabled. node_id=d9160dc27b026e787f1f81b4465a36f139cf00f32ff2820e42f20120
139
+ [2026-02-27 00:32:12,518 I 11379 11510] normal_task_submitter.cc:824: Number of alive nodes:0
python-core-worker-772052e39bd349442253d65d82fc94825e9e58c75098b1b473bedce2_12586.log ADDED
@@ -0,0 +1,72 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2026-02-27 00:31:33,408 I 12586 12586] core_worker_process.cc:773: Constructing CoreWorkerProcess. pid: 12586
2
+ [2026-02-27 00:31:33,415 I 12586 12586] event.cc:499: Ray Event initialized for CORE_WORKER
3
+ [2026-02-27 00:31:33,415 I 12586 12586] event.cc:499: Ray Event initialized for EXPORT_TASK
4
+ [2026-02-27 00:31:33,415 I 12586 12586] event.cc:332: Set ray event level to warning
5
+ [2026-02-27 00:31:33,416 I 12586 12586] event_aggregator_client.h:50: Initiating the local event aggregator client with port: 55678
6
+ [2026-02-27 00:31:33,418 I 12586 12586] grpc_server.cc:143: worker server started, listening on port 50477.
7
+ [2026-02-27 00:31:33,437 I 12586 12586] core_worker_process.cc:261: Initializing worker at address: 10.128.0.163:50477 worker_id=772052e39bd349442253d65d82fc94825e9e58c75098b1b473bedce2 node_id=d9160dc27b026e787f1f81b4465a36f139cf00f32ff2820e42f20120
8
+ [2026-02-27 00:31:33,438 I 12586 12586] task_event_buffer.cc:480: Reporting task events to GCS every 1000ms.
9
+ [2026-02-27 00:31:33,439 I 12586 12586] core_worker.cc:515: Adjusted worker niceness to 15
10
+ [2026-02-27 00:31:33,440 I 12586 12586] metrics_agent_client.cc:42: Initializing exporter ...
11
+ [2026-02-27 00:31:33,440 I 12586 12950] core_worker.cc:455: Event stats:
12
+
13
+
14
+ Global stats: 12 total (10 active)
15
+ Queueing time: mean = 0.01ms, max = 0.09ms, min = 0.08ms, total = 0.16ms
16
+ Execution time: mean = 0.00ms, total = 0.04ms
17
+ Event stats:
18
+ PeriodicalRunner.RunFnPeriodically - 7 total (5 active, 1 running), Execution time: mean = 0.01ms, total = 0.04ms, Queueing time: mean = 0.02ms, max = 0.09ms, min = 0.08ms, total = 0.16ms
19
+ ray::rpc::WorkerInfoGcsService.grpc_client.AddWorkerInfo - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
20
+ ray::rpc::InternalPubSubGcsService.grpc_client.GcsSubscriberCommandBatch - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
21
+ Publisher.CheckDeadSubscribers - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
22
+ ray::rpc::InternalPubSubGcsService.grpc_client.GcsSubscriberPoll - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
23
+ CoreWorker.ExitIfParentRayletDies - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
24
+
25
+ -----------------
26
+ Task execution event stats:
27
+
28
+ Global stats: 0 total (0 active)
29
+ Queueing time: mean = -nanms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
30
+ Execution time: mean = -nanms, total = 0.00ms
31
+ Event stats:
32
+
33
+ -----------------
34
+ Task Event stats:
35
+
36
+ IO Service Stats:
37
+
38
+ Global stats: 3 total (2 active)
39
+ Queueing time: mean = 0.02ms, max = 0.06ms, min = 0.06ms, total = 0.06ms
40
+ Execution time: mean = 0.08ms, total = 0.24ms
41
+ Event stats:
42
+ PeriodicalRunner.RunFnPeriodically - 1 total (0 active), Execution time: mean = 0.24ms, total = 0.24ms, Queueing time: mean = 0.06ms, max = 0.06ms, min = 0.06ms, total = 0.06ms
43
+ ray::rpc::TaskInfoGcsService.grpc_client.AddTaskEventData - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
44
+ CoreWorker.deadline_timer.flush_task_events - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
45
+ Other Stats:
46
+ gcs_grpc_in_progress:1
47
+ event_aggregator_grpc_in_progress:0
48
+ current number of task status events in buffer: 0
49
+ current number of profile events in buffer: 0
50
+ current number of dropped task attempts tracked: 0
51
+ total task events sent: 0 MiB
52
+ total number of task attempts sent: 0
53
+ total number of task attempts dropped reported: 0
54
+ total number of sent failure: 0
55
+ num status task events dropped: 0
56
+ num profile task events dropped: 0
57
+ num ray task events reported to aggregator: 0
58
+ num ray task events failed to report to aggregator: 0
59
+ num of task attempts dropped reported to aggregator: 0
60
+ num of failed requests to aggregator: 0
61
+
62
+ [2026-02-27 00:31:33,446 I 12586 12950] metrics_agent_client.cc:54: Exporter initialized.
63
+ [2026-02-27 00:31:33,447 I 12586 12950] accessor.cc:540: Received address and liveness notification for node, IsAlive = 1 node_id=d9160dc27b026e787f1f81b4465a36f139cf00f32ff2820e42f20120
64
+ [2026-02-27 00:31:33,447 I 12586 12950] normal_task_submitter.cc:824: Number of alive nodes:1
65
+ [2026-02-27 00:31:33,457 I 12586 12586] actor_task_submitter.cc:74: Set actor max pending calls to -1 actor_id=0cc31f28648b86996c84113a01000000
66
+ [2026-02-27 00:31:33,458 I 12586 12586] core_worker.cc:2903: Creating actor actor_id=0cc31f28648b86996c84113a01000000
67
+ [2026-02-27 00:31:44,889 I 12586 12586] task_receiver.cc:142: Actor creation task finished, task_id: ffffffffffffffff0cc31f28648b86996c84113a01000000, actor_id: 0cc31f28648b86996c84113a01000000, actor_repr_name:
68
+ [2026-02-27 00:32:12,462 I 12586 12950] core_worker_shutdown_executor.cc:217: Try killing all child processes of this worker as it exits. Child process pids:
69
+ [2026-02-27 00:32:12,462 I 12586 12950] core_worker_shutdown_executor.cc:262: Sending disconnect message to the local raylet.
70
+ [2026-02-27 00:32:12,462 I 12586 12950] raylet_ipc_client.cc:135: RayletIpcClient::Disconnect, exit_type=INTENDED_USER_EXIT, exit_detail=Worker force exited because its job has finished, has creation_task_exception_pb_bytes=0
71
+ [2026-02-27 00:32:12,465 I 12586 12950] core_worker_shutdown_executor.cc:279: Disconnected from the local raylet.
72
+ [2026-02-27 00:32:12,465 W 12586 12950] core_worker_shutdown_executor.cc:288: Quick exit - terminating process immediately
python-core-worker-8d27d27b6a5c820150d6a54cd27fe296fd0409567d2b4685b9a84fc8_11896.log ADDED
@@ -0,0 +1,195 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2026-02-27 00:30:38,465 I 11896 11896] core_worker_process.cc:773: Constructing CoreWorkerProcess. pid: 11896
2
+ [2026-02-27 00:30:38,468 I 11896 11896] event.cc:499: Ray Event initialized for CORE_WORKER
3
+ [2026-02-27 00:30:38,469 I 11896 11896] event.cc:499: Ray Event initialized for EXPORT_TASK
4
+ [2026-02-27 00:30:38,469 I 11896 11896] event.cc:332: Set ray event level to warning
5
+ [2026-02-27 00:30:38,469 I 11896 11896] event_aggregator_client.h:50: Initiating the local event aggregator client with port: 55678
6
+ [2026-02-27 00:30:38,470 I 11896 11896] grpc_server.cc:143: worker server started, listening on port 50157.
7
+ [2026-02-27 00:30:38,480 I 11896 11896] core_worker_process.cc:261: Initializing worker at address: 10.128.0.163:50157 worker_id=8d27d27b6a5c820150d6a54cd27fe296fd0409567d2b4685b9a84fc8 node_id=d9160dc27b026e787f1f81b4465a36f139cf00f32ff2820e42f20120
8
+ [2026-02-27 00:30:38,481 I 11896 11896] task_event_buffer.cc:480: Reporting task events to GCS every 1000ms.
9
+ [2026-02-27 00:30:38,482 I 11896 11896] core_worker.cc:515: Adjusted worker niceness to 15
10
+ [2026-02-27 00:30:38,482 I 11896 11946] core_worker.cc:455: Event stats:
11
+
12
+
13
+ Global stats: 12 total (10 active)
14
+ Queueing time: mean = 0.01ms, max = 0.04ms, min = 0.04ms, total = 0.07ms
15
+ Execution time: mean = 0.00ms, total = 0.06ms
16
+ Event stats:
17
+ PeriodicalRunner.RunFnPeriodically - 7 total (5 active, 1 running), Execution time: mean = 0.01ms, total = 0.06ms, Queueing time: mean = 0.01ms, max = 0.04ms, min = 0.04ms, total = 0.07ms
18
+ Publisher.CheckDeadSubscribers - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
19
+ ray::rpc::WorkerInfoGcsService.grpc_client.AddWorkerInfo - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
20
+ ray::rpc::InternalPubSubGcsService.grpc_client.GcsSubscriberCommandBatch - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
21
+ ray::rpc::InternalPubSubGcsService.grpc_client.GcsSubscriberPoll - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
22
+ CoreWorker.ExitIfParentRayletDies - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
23
+
24
+ -----------------
25
+ Task execution event stats:
26
+
27
+ Global stats: 0 total (0 active)
28
+ Queueing time: mean = -nanms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
29
+ Execution time: mean = -nanms, total = 0.00ms
30
+ Event stats:
31
+
32
+ -----------------
33
+ Task Event stats:
34
+
35
+ IO Service Stats:
36
+
37
+ Global stats: 4 total (1 active)
38
+ Queueing time: mean = 0.01ms, max = 0.02ms, min = 0.02ms, total = 0.04ms
39
+ Execution time: mean = 0.23ms, total = 0.94ms
40
+ Event stats:
41
+ ray::rpc::TaskInfoGcsService.grpc_client.AddTaskEventData.OnReplyReceived - 1 total (0 active), Execution time: mean = 0.02ms, total = 0.02ms, Queueing time: mean = 0.02ms, max = 0.02ms, min = 0.02ms, total = 0.02ms
42
+ PeriodicalRunner.RunFnPeriodically - 1 total (0 active), Execution time: mean = 0.29ms, total = 0.29ms, Queueing time: mean = 0.02ms, max = 0.02ms, min = 0.02ms, total = 0.02ms
43
+ CoreWorker.deadline_timer.flush_task_events - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
44
+ ray::rpc::TaskInfoGcsService.grpc_client.AddTaskEventData - 1 total (0 active), Execution time: mean = 0.63ms, total = 0.63ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
45
+ Other Stats:
46
+ gcs_grpc_in_progress:0
47
+ event_aggregator_grpc_in_progress:0
48
+ current number of task status events in buffer: 0
49
+ current number of profile events in buffer: 0
50
+ current number of dropped task attempts tracked: 0
51
+ total task events sent: 0 MiB
52
+ total number of task attempts sent: 0
53
+ total number of task attempts dropped reported: 0
54
+ total number of sent failure: 0
55
+ num status task events dropped: 0
56
+ num profile task events dropped: 0
57
+ num ray task events reported to aggregator: 0
58
+ num ray task events failed to report to aggregator: 0
59
+ num of task attempts dropped reported to aggregator: 0
60
+ num of failed requests to aggregator: 0
61
+
62
+ [2026-02-27 00:30:38,482 I 11896 11896] metrics_agent_client.cc:42: Initializing exporter ...
63
+ [2026-02-27 00:30:38,485 I 11896 11946] accessor.cc:540: Received address and liveness notification for node, IsAlive = 1 node_id=d9160dc27b026e787f1f81b4465a36f139cf00f32ff2820e42f20120
64
+ [2026-02-27 00:30:38,485 I 11896 11946] normal_task_submitter.cc:824: Number of alive nodes:1
65
+ [2026-02-27 00:30:38,485 I 11896 11946] metrics_agent_client.cc:54: Exporter initialized.
66
+ [2026-02-27 00:30:38,486 I 11896 11896] actor_task_submitter.cc:74: Set actor max pending calls to -1 actor_id=0f47a291c83181c09abd27b101000000
67
+ [2026-02-27 00:30:38,486 I 11896 11896] core_worker.cc:2903: Creating actor actor_id=0f47a291c83181c09abd27b101000000
68
+ [2026-02-27 00:30:50,693 I 11896 11896] task_receiver.cc:142: Actor creation task finished, task_id: ffffffffffffffff0f47a291c83181c09abd27b101000000, actor_id: 0f47a291c83181c09abd27b101000000, actor_repr_name:
69
+ [2026-02-27 00:30:57,303 I 11896 11896] core_worker.cc:2301: Submitting Placement Group creation to GCS placement_group_id=6834375140417d94aa5cc2a5c3d701000000
70
+ [2026-02-27 00:30:59,438 W 11896 11896] actor_manager.cc:110: Failed to look up actor with name 'B8Qm6MWorkerDict_0:0'. This could because 1. You are trying to look up a named actor you didn't create. 2. The named actor died. 3. You did not use a namespace matching the namespace of the actor.
71
+ [2026-02-27 00:30:59,654 I 11896 11896] actor_task_submitter.cc:74: Set actor max pending calls to -1 actor_id=ddbc05b6418ab5a7f12dad2101000000
72
+ [2026-02-27 00:30:59,676 I 11896 11946] actor_manager.cc:236: received notification on actor, state: PENDING_CREATION, ip address: , port: 0, num_restarts: 0, death context type=CONTEXT_NOT_SET actor_id=ddbc05b6418ab5a7f12dad2101000000 worker_id=NIL_ID node_id=d9160dc27b026e787f1f81b4465a36f139cf00f32ff2820e42f20120
73
+ [2026-02-27 00:31:16,005 I 11896 11946] actor_manager.cc:236: received notification on actor, state: ALIVE, ip address: 10.128.0.163, port: 50215, num_restarts: 0, death context type=CONTEXT_NOT_SET actor_id=ddbc05b6418ab5a7f12dad2101000000 worker_id=a711ab381f1202e338fc2083afa6dd5133aebf91969e1e83b35a9610 node_id=d9160dc27b026e787f1f81b4465a36f139cf00f32ff2820e42f20120
74
+ [2026-02-27 00:31:29,833 W 11896 11896] actor_manager.cc:110: Failed to look up actor with name 'reward_loop_worker_0'. This could because 1. You are trying to look up a named actor you didn't create. 2. The named actor died. 3. You did not use a namespace matching the namespace of the actor.
75
+ [2026-02-27 00:31:29,859 I 11896 11896] actor_task_submitter.cc:74: Set actor max pending calls to -1 actor_id=2f225182ea5eab0591f8ce2101000000
76
+ [2026-02-27 00:31:29,862 W 11896 11896] actor_manager.cc:110: Failed to look up actor with name 'reward_loop_worker_1'. This could because 1. You are trying to look up a named actor you didn't create. 2. The named actor died. 3. You did not use a namespace matching the namespace of the actor.
77
+ [2026-02-27 00:31:29,871 I 11896 11896] actor_task_submitter.cc:74: Set actor max pending calls to -1 actor_id=2110253750be447ceb2be14f01000000
78
+ [2026-02-27 00:31:29,874 W 11896 11896] actor_manager.cc:110: Failed to look up actor with name 'reward_loop_worker_2'. This could because 1. You are trying to look up a named actor you didn't create. 2. The named actor died. 3. You did not use a namespace matching the namespace of the actor.
79
+ [2026-02-27 00:31:29,884 I 11896 11896] actor_task_submitter.cc:74: Set actor max pending calls to -1 actor_id=87ab4f0b8032c66348227f5a01000000
80
+ [2026-02-27 00:31:29,887 W 11896 11896] actor_manager.cc:110: Failed to look up actor with name 'reward_loop_worker_3'. This could because 1. You are trying to look up a named actor you didn't create. 2. The named actor died. 3. You did not use a namespace matching the namespace of the actor.
81
+ [2026-02-27 00:31:29,898 I 11896 11896] actor_task_submitter.cc:74: Set actor max pending calls to -1 actor_id=83a8e5ec085f16d6fdcd7dc801000000
82
+ [2026-02-27 00:31:29,902 W 11896 11896] actor_manager.cc:110: Failed to look up actor with name 'reward_loop_worker_4'. This could because 1. You are trying to look up a named actor you didn't create. 2. The named actor died. 3. You did not use a namespace matching the namespace of the actor.
83
+ [2026-02-27 00:31:29,914 I 11896 11896] actor_task_submitter.cc:74: Set actor max pending calls to -1 actor_id=ae31a85bdd5e21bfaec5bc9e01000000
84
+ [2026-02-27 00:31:29,918 W 11896 11896] actor_manager.cc:110: Failed to look up actor with name 'reward_loop_worker_5'. This could because 1. You are trying to look up a named actor you didn't create. 2. The named actor died. 3. You did not use a namespace matching the namespace of the actor.
85
+ [2026-02-27 00:31:29,931 I 11896 11896] actor_task_submitter.cc:74: Set actor max pending calls to -1 actor_id=a0546db81b53339607d0f80c01000000
86
+ [2026-02-27 00:31:29,936 W 11896 11896] actor_manager.cc:110: Failed to look up actor with name 'reward_loop_worker_6'. This could because 1. You are trying to look up a named actor you didn't create. 2. The named actor died. 3. You did not use a namespace matching the namespace of the actor.
87
+ [2026-02-27 00:31:29,951 I 11896 11896] actor_task_submitter.cc:74: Set actor max pending calls to -1 actor_id=0cc31f28648b86996c84113a01000000
88
+ [2026-02-27 00:31:29,962 W 11896 11896] actor_manager.cc:110: Failed to look up actor with name 'reward_loop_worker_7'. This could because 1. You are trying to look up a named actor you didn't create. 2. The named actor died. 3. You did not use a namespace matching the namespace of the actor.
89
+ [2026-02-27 00:31:29,978 I 11896 11896] actor_task_submitter.cc:74: Set actor max pending calls to -1 actor_id=1b6a74cdbf74d457007f760101000000
90
+ [2026-02-27 00:31:38,013 W 11896 11896] actor_manager.cc:110: Failed to look up actor with name 'vllm_server_0_0'. This could because 1. You are trying to look up a named actor you didn't create. 2. The named actor died. 3. You did not use a namespace matching the namespace of the actor.
91
+ [2026-02-27 00:31:38,159 I 11896 11896] actor_task_submitter.cc:74: Set actor max pending calls to -1 actor_id=8b6cdab5e264a7c65d1ca0d701000000
92
+ [2026-02-27 00:31:38,163 I 11896 11946] actor_manager.cc:236: received notification on actor, state: PENDING_CREATION, ip address: , port: 0, num_restarts: 0, death context type=CONTEXT_NOT_SET actor_id=8b6cdab5e264a7c65d1ca0d701000000 worker_id=NIL_ID node_id=d9160dc27b026e787f1f81b4465a36f139cf00f32ff2820e42f20120
93
+ [2026-02-27 00:31:38,483 I 11896 11946] core_worker.cc:455: Event stats:
94
+
95
+
96
+ Global stats: 1007 total (28 active)
97
+ Queueing time: mean = 0.08ms, max = 11.78ms, min = 0.01ms, total = 83.20ms
98
+ Execution time: mean = 81.62ms, total = 82189.78ms
99
+ Event stats:
100
+ CoreWorker.RecoverObjects - 600 total (1 active), Execution time: mean = 0.01ms, total = 4.99ms, Queueing time: mean = 0.08ms, max = 10.23ms, min = 0.03ms, total = 48.07ms
101
+ NodeManagerService.grpc_client.ReportWorkerBacklog - 60 total (0 active), Execution time: mean = 0.75ms, total = 45.19ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
102
+ CoreWorker.ExitIfParentRayletDies - 60 total (1 active), Execution time: mean = 0.01ms, total = 0.67ms, Queueing time: mean = 0.15ms, max = 6.22ms, min = 0.02ms, total = 8.94ms
103
+ CoreWorker.InternalHeartbeat - 60 total (1 active), Execution time: mean = 0.17ms, total = 10.41ms, Queueing time: mean = 0.24ms, max = 11.78ms, min = 0.02ms, total = 14.41ms
104
+ NodeManagerService.grpc_client.ReportWorkerBacklog.OnReplyReceived - 60 total (0 active), Execution time: mean = 0.02ms, total = 1.48ms, Queueing time: mean = 0.04ms, max = 0.64ms, min = 0.01ms, total = 2.54ms
105
+ CoreWorker.RecordMetrics - 12 total (1 active), Execution time: mean = 1.11ms, total = 13.37ms, Queueing time: mean = 0.04ms, max = 0.08ms, min = 0.03ms, total = 0.49ms
106
+ CoreWorker.SubmitTask - 12 total (0 active), Execution time: mean = 0.41ms, total = 4.96ms, Queueing time: mean = 0.11ms, max = 1.00ms, min = 0.02ms, total = 1.36ms
107
+ CoreWorkerService.grpc_server.WaitForActorRefDeleted - 10 total (10 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
108
+ CoreWorkerService.grpc_server.WaitForActorRefDeleted.HandleRequestImpl - 10 total (0 active), Execution time: mean = 0.06ms, total = 0.60ms, Queueing time: mean = 0.03ms, max = 0.06ms, min = 0.01ms, total = 0.28ms
109
+ ray::rpc::ActorInfoGcsService.grpc_client.CreateActor - 10 total (9 active), Execution time: mean = 1634.00ms, total = 16340.03ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
110
+ ray::rpc::ActorInfoGcsService.grpc_client.RegisterActor.OnReplyReceived - 10 total (0 active), Execution time: mean = 0.08ms, total = 0.75ms, Queueing time: mean = 0.11ms, max = 0.83ms, min = 0.02ms, total = 1.11ms
111
+ ray::rpc::ActorInfoGcsService.grpc_client.GetNamedActorInfo - 10 total (0 active), Execution time: mean = 1.19ms, total = 11.90ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
112
+ ray::rpc::ActorInfoGcsService.grpc_client.GetNamedActorInfo.OnReplyReceived - 10 total (0 active), Execution time: mean = 0.07ms, total = 0.65ms, Queueing time: mean = 0.02ms, max = 0.03ms, min = 0.01ms, total = 0.24ms
113
+ ray::rpc::ActorInfoGcsService.grpc_client.RegisterActor - 10 total (0 active), Execution time: mean = 2.49ms, total = 24.94ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
114
+ PeriodicalRunner.RunFnPeriodically - 7 total (0 active), Execution time: mean = 0.12ms, total = 0.83ms, Queueing time: mean = 0.35ms, max = 0.81ms, min = 0.04ms, total = 2.42ms
115
+ CoreWorker.TryDelPendingObjectRefStreams - 6 total (1 active), Execution time: mean = 0.00ms, total = 0.03ms, Queueing time: mean = 0.04ms, max = 0.06ms, min = 0.02ms, total = 0.22ms
116
+ CoreWorkerService.grpc_client.PushTask.OnReplyReceived - 5 total (0 active), Execution time: mean = 0.24ms, total = 1.20ms, Queueing time: mean = 0.03ms, max = 0.05ms, min = 0.02ms, total = 0.16ms
117
+ CoreWorkerService.grpc_client.PushTask - 5 total (0 active), Execution time: mean = 2822.87ms, total = 14114.34ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
118
+ CoreWorkerService.grpc_server.GetCoreWorkerStats.HandleRequestImpl - 4 total (0 active), Execution time: mean = 0.14ms, total = 0.55ms, Queueing time: mean = 0.03ms, max = 0.05ms, min = 0.02ms, total = 0.13ms
119
+ CoreWorkerService.grpc_server.GetCoreWorkerStats - 4 total (0 active), Execution time: mean = 0.50ms, total = 2.01ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
120
+ ActorTaskSubmitter::SubmitTask - 4 total (0 active), Execution time: mean = 0.24ms, total = 0.94ms, Queueing time: mean = 0.04ms, max = 0.06ms, min = 0.03ms, total = 0.15ms
121
+ ray::rpc::InternalPubSubGcsService.grpc_client.GcsSubscriberCommandBatch.OnReplyReceived - 3 total (0 active), Execution time: mean = 0.16ms, total = 0.48ms, Queueing time: mean = 0.13ms, max = 0.36ms, min = 0.02ms, total = 0.39ms
122
+ ray::rpc::InternalPubSubGcsService.grpc_client.GcsSubscriberCommandBatch - 3 total (0 active), Execution time: mean = 1.45ms, total = 4.36ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
123
+ NodeManagerService.grpc_client.RequestWorkerLease - 2 total (0 active), Execution time: mean = 928.32ms, total = 1856.64ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
124
+ CoreWorkerService.grpc_server.PushTask.HandleRequestImpl - 2 total (0 active), Execution time: mean = 0.28ms, total = 0.57ms, Queueing time: mean = 0.24ms, max = 0.44ms, min = 0.04ms, total = 0.48ms
125
+ NodeManagerService.grpc_client.ReturnWorkerLease - 2 total (0 active), Execution time: mean = 0.98ms, total = 1.96ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
126
+ NodeManagerService.grpc_client.RequestWorkerLease.OnReplyReceived - 2 total (0 active), Execution time: mean = 0.56ms, total = 1.11ms, Queueing time: mean = 0.03ms, max = 0.03ms, min = 0.03ms, total = 0.06ms
127
+ CoreWorkerService.grpc_server.PushTask - 2 total (1 active), Execution time: mean = 6104.32ms, total = 12208.64ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
128
+ ray::rpc::ActorInfoGcsService.grpc_client.GetActorInfo.OnReplyReceived - 2 total (0 active), Execution time: mean = 0.51ms, total = 1.02ms, Queueing time: mean = 0.03ms, max = 0.03ms, min = 0.02ms, total = 0.05ms
129
+ ray::rpc::ActorInfoGcsService.grpc_client.GetActorInfo - 2 total (0 active), Execution time: mean = 1.69ms, total = 3.37ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
130
+ NodeManagerService.grpc_client.ReturnWorkerLease.OnReplyReceived - 2 total (0 active), Execution time: mean = 0.04ms, total = 0.08ms, Queueing time: mean = 0.03ms, max = 0.04ms, min = 0.02ms, total = 0.06ms
131
+ ray::rpc::InternalPubSubGcsService.grpc_client.GcsSubscriberPoll - 2 total (1 active), Execution time: mean = 18761.67ms, total = 37523.35ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
132
+ Subscriber.HandlePublishedMessage_GCS_ACTOR_CHANNEL - 1 total (0 active), Execution time: mean = 0.58ms, total = 0.58ms, Queueing time: mean = 0.27ms, max = 0.27ms, min = 0.27ms, total = 0.27ms
133
+ ray::rpc::ActorInfoGcsService.grpc_client.CreateActor.OnReplyReceived - 1 total (0 active), Execution time: mean = 0.10ms, total = 0.10ms, Queueing time: mean = 0.35ms, max = 0.35ms, min = 0.35ms, total = 0.35ms
134
+ CoreWorkerMemoryStore.Put.get_async_callbacks - 1 total (0 active), Execution time: mean = 0.44ms, total = 0.44ms, Queueing time: mean = 0.06ms, max = 0.06ms, min = 0.06ms, total = 0.06ms
135
+ ReporterService.grpc_client.HealthCheck - 1 total (0 active), Execution time: mean = 2.40ms, total = 2.40ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
136
+ ray::rpc::PlacementGroupInfoGcsService.grpc_client.CreatePlacementGroup.OnReplyReceived - 1 total (0 active), Execution time: mean = 0.06ms, total = 0.06ms, Queueing time: mean = 0.02ms, max = 0.02ms, min = 0.02ms, total = 0.02ms
137
+ ray::rpc::NodeInfoGcsService.grpc_client.GetAllNodeAddressAndLiveness - 1 total (0 active), Execution time: mean = 1.97ms, total = 1.97ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
138
+ ray::rpc::WorkerInfoGcsService.grpc_client.AddWorkerInfo - 1 total (0 active), Execution time: mean = 0.75ms, total = 0.75ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
139
+ ray::rpc::NodeInfoGcsService.grpc_client.GetAllNodeAddressAndLiveness.OnReplyReceived - 1 total (0 active), Execution time: mean = 0.27ms, total = 0.27ms, Queueing time: mean = 0.11ms, max = 0.11ms, min = 0.11ms, total = 0.11ms
140
+ ray::rpc::WorkerInfoGcsService.grpc_client.AddWorkerInfo.OnReplyReceived - 1 total (0 active), Execution time: mean = 0.03ms, total = 0.03ms, Queueing time: mean = 0.42ms, max = 0.42ms, min = 0.42ms, total = 0.42ms
141
+ ReporterService.grpc_client.HealthCheck.OnReplyReceived - 1 total (0 active), Execution time: mean = 0.37ms, total = 0.37ms, Queueing time: mean = 0.36ms, max = 0.36ms, min = 0.36ms, total = 0.36ms
142
+ ray::rpc::InternalPubSubGcsService.grpc_client.GcsSubscriberPoll.OnReplyReceived - 1 total (0 active), Execution time: mean = 0.37ms, total = 0.37ms, Queueing time: mean = 0.03ms, max = 0.03ms, min = 0.03ms, total = 0.03ms
143
+ Publisher.CheckDeadSubscribers - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
144
+ ray::rpc::PlacementGroupInfoGcsService.grpc_client.CreatePlacementGroup - 1 total (0 active), Execution time: mean = 0.99ms, total = 0.99ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
145
+ CoreWorker.PrintEventStats - 1 total (1 active, 1 running), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
146
+
147
+ -----------------
148
+ Task execution event stats:
149
+
150
+ Global stats: 5 total (2 active)
151
+ Queueing time: mean = 2439.62ms, max = 12198.08ms, min = 0.00ms, total = 12198.10ms
152
+ Execution time: mean = 2441.52ms, total = 12207.59ms
153
+ Event stats:
154
+ CoreWorker.CheckSignal - 2 total (1 active), Execution time: mean = 0.01ms, total = 0.01ms, Queueing time: mean = 6099.04ms, max = 12198.08ms, min = 12198.08ms, total = 12198.08ms
155
+ PeriodicalRunner.RunFnPeriodically - 1 total (0 active), Execution time: mean = 0.01ms, total = 0.01ms, Queueing time: mean = 0.00ms, max = 0.00ms, min = 0.00ms, total = 0.00ms
156
+ CoreWorker.HandlePushTask - 1 total (0 active), Execution time: mean = 12207.57ms, total = 12207.57ms, Queueing time: mean = 0.02ms, max = 0.02ms, min = 0.02ms, total = 0.02ms
157
+ CoreWorker.HandlePushTaskActor - 1 total (1 active, 1 running), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
158
+
159
+ -----------------
160
+ Task Event stats:
161
+
162
+ IO Service Stats:
163
+
164
+ Global stats: 181 total (1 active)
165
+ Queueing time: mean = 0.03ms, max = 0.34ms, min = 0.02ms, total = 5.24ms
166
+ Execution time: mean = 0.40ms, total = 72.68ms
167
+ Event stats:
168
+ ray::rpc::TaskInfoGcsService.grpc_client.AddTaskEventData.OnReplyReceived - 60 total (0 active), Execution time: mean = 0.04ms, total = 2.48ms, Queueing time: mean = 0.04ms, max = 0.34ms, min = 0.02ms, total = 2.39ms
169
+ CoreWorker.deadline_timer.flush_task_events - 60 total (1 active), Execution time: mean = 0.29ms, total = 17.39ms, Queueing time: mean = 0.05ms, max = 0.09ms, min = 0.02ms, total = 2.83ms
170
+ ray::rpc::TaskInfoGcsService.grpc_client.AddTaskEventData - 60 total (0 active), Execution time: mean = 0.88ms, total = 52.52ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
171
+ PeriodicalRunner.RunFnPeriodically - 1 total (0 active), Execution time: mean = 0.29ms, total = 0.29ms, Queueing time: mean = 0.02ms, max = 0.02ms, min = 0.02ms, total = 0.02ms
172
+ Other Stats:
173
+ gcs_grpc_in_progress:0
174
+ event_aggregator_grpc_in_progress:0
175
+ current number of task status events in buffer: 8
176
+ current number of profile events in buffer: 3
177
+ current number of dropped task attempts tracked: 0
178
+ total task events sent: 0.0106544 MiB
179
+ total number of task attempts sent: 26
180
+ total number of task attempts dropped reported: 0
181
+ total number of sent failure: 0
182
+ num status task events dropped: 0
183
+ num profile task events dropped: 0
184
+ num ray task events reported to aggregator: 0
185
+ num ray task events failed to report to aggregator: 0
186
+ num of task attempts dropped reported to aggregator: 0
187
+ num of failed requests to aggregator: 0
188
+
189
+ [2026-02-27 00:32:01,977 I 11896 11946] actor_manager.cc:236: received notification on actor, state: ALIVE, ip address: 10.128.0.163, port: 50267, num_restarts: 0, death context type=CONTEXT_NOT_SET actor_id=8b6cdab5e264a7c65d1ca0d701000000 worker_id=a99abd04b70eed71bbc2b85849964e1f45cdec8a7b96f35e101ab940 node_id=d9160dc27b026e787f1f81b4465a36f139cf00f32ff2820e42f20120
190
+ [2026-02-27 00:32:12,448 I 11896 11946] core_worker.cc:4087: Force kill actor request has received. exiting immediately... The actor is dead because its owner has died. Owner Id: 01000000ffffffffffffffffffffffffffffffffffffffffffffffff Owner Ip address: 10.128.0.163 Owner worker exit type: INTENDED_USER_EXIT Worker exit detail: Owner's worker process has crashed.
191
+ [2026-02-27 00:32:12,477 I 11896 11946] core_worker_shutdown_executor.cc:217: Try killing all child processes of this worker as it exits. Child process pids:
192
+ [2026-02-27 00:32:12,482 I 11896 11946] core_worker_shutdown_executor.cc:262: Sending disconnect message to the local raylet.
193
+ [2026-02-27 00:32:12,482 I 11896 11946] raylet_ipc_client.cc:135: RayletIpcClient::Disconnect, exit_type=INTENDED_SYSTEM_EXIT, exit_detail=Worker exits because the actor is killed. The actor is dead because its owner has died. Owner Id: 01000000ffffffffffffffffffffffffffffffffffffffffffffffff Owner Ip address: 10.128.0.163 Owner worker exit type: INTENDED_USER_EXIT Worker exit detail: Owner's worker process has crashed., has creation_task_exception_pb_bytes=0
194
+ [2026-02-27 00:32:12,487 I 11896 11946] core_worker_shutdown_executor.cc:279: Disconnected from the local raylet.
195
+ [2026-02-27 00:32:12,487 W 11896 11946] core_worker_shutdown_executor.cc:288: Quick exit - terminating process immediately
python-core-worker-9224fcd6abcfd04deeca6990e3ac522c58f6eec637ba09c0e927aaef_12481.log ADDED
@@ -0,0 +1,72 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2026-02-27 00:31:33,077 I 12481 12481] core_worker_process.cc:773: Constructing CoreWorkerProcess. pid: 12481
2
+ [2026-02-27 00:31:33,086 I 12481 12481] event.cc:499: Ray Event initialized for CORE_WORKER
3
+ [2026-02-27 00:31:33,089 I 12481 12481] event.cc:499: Ray Event initialized for EXPORT_TASK
4
+ [2026-02-27 00:31:33,089 I 12481 12481] event.cc:332: Set ray event level to warning
5
+ [2026-02-27 00:31:33,089 I 12481 12481] event_aggregator_client.h:50: Initiating the local event aggregator client with port: 55678
6
+ [2026-02-27 00:31:33,091 I 12481 12481] grpc_server.cc:143: worker server started, listening on port 50221.
7
+ [2026-02-27 00:31:33,109 I 12481 12481] core_worker_process.cc:261: Initializing worker at address: 10.128.0.163:50221 worker_id=9224fcd6abcfd04deeca6990e3ac522c58f6eec637ba09c0e927aaef node_id=d9160dc27b026e787f1f81b4465a36f139cf00f32ff2820e42f20120
8
+ [2026-02-27 00:31:33,110 I 12481 12481] task_event_buffer.cc:480: Reporting task events to GCS every 1000ms.
9
+ [2026-02-27 00:31:33,113 I 12481 12769] core_worker.cc:455: Event stats:
10
+
11
+
12
+ Global stats: 8 total (6 active)
13
+ Queueing time: mean = 0.01ms, max = 0.03ms, min = 0.03ms, total = 0.06ms
14
+ Execution time: mean = 0.00ms, total = 0.04ms
15
+ Event stats:
16
+ PeriodicalRunner.RunFnPeriodically - 3 total (1 active, 1 running), Execution time: mean = 0.01ms, total = 0.04ms, Queueing time: mean = 0.02ms, max = 0.03ms, min = 0.03ms, total = 0.06ms
17
+ Publisher.CheckDeadSubscribers - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
18
+ ray::rpc::WorkerInfoGcsService.grpc_client.AddWorkerInfo - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
19
+ ray::rpc::InternalPubSubGcsService.grpc_client.GcsSubscriberPoll - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
20
+ CoreWorker.ExitIfParentRayletDies - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
21
+ ray::rpc::InternalPubSubGcsService.grpc_client.GcsSubscriberCommandBatch - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
22
+
23
+ -----------------
24
+ Task execution event stats:
25
+
26
+ Global stats: 0 total (0 active)
27
+ Queueing time: mean = -nanms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
28
+ Execution time: mean = -nanms, total = 0.00ms
29
+ Event stats:
30
+
31
+ -----------------
32
+ Task Event stats:
33
+
34
+ IO Service Stats:
35
+
36
+ Global stats: 3 total (2 active)
37
+ Queueing time: mean = 0.04ms, max = 0.11ms, min = 0.11ms, total = 0.11ms
38
+ Execution time: mean = 0.09ms, total = 0.28ms
39
+ Event stats:
40
+ PeriodicalRunner.RunFnPeriodically - 1 total (0 active), Execution time: mean = 0.28ms, total = 0.28ms, Queueing time: mean = 0.11ms, max = 0.11ms, min = 0.11ms, total = 0.11ms
41
+ ray::rpc::TaskInfoGcsService.grpc_client.AddTaskEventData - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
42
+ CoreWorker.deadline_timer.flush_task_events - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
43
+ Other Stats:
44
+ gcs_grpc_in_progress:1
45
+ event_aggregator_grpc_in_progress:0
46
+ current number of task status events in buffer: 0
47
+ current number of profile events in buffer: 0
48
+ current number of dropped task attempts tracked: 0
49
+ total task events sent: 0 MiB
50
+ total number of task attempts sent: 0
51
+ total number of task attempts dropped reported: 0
52
+ total number of sent failure: 0
53
+ num status task events dropped: 0
54
+ num profile task events dropped: 0
55
+ num ray task events reported to aggregator: 0
56
+ num ray task events failed to report to aggregator: 0
57
+ num of task attempts dropped reported to aggregator: 0
58
+ num of failed requests to aggregator: 0
59
+
60
+ [2026-02-27 00:31:33,114 I 12481 12481] core_worker.cc:515: Adjusted worker niceness to 15
61
+ [2026-02-27 00:31:33,117 I 12481 12481] metrics_agent_client.cc:42: Initializing exporter ...
62
+ [2026-02-27 00:31:33,123 I 12481 12769] accessor.cc:540: Received address and liveness notification for node, IsAlive = 1 node_id=d9160dc27b026e787f1f81b4465a36f139cf00f32ff2820e42f20120
63
+ [2026-02-27 00:31:33,123 I 12481 12769] normal_task_submitter.cc:824: Number of alive nodes:1
64
+ [2026-02-27 00:31:33,125 I 12481 12481] actor_task_submitter.cc:74: Set actor max pending calls to -1 actor_id=2f225182ea5eab0591f8ce2101000000
65
+ [2026-02-27 00:31:33,125 I 12481 12481] core_worker.cc:2903: Creating actor actor_id=2f225182ea5eab0591f8ce2101000000
66
+ [2026-02-27 00:31:33,130 I 12481 12769] metrics_agent_client.cc:54: Exporter initialized.
67
+ [2026-02-27 00:31:44,873 I 12481 12481] task_receiver.cc:142: Actor creation task finished, task_id: ffffffffffffffff2f225182ea5eab0591f8ce2101000000, actor_id: 2f225182ea5eab0591f8ce2101000000, actor_repr_name:
68
+ [2026-02-27 00:32:12,487 I 12481 12769] core_worker_shutdown_executor.cc:217: Try killing all child processes of this worker as it exits. Child process pids:
69
+ [2026-02-27 00:32:12,487 I 12481 12769] core_worker_shutdown_executor.cc:262: Sending disconnect message to the local raylet.
70
+ [2026-02-27 00:32:12,487 I 12481 12769] raylet_ipc_client.cc:135: RayletIpcClient::Disconnect, exit_type=INTENDED_USER_EXIT, exit_detail=Worker force exited because its job has finished, has creation_task_exception_pb_bytes=0
71
+ [2026-02-27 00:32:12,503 I 12481 12769] core_worker_shutdown_executor.cc:279: Disconnected from the local raylet.
72
+ [2026-02-27 00:32:12,503 W 12481 12769] core_worker_shutdown_executor.cc:288: Quick exit - terminating process immediately
python-core-worker-af6e4d2eae80c226c783dd6717832e015ec8fc0144d801649c12abfe_12563.log ADDED
@@ -0,0 +1,76 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2026-02-27 00:31:33,361 I 12563 12563] core_worker_process.cc:773: Constructing CoreWorkerProcess. pid: 12563
2
+ [2026-02-27 00:31:33,369 I 12563 12563] event.cc:499: Ray Event initialized for CORE_WORKER
3
+ [2026-02-27 00:31:33,369 I 12563 12563] event.cc:499: Ray Event initialized for EXPORT_TASK
4
+ [2026-02-27 00:31:33,369 I 12563 12563] event.cc:332: Set ray event level to warning
5
+ [2026-02-27 00:31:33,369 I 12563 12563] event_aggregator_client.h:50: Initiating the local event aggregator client with port: 55678
6
+ [2026-02-27 00:31:33,372 I 12563 12563] grpc_server.cc:143: worker server started, listening on port 50485.
7
+ [2026-02-27 00:31:33,386 I 12563 12563] core_worker_process.cc:261: Initializing worker at address: 10.128.0.163:50485 worker_id=af6e4d2eae80c226c783dd6717832e015ec8fc0144d801649c12abfe node_id=d9160dc27b026e787f1f81b4465a36f139cf00f32ff2820e42f20120
8
+ [2026-02-27 00:31:33,388 I 12563 12563] task_event_buffer.cc:480: Reporting task events to GCS every 1000ms.
9
+ [2026-02-27 00:31:33,392 I 12563 12563] core_worker.cc:515: Adjusted worker niceness to 15
10
+ [2026-02-27 00:31:33,393 I 12563 12886] core_worker.cc:455: Event stats:
11
+
12
+
13
+ Global stats: 15 total (9 active)
14
+ Queueing time: mean = 0.11ms, max = 1.37ms, min = 0.03ms, total = 1.64ms
15
+ Execution time: mean = 0.13ms, total = 1.93ms
16
+ Event stats:
17
+ PeriodicalRunner.RunFnPeriodically - 7 total (5 active, 1 running), Execution time: mean = 0.00ms, total = 0.03ms, Queueing time: mean = 0.22ms, max = 1.37ms, min = 0.20ms, total = 1.57ms
18
+ Publisher.CheckDeadSubscribers - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
19
+ ray::rpc::WorkerInfoGcsService.grpc_client.AddWorkerInfo - 1 total (0 active), Execution time: mean = 0.70ms, total = 0.70ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
20
+ ray::rpc::InternalPubSubGcsService.grpc_client.GcsSubscriberCommandBatch - 1 total (0 active), Execution time: mean = 0.60ms, total = 0.60ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
21
+ ray::rpc::InternalPubSubGcsService.grpc_client.GcsSubscriberCommandBatch.OnReplyReceived - 1 total (0 active), Execution time: mean = 0.56ms, total = 0.56ms, Queueing time: mean = 0.03ms, max = 0.03ms, min = 0.03ms, total = 0.03ms
22
+ ray::rpc::NodeInfoGcsService.grpc_client.GetAllNodeAddressAndLiveness - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
23
+ ray::rpc::InternalPubSubGcsService.grpc_client.GcsSubscriberPoll - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
24
+ ray::rpc::WorkerInfoGcsService.grpc_client.AddWorkerInfo.OnReplyReceived - 1 total (0 active), Execution time: mean = 0.04ms, total = 0.04ms, Queueing time: mean = 0.04ms, max = 0.04ms, min = 0.04ms, total = 0.04ms
25
+ CoreWorker.ExitIfParentRayletDies - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
26
+
27
+ -----------------
28
+ Task execution event stats:
29
+
30
+ Global stats: 0 total (0 active)
31
+ Queueing time: mean = -nanms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
32
+ Execution time: mean = -nanms, total = 0.00ms
33
+ Event stats:
34
+
35
+ -----------------
36
+ Task Event stats:
37
+
38
+ IO Service Stats:
39
+
40
+ Global stats: 4 total (1 active)
41
+ Queueing time: mean = 0.01ms, max = 0.02ms, min = 0.02ms, total = 0.03ms
42
+ Execution time: mean = 0.49ms, total = 1.98ms
43
+ Event stats:
44
+ ray::rpc::TaskInfoGcsService.grpc_client.AddTaskEventData - 1 total (0 active), Execution time: mean = 1.16ms, total = 1.16ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
45
+ CoreWorker.deadline_timer.flush_task_events - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
46
+ PeriodicalRunner.RunFnPeriodically - 1 total (0 active), Execution time: mean = 0.79ms, total = 0.79ms, Queueing time: mean = 0.02ms, max = 0.02ms, min = 0.02ms, total = 0.02ms
47
+ ray::rpc::TaskInfoGcsService.grpc_client.AddTaskEventData.OnReplyReceived - 1 total (0 active), Execution time: mean = 0.02ms, total = 0.02ms, Queueing time: mean = 0.02ms, max = 0.02ms, min = 0.02ms, total = 0.02ms
48
+ Other Stats:
49
+ gcs_grpc_in_progress:0
50
+ event_aggregator_grpc_in_progress:0
51
+ current number of task status events in buffer: 0
52
+ current number of profile events in buffer: 0
53
+ current number of dropped task attempts tracked: 0
54
+ total task events sent: 0 MiB
55
+ total number of task attempts sent: 0
56
+ total number of task attempts dropped reported: 0
57
+ total number of sent failure: 0
58
+ num status task events dropped: 0
59
+ num profile task events dropped: 0
60
+ num ray task events reported to aggregator: 0
61
+ num ray task events failed to report to aggregator: 0
62
+ num of task attempts dropped reported to aggregator: 0
63
+ num of failed requests to aggregator: 0
64
+
65
+ [2026-02-27 00:31:33,393 I 12563 12563] metrics_agent_client.cc:42: Initializing exporter ...
66
+ [2026-02-27 00:31:33,395 I 12563 12886] accessor.cc:540: Received address and liveness notification for node, IsAlive = 1 node_id=d9160dc27b026e787f1f81b4465a36f139cf00f32ff2820e42f20120
67
+ [2026-02-27 00:31:33,395 I 12563 12886] normal_task_submitter.cc:824: Number of alive nodes:1
68
+ [2026-02-27 00:31:33,397 I 12563 12886] metrics_agent_client.cc:54: Exporter initialized.
69
+ [2026-02-27 00:31:33,398 I 12563 12563] actor_task_submitter.cc:74: Set actor max pending calls to -1 actor_id=83a8e5ec085f16d6fdcd7dc801000000
70
+ [2026-02-27 00:31:33,398 I 12563 12563] core_worker.cc:2903: Creating actor actor_id=83a8e5ec085f16d6fdcd7dc801000000
71
+ [2026-02-27 00:31:44,963 I 12563 12563] task_receiver.cc:142: Actor creation task finished, task_id: ffffffffffffffff83a8e5ec085f16d6fdcd7dc801000000, actor_id: 83a8e5ec085f16d6fdcd7dc801000000, actor_repr_name:
72
+ [2026-02-27 00:32:12,483 I 12563 12886] core_worker_shutdown_executor.cc:217: Try killing all child processes of this worker as it exits. Child process pids:
73
+ [2026-02-27 00:32:12,484 I 12563 12886] core_worker_shutdown_executor.cc:262: Sending disconnect message to the local raylet.
74
+ [2026-02-27 00:32:12,484 I 12563 12886] raylet_ipc_client.cc:135: RayletIpcClient::Disconnect, exit_type=INTENDED_USER_EXIT, exit_detail=Worker force exited because its job has finished, has creation_task_exception_pb_bytes=0
75
+ [2026-02-27 00:32:12,492 I 12563 12886] core_worker_shutdown_executor.cc:279: Disconnected from the local raylet.
76
+ [2026-02-27 00:32:12,492 W 12563 12886] core_worker_shutdown_executor.cc:288: Quick exit - terminating process immediately
python-core-worker-bb0f50c5405699ae07f957ec3f7c03f2bdf40be03f6e39b39232dc16_11378.log ADDED
@@ -0,0 +1,82 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2026-02-27 00:30:32,596 I 11378 11378] core_worker_process.cc:773: Constructing CoreWorkerProcess. pid: 11378
2
+ [2026-02-27 00:30:32,602 I 11378 11378] event.cc:499: Ray Event initialized for CORE_WORKER
3
+ [2026-02-27 00:30:32,602 I 11378 11378] event.cc:499: Ray Event initialized for EXPORT_TASK
4
+ [2026-02-27 00:30:32,602 I 11378 11378] event.cc:332: Set ray event level to warning
5
+ [2026-02-27 00:30:32,602 I 11378 11378] event_aggregator_client.h:50: Initiating the local event aggregator client with port: 55678
6
+ [2026-02-27 00:30:32,604 I 11378 11378] grpc_server.cc:143: worker server started, listening on port 50015.
7
+ [2026-02-27 00:30:32,619 I 11378 11378] core_worker_process.cc:261: Initializing worker at address: 10.128.0.163:50015 worker_id=bb0f50c5405699ae07f957ec3f7c03f2bdf40be03f6e39b39232dc16 node_id=d9160dc27b026e787f1f81b4465a36f139cf00f32ff2820e42f20120
8
+ [2026-02-27 00:30:32,620 I 11378 11378] task_event_buffer.cc:480: Reporting task events to GCS every 1000ms.
9
+ [2026-02-27 00:30:32,622 I 11378 11589] core_worker.cc:455: Event stats:
10
+
11
+
12
+ Global stats: 8 total (6 active)
13
+ Queueing time: mean = 0.01ms, max = 0.02ms, min = 0.02ms, total = 0.05ms
14
+ Execution time: mean = 0.00ms, total = 0.04ms
15
+ Event stats:
16
+ PeriodicalRunner.RunFnPeriodically - 3 total (1 active, 1 running), Execution time: mean = 0.01ms, total = 0.04ms, Queueing time: mean = 0.02ms, max = 0.02ms, min = 0.02ms, total = 0.05ms
17
+ Publisher.CheckDeadSubscribers - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
18
+ ray::rpc::WorkerInfoGcsService.grpc_client.AddWorkerInfo - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
19
+ ray::rpc::InternalPubSubGcsService.grpc_client.GcsSubscriberCommandBatch - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
20
+ CoreWorker.ExitIfParentRayletDies - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
21
+ ray::rpc::InternalPubSubGcsService.grpc_client.GcsSubscriberPoll - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
22
+
23
+ -----------------
24
+ Task execution event stats:
25
+
26
+ Global stats: 0 total (0 active)
27
+ Queueing time: mean = -nanms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
28
+ Execution time: mean = -nanms, total = 0.00ms
29
+ Event stats:
30
+
31
+ -----------------
32
+ Task Event stats:
33
+
34
+ IO Service Stats:
35
+
36
+ Global stats: 1 total (1 active)
37
+ Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
38
+ Execution time: mean = 0.00ms, total = 0.00ms
39
+ Event stats:
40
+ PeriodicalRunner.RunFnPeriodically - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
41
+ Other Stats:
42
+ gcs_grpc_in_progress:0
43
+ event_aggregator_grpc_in_progress:0
44
+ current number of task status events in buffer: 0
45
+ current number of profile events in buffer: 0
46
+ current number of dropped task attempts tracked: 0
47
+ total task events sent: 0 MiB
48
+ total number of task attempts sent: 0
49
+ total number of task attempts dropped reported: 0
50
+ total number of sent failure: 0
51
+ num status task events dropped: 0
52
+ num profile task events dropped: 0
53
+ num ray task events reported to aggregator: 0
54
+ num ray task events failed to report to aggregator: 0
55
+ num of task attempts dropped reported to aggregator: 0
56
+ num of failed requests to aggregator: 0
57
+
58
+ [2026-02-27 00:30:32,623 I 11378 11378] core_worker.cc:515: Adjusted worker niceness to 15
59
+ [2026-02-27 00:30:32,624 I 11378 11378] metrics_agent_client.cc:42: Initializing exporter ...
60
+ [2026-02-27 00:30:32,625 I 11378 11589] accessor.cc:540: Received address and liveness notification for node, IsAlive = 1 node_id=d9160dc27b026e787f1f81b4465a36f139cf00f32ff2820e42f20120
61
+ [2026-02-27 00:30:32,625 I 11378 11589] normal_task_submitter.cc:824: Number of alive nodes:1
62
+ [2026-02-27 00:30:34,626 I 11378 11589] metrics_agent_client.cc:54: Exporter initialized.
63
+ [2026-02-27 00:30:57,500 I 11378 11589] core_worker_shutdown_executor.cc:184: Executing handle exit: INTENDED_SYSTEM_EXIT - Worker exited because it was idle for a long time (timeout: -1ms)
64
+ [2026-02-27 00:30:57,500 I 11378 11589] core_worker_shutdown_executor.cc:94: Executing worker exit: INTENDED_SYSTEM_EXIT - Worker exited because it was idle for a long time (timeout: 10000ms)
65
+ [2026-02-27 00:30:57,500 I 11378 11378] core_worker_shutdown_executor.cc:128: Wait for currently executing tasks in the underlying thread pools to finish.
66
+ [2026-02-27 00:30:57,500 I 11378 11378] core_worker_shutdown_executor.cc:162: Releasing local references, then draining reference counter.
67
+ [2026-02-27 00:30:57,505 I 11378 11378] core_worker_shutdown_executor.cc:217: Try killing all child processes of this worker as it exits. Child process pids:
68
+ [2026-02-27 00:30:57,505 I 11378 11378] core_worker_shutdown_executor.cc:262: Sending disconnect message to the local raylet.
69
+ [2026-02-27 00:30:57,506 I 11378 11378] raylet_ipc_client.cc:135: RayletIpcClient::Disconnect, exit_type=INTENDED_SYSTEM_EXIT, exit_detail=Worker exited because it was idle for a long time, has creation_task_exception_pb_bytes=0
70
+ [2026-02-27 00:30:57,507 I 11378 11378] core_worker_shutdown_executor.cc:279: Disconnected from the local raylet.
71
+ [2026-02-27 00:30:57,507 I 11378 11378] task_event_buffer.cc:491: Shutting down TaskEventBuffer.
72
+ [2026-02-27 00:30:57,508 I 11378 11597] task_event_buffer.cc:459: Task event buffer io service stopped.
73
+ [2026-02-27 00:30:57,509 I 11378 11378] core_worker_shutdown_executor.cc:54: Waiting for joining a core worker io thread. If it hangs here, there might be deadlock or a high load in the core worker io service.
74
+ [2026-02-27 00:30:57,509 I 11378 11589] core_worker_process.cc:194: Core worker main io service stopped.
75
+ [2026-02-27 00:30:57,512 I 11378 11378] core_worker_shutdown_executor.cc:72: Disconnecting a GCS client.
76
+ [2026-02-27 00:30:57,512 I 11378 11378] core_worker_shutdown_executor.cc:79: Core worker ready to be deallocated.
77
+ [2026-02-27 00:30:57,512 I 11378 11378] core_worker_process.cc:950: Task execution loop terminated. Removing the global worker.
78
+ [2026-02-27 00:30:57,512 I 11378 11378] core_worker.cc:539: Core worker is destructed
79
+ [2026-02-27 00:30:57,512 I 11378 11378] task_event_buffer.cc:491: Shutting down TaskEventBuffer.
80
+ [2026-02-27 00:30:57,760 I 11378 11378] core_worker_process.cc:846: Destructing CoreWorkerProcessImpl. pid: 11378
81
+ [2026-02-27 00:30:57,762 I 11378 11378] stats.h:149: Stats module has shutdown.
82
+ [2026-02-27 00:30:57,785 W 11378 11378] core_worker_process.cc:860: The core worker process is not initialized yet or already shutdown.
python-core-worker-e98598eeddae739fb0211beef22a201ab9028016b2b64fe185d8c813_12584.log ADDED
@@ -0,0 +1,72 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2026-02-27 00:31:33,233 I 12584 12584] core_worker_process.cc:773: Constructing CoreWorkerProcess. pid: 12584
2
+ [2026-02-27 00:31:33,240 I 12584 12584] event.cc:499: Ray Event initialized for CORE_WORKER
3
+ [2026-02-27 00:31:33,240 I 12584 12584] event.cc:499: Ray Event initialized for EXPORT_TASK
4
+ [2026-02-27 00:31:33,240 I 12584 12584] event.cc:332: Set ray event level to warning
5
+ [2026-02-27 00:31:33,240 I 12584 12584] event_aggregator_client.h:50: Initiating the local event aggregator client with port: 55678
6
+ [2026-02-27 00:31:33,242 I 12584 12584] grpc_server.cc:143: worker server started, listening on port 50207.
7
+ [2026-02-27 00:31:33,257 I 12584 12584] core_worker_process.cc:261: Initializing worker at address: 10.128.0.163:50207 worker_id=e98598eeddae739fb0211beef22a201ab9028016b2b64fe185d8c813 node_id=d9160dc27b026e787f1f81b4465a36f139cf00f32ff2820e42f20120
8
+ [2026-02-27 00:31:33,258 I 12584 12584] task_event_buffer.cc:480: Reporting task events to GCS every 1000ms.
9
+ [2026-02-27 00:31:33,260 I 12584 12584] core_worker.cc:515: Adjusted worker niceness to 15
10
+ [2026-02-27 00:31:33,260 I 12584 12821] core_worker.cc:455: Event stats:
11
+
12
+
13
+ Global stats: 12 total (10 active)
14
+ Queueing time: mean = 0.01ms, max = 0.06ms, min = 0.03ms, total = 0.09ms
15
+ Execution time: mean = 0.00ms, total = 0.05ms
16
+ Event stats:
17
+ PeriodicalRunner.RunFnPeriodically - 7 total (5 active, 1 running), Execution time: mean = 0.01ms, total = 0.05ms, Queueing time: mean = 0.01ms, max = 0.06ms, min = 0.03ms, total = 0.09ms
18
+ Publisher.CheckDeadSubscribers - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
19
+ ray::rpc::WorkerInfoGcsService.grpc_client.AddWorkerInfo - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
20
+ ray::rpc::InternalPubSubGcsService.grpc_client.GcsSubscriberCommandBatch - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
21
+ ray::rpc::InternalPubSubGcsService.grpc_client.GcsSubscriberPoll - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
22
+ CoreWorker.ExitIfParentRayletDies - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
23
+
24
+ -----------------
25
+ Task execution event stats:
26
+
27
+ Global stats: 0 total (0 active)
28
+ Queueing time: mean = -nanms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
29
+ Execution time: mean = -nanms, total = 0.00ms
30
+ Event stats:
31
+
32
+ -----------------
33
+ Task Event stats:
34
+
35
+ IO Service Stats:
36
+
37
+ Global stats: 3 total (2 active)
38
+ Queueing time: mean = 0.45ms, max = 1.35ms, min = 1.35ms, total = 1.35ms
39
+ Execution time: mean = 0.10ms, total = 0.29ms
40
+ Event stats:
41
+ PeriodicalRunner.RunFnPeriodically - 1 total (0 active), Execution time: mean = 0.29ms, total = 0.29ms, Queueing time: mean = 1.35ms, max = 1.35ms, min = 1.35ms, total = 1.35ms
42
+ ray::rpc::TaskInfoGcsService.grpc_client.AddTaskEventData - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
43
+ CoreWorker.deadline_timer.flush_task_events - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
44
+ Other Stats:
45
+ gcs_grpc_in_progress:1
46
+ event_aggregator_grpc_in_progress:0
47
+ current number of task status events in buffer: 0
48
+ current number of profile events in buffer: 0
49
+ current number of dropped task attempts tracked: 0
50
+ total task events sent: 0 MiB
51
+ total number of task attempts sent: 0
52
+ total number of task attempts dropped reported: 0
53
+ total number of sent failure: 0
54
+ num status task events dropped: 0
55
+ num profile task events dropped: 0
56
+ num ray task events reported to aggregator: 0
57
+ num ray task events failed to report to aggregator: 0
58
+ num of task attempts dropped reported to aggregator: 0
59
+ num of failed requests to aggregator: 0
60
+
61
+ [2026-02-27 00:31:33,261 I 12584 12584] metrics_agent_client.cc:42: Initializing exporter ...
62
+ [2026-02-27 00:31:33,263 I 12584 12821] accessor.cc:540: Received address and liveness notification for node, IsAlive = 1 node_id=d9160dc27b026e787f1f81b4465a36f139cf00f32ff2820e42f20120
63
+ [2026-02-27 00:31:33,263 I 12584 12821] normal_task_submitter.cc:824: Number of alive nodes:1
64
+ [2026-02-27 00:31:33,266 I 12584 12821] metrics_agent_client.cc:54: Exporter initialized.
65
+ [2026-02-27 00:31:33,273 I 12584 12584] actor_task_submitter.cc:74: Set actor max pending calls to -1 actor_id=87ab4f0b8032c66348227f5a01000000
66
+ [2026-02-27 00:31:33,273 I 12584 12584] core_worker.cc:2903: Creating actor actor_id=87ab4f0b8032c66348227f5a01000000
67
+ [2026-02-27 00:31:44,904 I 12584 12584] task_receiver.cc:142: Actor creation task finished, task_id: ffffffffffffffff87ab4f0b8032c66348227f5a01000000, actor_id: 87ab4f0b8032c66348227f5a01000000, actor_repr_name:
68
+ [2026-02-27 00:32:12,472 I 12584 12821] core_worker_shutdown_executor.cc:217: Try killing all child processes of this worker as it exits. Child process pids:
69
+ [2026-02-27 00:32:12,472 I 12584 12821] core_worker_shutdown_executor.cc:262: Sending disconnect message to the local raylet.
70
+ [2026-02-27 00:32:12,472 I 12584 12821] raylet_ipc_client.cc:135: RayletIpcClient::Disconnect, exit_type=INTENDED_USER_EXIT, exit_detail=Worker force exited because its job has finished, has creation_task_exception_pb_bytes=0
71
+ [2026-02-27 00:32:12,476 I 12584 12821] core_worker_shutdown_executor.cc:279: Disconnected from the local raylet.
72
+ [2026-02-27 00:32:12,476 W 12584 12821] core_worker_shutdown_executor.cc:288: Quick exit - terminating process immediately
python-core-worker-f46103e29121f0b748164b47d1653310da7f304c2c8c8df73871f0e5_12507.log ADDED
@@ -0,0 +1,72 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2026-02-27 00:31:33,383 I 12507 12507] core_worker_process.cc:773: Constructing CoreWorkerProcess. pid: 12507
2
+ [2026-02-27 00:31:33,386 I 12507 12507] event.cc:499: Ray Event initialized for CORE_WORKER
3
+ [2026-02-27 00:31:33,387 I 12507 12507] event.cc:499: Ray Event initialized for EXPORT_TASK
4
+ [2026-02-27 00:31:33,387 I 12507 12507] event.cc:332: Set ray event level to warning
5
+ [2026-02-27 00:31:33,387 I 12507 12507] event_aggregator_client.h:50: Initiating the local event aggregator client with port: 55678
6
+ [2026-02-27 00:31:33,389 I 12507 12507] grpc_server.cc:143: worker server started, listening on port 50291.
7
+ [2026-02-27 00:31:33,413 I 12507 12507] core_worker_process.cc:261: Initializing worker at address: 10.128.0.163:50291 worker_id=f46103e29121f0b748164b47d1653310da7f304c2c8c8df73871f0e5 node_id=d9160dc27b026e787f1f81b4465a36f139cf00f32ff2820e42f20120
8
+ [2026-02-27 00:31:33,414 I 12507 12507] task_event_buffer.cc:480: Reporting task events to GCS every 1000ms.
9
+ [2026-02-27 00:31:33,416 I 12507 12507] core_worker.cc:515: Adjusted worker niceness to 15
10
+ [2026-02-27 00:31:33,416 I 12507 12923] core_worker.cc:455: Event stats:
11
+
12
+
13
+ Global stats: 12 total (10 active)
14
+ Queueing time: mean = 0.01ms, max = 0.03ms, min = 0.03ms, total = 0.07ms
15
+ Execution time: mean = 0.00ms, total = 0.04ms
16
+ Event stats:
17
+ PeriodicalRunner.RunFnPeriodically - 7 total (5 active, 1 running), Execution time: mean = 0.01ms, total = 0.04ms, Queueing time: mean = 0.01ms, max = 0.03ms, min = 0.03ms, total = 0.07ms
18
+ Publisher.CheckDeadSubscribers - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
19
+ ray::rpc::InternalPubSubGcsService.grpc_client.GcsSubscriberPoll - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
20
+ CoreWorker.ExitIfParentRayletDies - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
21
+ ray::rpc::WorkerInfoGcsService.grpc_client.AddWorkerInfo - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
22
+ ray::rpc::InternalPubSubGcsService.grpc_client.GcsSubscriberCommandBatch - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
23
+
24
+ -----------------
25
+ Task execution event stats:
26
+
27
+ Global stats: 0 total (0 active)
28
+ Queueing time: mean = -nanms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
29
+ Execution time: mean = -nanms, total = 0.00ms
30
+ Event stats:
31
+
32
+ -----------------
33
+ Task Event stats:
34
+
35
+ IO Service Stats:
36
+
37
+ Global stats: 3 total (2 active)
38
+ Queueing time: mean = 0.52ms, max = 1.57ms, min = 1.57ms, total = 1.57ms
39
+ Execution time: mean = 0.08ms, total = 0.24ms
40
+ Event stats:
41
+ CoreWorker.deadline_timer.flush_task_events - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
42
+ PeriodicalRunner.RunFnPeriodically - 1 total (0 active), Execution time: mean = 0.24ms, total = 0.24ms, Queueing time: mean = 1.57ms, max = 1.57ms, min = 1.57ms, total = 1.57ms
43
+ ray::rpc::TaskInfoGcsService.grpc_client.AddTaskEventData - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
44
+ Other Stats:
45
+ gcs_grpc_in_progress:1
46
+ event_aggregator_grpc_in_progress:0
47
+ current number of task status events in buffer: 0
48
+ current number of profile events in buffer: 0
49
+ current number of dropped task attempts tracked: 0
50
+ total task events sent: 0 MiB
51
+ total number of task attempts sent: 0
52
+ total number of task attempts dropped reported: 0
53
+ total number of sent failure: 0
54
+ num status task events dropped: 0
55
+ num profile task events dropped: 0
56
+ num ray task events reported to aggregator: 0
57
+ num ray task events failed to report to aggregator: 0
58
+ num of task attempts dropped reported to aggregator: 0
59
+ num of failed requests to aggregator: 0
60
+
61
+ [2026-02-27 00:31:33,416 I 12507 12507] metrics_agent_client.cc:42: Initializing exporter ...
62
+ [2026-02-27 00:31:33,420 I 12507 12923] accessor.cc:540: Received address and liveness notification for node, IsAlive = 1 node_id=d9160dc27b026e787f1f81b4465a36f139cf00f32ff2820e42f20120
63
+ [2026-02-27 00:31:33,421 I 12507 12923] normal_task_submitter.cc:824: Number of alive nodes:1
64
+ [2026-02-27 00:31:33,422 I 12507 12923] metrics_agent_client.cc:54: Exporter initialized.
65
+ [2026-02-27 00:31:33,423 I 12507 12507] actor_task_submitter.cc:74: Set actor max pending calls to -1 actor_id=a0546db81b53339607d0f80c01000000
66
+ [2026-02-27 00:31:33,423 I 12507 12507] core_worker.cc:2903: Creating actor actor_id=a0546db81b53339607d0f80c01000000
67
+ [2026-02-27 00:31:44,840 I 12507 12507] task_receiver.cc:142: Actor creation task finished, task_id: ffffffffffffffffa0546db81b53339607d0f80c01000000, actor_id: a0546db81b53339607d0f80c01000000, actor_repr_name:
68
+ [2026-02-27 00:32:12,464 I 12507 12923] core_worker_shutdown_executor.cc:217: Try killing all child processes of this worker as it exits. Child process pids:
69
+ [2026-02-27 00:32:12,464 I 12507 12923] core_worker_shutdown_executor.cc:262: Sending disconnect message to the local raylet.
70
+ [2026-02-27 00:32:12,464 I 12507 12923] raylet_ipc_client.cc:135: RayletIpcClient::Disconnect, exit_type=INTENDED_USER_EXIT, exit_detail=Worker force exited because its job has finished, has creation_task_exception_pb_bytes=0
71
+ [2026-02-27 00:32:12,469 I 12507 12923] core_worker_shutdown_executor.cc:279: Disconnected from the local raylet.
72
+ [2026-02-27 00:32:12,469 W 12507 12923] core_worker_shutdown_executor.cc:288: Quick exit - terminating process immediately
python-core-worker-fc14e0d4e4b6acb4ecead813c2d960587eefa7859aac6d8e19aeec98_11374.log ADDED
@@ -0,0 +1,84 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2026-02-27 00:30:32,719 I 11374 11374] core_worker_process.cc:773: Constructing CoreWorkerProcess. pid: 11374
2
+ [2026-02-27 00:30:32,721 I 11374 11374] event.cc:499: Ray Event initialized for CORE_WORKER
3
+ [2026-02-27 00:30:32,722 I 11374 11374] event.cc:499: Ray Event initialized for EXPORT_TASK
4
+ [2026-02-27 00:30:32,722 I 11374 11374] event.cc:332: Set ray event level to warning
5
+ [2026-02-27 00:30:32,722 I 11374 11374] event_aggregator_client.h:50: Initiating the local event aggregator client with port: 55678
6
+ [2026-02-27 00:30:32,723 I 11374 11374] grpc_server.cc:143: worker server started, listening on port 50037.
7
+ [2026-02-27 00:30:32,733 I 11374 11374] core_worker_process.cc:261: Initializing worker at address: 10.128.0.163:50037 worker_id=fc14e0d4e4b6acb4ecead813c2d960587eefa7859aac6d8e19aeec98 node_id=d9160dc27b026e787f1f81b4465a36f139cf00f32ff2820e42f20120
8
+ [2026-02-27 00:30:32,734 I 11374 11374] task_event_buffer.cc:480: Reporting task events to GCS every 1000ms.
9
+ [2026-02-27 00:30:32,735 I 11374 11729] core_worker.cc:455: Event stats:
10
+
11
+
12
+ Global stats: 8 total (6 active)
13
+ Queueing time: mean = 0.01ms, max = 0.04ms, min = 0.02ms, total = 0.07ms
14
+ Execution time: mean = 0.02ms, total = 0.17ms
15
+ Event stats:
16
+ PeriodicalRunner.RunFnPeriodically - 3 total (1 active, 1 running), Execution time: mean = 0.06ms, total = 0.17ms, Queueing time: mean = 0.02ms, max = 0.04ms, min = 0.02ms, total = 0.07ms
17
+ ray::rpc::WorkerInfoGcsService.grpc_client.AddWorkerInfo - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
18
+ Publisher.CheckDeadSubscribers - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
19
+ ray::rpc::InternalPubSubGcsService.grpc_client.GcsSubscriberCommandBatch - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
20
+ ray::rpc::InternalPubSubGcsService.grpc_client.GcsSubscriberPoll - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
21
+ CoreWorker.ExitIfParentRayletDies - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
22
+
23
+ -----------------
24
+ Task execution event stats:
25
+
26
+ Global stats: 0 total (0 active)
27
+ Queueing time: mean = -nanms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
28
+ Execution time: mean = -nanms, total = 0.00ms
29
+ Event stats:
30
+
31
+ -----------------
32
+ Task Event stats:
33
+
34
+ IO Service Stats:
35
+
36
+ Global stats: 4 total (1 active)
37
+ Queueing time: mean = 0.01ms, max = 0.01ms, min = 0.01ms, total = 0.03ms
38
+ Execution time: mean = 0.20ms, total = 0.80ms
39
+ Event stats:
40
+ CoreWorker.deadline_timer.flush_task_events - 1 total (1 active), Execution time: mean = 0.00ms, total = 0.00ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
41
+ ray::rpc::TaskInfoGcsService.grpc_client.AddTaskEventData - 1 total (0 active), Execution time: mean = 0.56ms, total = 0.56ms, Queueing time: mean = 0.00ms, max = -0.00ms, min = 9223372036854.78ms, total = 0.00ms
42
+ PeriodicalRunner.RunFnPeriodically - 1 total (0 active), Execution time: mean = 0.22ms, total = 0.22ms, Queueing time: mean = 0.01ms, max = 0.01ms, min = 0.01ms, total = 0.01ms
43
+ ray::rpc::TaskInfoGcsService.grpc_client.AddTaskEventData.OnReplyReceived - 1 total (0 active), Execution time: mean = 0.02ms, total = 0.02ms, Queueing time: mean = 0.01ms, max = 0.01ms, min = 0.01ms, total = 0.01ms
44
+ Other Stats:
45
+ gcs_grpc_in_progress:0
46
+ event_aggregator_grpc_in_progress:0
47
+ current number of task status events in buffer: 0
48
+ current number of profile events in buffer: 0
49
+ current number of dropped task attempts tracked: 0
50
+ total task events sent: 0 MiB
51
+ total number of task attempts sent: 0
52
+ total number of task attempts dropped reported: 0
53
+ total number of sent failure: 0
54
+ num status task events dropped: 0
55
+ num profile task events dropped: 0
56
+ num ray task events reported to aggregator: 0
57
+ num ray task events failed to report to aggregator: 0
58
+ num of task attempts dropped reported to aggregator: 0
59
+ num of failed requests to aggregator: 0
60
+
61
+ [2026-02-27 00:30:32,735 I 11374 11374] core_worker.cc:515: Adjusted worker niceness to 15
62
+ [2026-02-27 00:30:32,736 I 11374 11374] metrics_agent_client.cc:42: Initializing exporter ...
63
+ [2026-02-27 00:30:32,736 I 11374 11729] accessor.cc:540: Received address and liveness notification for node, IsAlive = 1 node_id=d9160dc27b026e787f1f81b4465a36f139cf00f32ff2820e42f20120
64
+ [2026-02-27 00:30:32,736 I 11374 11729] normal_task_submitter.cc:824: Number of alive nodes:1
65
+ [2026-02-27 00:30:33,892 I 11374 11729] core_worker_shutdown_executor.cc:184: Executing handle exit: INTENDED_SYSTEM_EXIT - Worker exited because it was idle for a long time (timeout: -1ms)
66
+ [2026-02-27 00:30:33,893 I 11374 11729] core_worker_shutdown_executor.cc:94: Executing worker exit: INTENDED_SYSTEM_EXIT - Worker exited because it was idle for a long time (timeout: 10000ms)
67
+ [2026-02-27 00:30:33,893 I 11374 11374] core_worker_shutdown_executor.cc:128: Wait for currently executing tasks in the underlying thread pools to finish.
68
+ [2026-02-27 00:30:33,893 I 11374 11374] core_worker_shutdown_executor.cc:162: Releasing local references, then draining reference counter.
69
+ [2026-02-27 00:30:33,897 I 11374 11374] core_worker_shutdown_executor.cc:217: Try killing all child processes of this worker as it exits. Child process pids:
70
+ [2026-02-27 00:30:33,897 I 11374 11374] core_worker_shutdown_executor.cc:262: Sending disconnect message to the local raylet.
71
+ [2026-02-27 00:30:33,898 I 11374 11374] raylet_ipc_client.cc:135: RayletIpcClient::Disconnect, exit_type=INTENDED_SYSTEM_EXIT, exit_detail=Worker exited because it was idle for a long time, has creation_task_exception_pb_bytes=0
72
+ [2026-02-27 00:30:33,899 I 11374 11374] core_worker_shutdown_executor.cc:279: Disconnected from the local raylet.
73
+ [2026-02-27 00:30:33,899 I 11374 11374] task_event_buffer.cc:491: Shutting down TaskEventBuffer.
74
+ [2026-02-27 00:30:33,899 I 11374 11737] task_event_buffer.cc:459: Task event buffer io service stopped.
75
+ [2026-02-27 00:30:33,899 I 11374 11374] core_worker_shutdown_executor.cc:54: Waiting for joining a core worker io thread. If it hangs here, there might be deadlock or a high load in the core worker io service.
76
+ [2026-02-27 00:30:33,899 I 11374 11729] core_worker_process.cc:194: Core worker main io service stopped.
77
+ [2026-02-27 00:30:33,902 I 11374 11374] core_worker_shutdown_executor.cc:72: Disconnecting a GCS client.
78
+ [2026-02-27 00:30:33,902 I 11374 11374] core_worker_shutdown_executor.cc:79: Core worker ready to be deallocated.
79
+ [2026-02-27 00:30:33,902 I 11374 11374] core_worker_process.cc:950: Task execution loop terminated. Removing the global worker.
80
+ [2026-02-27 00:30:33,903 I 11374 11374] core_worker.cc:539: Core worker is destructed
81
+ [2026-02-27 00:30:33,903 I 11374 11374] task_event_buffer.cc:491: Shutting down TaskEventBuffer.
82
+ [2026-02-27 00:30:33,903 I 11374 11374] core_worker_process.cc:846: Destructing CoreWorkerProcessImpl. pid: 11374
83
+ [2026-02-27 00:30:33,903 I 11374 11374] stats.h:149: Stats module has shutdown.
84
+ [2026-02-27 00:30:33,934 W 11374 11374] core_worker_process.cc:860: The core worker process is not initialized yet or already shutdown.
runtime_env_agent.err ADDED
File without changes
runtime_env_agent.log ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2026-02-27 00:30:32,551 INFO runtime_env_agent.py:193 -- Starting runtime env agent at pid 11359
2
+ 2026-02-27 00:30:32,551 INFO runtime_env_agent.py:194 -- Parent raylet pid is 11302
3
+ 2026-02-27 00:30:32,555 INFO runtime_env_agent.py:250 -- Listening to address 10.128.0.163, port 57999
4
+ 2026-02-27 00:30:32,581 INFO runtime_env_agent.py:378 -- Creating runtime env: {"env_vars": {"CUDA_DEVICE_MAX_CONNECTIONS": "1", "HCCL_HOST_SOCKET_PORT_RANGE": "auto", "HCCL_NPU_SOCKET_PORT_RANGE": "auto", "NCCL_CUMEM_ENABLE": "0", "NCCL_DEBUG": "WARN", "TOKENIZERS_PARALLELISM": "true", "VLLM_ALLOW_RUNTIME_LORA_UPDATING": "true", "VLLM_DISABLE_COMPILE_CACHE": "1", "VLLM_LOGGING_LEVEL": "WARN"}} with timeout 600 seconds.
5
+ 2026-02-27 00:30:32,582 INFO runtime_env_agent.py:428 -- Successfully created runtime env: {"env_vars": {"CUDA_DEVICE_MAX_CONNECTIONS": "1", "HCCL_HOST_SOCKET_PORT_RANGE": "auto", "HCCL_NPU_SOCKET_PORT_RANGE": "auto", "NCCL_CUMEM_ENABLE": "0", "NCCL_DEBUG": "WARN", "TOKENIZERS_PARALLELISM": "true", "VLLM_ALLOW_RUNTIME_LORA_UPDATING": "true", "VLLM_DISABLE_COMPILE_CACHE": "1", "VLLM_LOGGING_LEVEL": "WARN"}}, context: {"command_prefix": [], "env_vars": {"CUDA_DEVICE_MAX_CONNECTIONS": "1", "HCCL_HOST_SOCKET_PORT_RANGE": "auto", "HCCL_NPU_SOCKET_PORT_RANGE": "auto", "NCCL_CUMEM_ENABLE": "0", "NCCL_DEBUG": "WARN", "TOKENIZERS_PARALLELISM": "true", "VLLM_ALLOW_RUNTIME_LORA_UPDATING": "true", "VLLM_DISABLE_COMPILE_CACHE": "1", "VLLM_LOGGING_LEVEL": "WARN"}, "py_executable": "/usr/bin/python3", "override_worker_entrypoint": null, "java_jars": []}
6
+ 2026-02-27 00:30:32,872 INFO runtime_env_agent.py:378 -- Creating runtime env: {"_nsight":{"cuda-graph-trace":"graph","cuda-memory-usage":"true","trace":"cuda,nvtx,cublas,ucx"},"env_vars":{"CUDA_DEVICE_MAX_CONNECTIONS":"1","HCCL_HOST_SOCKET_PORT_RANGE":"auto","HCCL_NPU_SOCKET_PORT_RANGE":"auto","NCCL_CUMEM_ENABLE":"0","NCCL_DEBUG":"WARN","TOKENIZERS_PARALLELISM":"true","VLLM_ALLOW_RUNTIME_LORA_UPDATING":"true","VLLM_DISABLE_COMPILE_CACHE":"1","VLLM_LOGGING_LEVEL":"WARN"}} with timeout 600 seconds.
7
+ 2026-02-27 00:30:36,525 INFO runtime_env_agent.py:428 -- Successfully created runtime env: {"_nsight":{"cuda-graph-trace":"graph","cuda-memory-usage":"true","trace":"cuda,nvtx,cublas,ucx"},"env_vars":{"CUDA_DEVICE_MAX_CONNECTIONS":"1","HCCL_HOST_SOCKET_PORT_RANGE":"auto","HCCL_NPU_SOCKET_PORT_RANGE":"auto","NCCL_CUMEM_ENABLE":"0","NCCL_DEBUG":"WARN","TOKENIZERS_PARALLELISM":"true","VLLM_ALLOW_RUNTIME_LORA_UPDATING":"true","VLLM_DISABLE_COMPILE_CACHE":"1","VLLM_LOGGING_LEVEL":"WARN"}}, context: {"command_prefix": [], "env_vars": {"CUDA_DEVICE_MAX_CONNECTIONS": "1", "HCCL_HOST_SOCKET_PORT_RANGE": "auto", "HCCL_NPU_SOCKET_PORT_RANGE": "auto", "NCCL_CUMEM_ENABLE": "0", "NCCL_DEBUG": "WARN", "TOKENIZERS_PARALLELISM": "true", "VLLM_ALLOW_RUNTIME_LORA_UPDATING": "true", "VLLM_DISABLE_COMPILE_CACHE": "1", "VLLM_LOGGING_LEVEL": "WARN"}, "py_executable": "nsys profile --cuda-graph-trace=graph --cuda-memory-usage=true --trace=cuda,nvtx,cublas,ucx -o /tmp/ray/session_2026-02-27_00-30-26_175126_10593/logs/nsight/'worker_process_%p' python", "override_worker_entrypoint": null, "java_jars": []}
8
+ 2026-02-27 00:30:57,313 INFO runtime_env_agent.py:467 -- Runtime env already created successfully. Env: {"_nsight":{"cuda-graph-trace":"graph","cuda-memory-usage":"true","trace":"cuda,nvtx,cublas,ucx"},"env_vars":{"CUDA_DEVICE_MAX_CONNECTIONS":"1","HCCL_HOST_SOCKET_PORT_RANGE":"auto","HCCL_NPU_SOCKET_PORT_RANGE":"auto","NCCL_CUMEM_ENABLE":"0","NCCL_DEBUG":"WARN","TOKENIZERS_PARALLELISM":"true","VLLM_ALLOW_RUNTIME_LORA_UPDATING":"true","VLLM_DISABLE_COMPILE_CACHE":"1","VLLM_LOGGING_LEVEL":"WARN"}}, context: {"command_prefix": [], "env_vars": {"CUDA_DEVICE_MAX_CONNECTIONS": "1", "HCCL_HOST_SOCKET_PORT_RANGE": "auto", "HCCL_NPU_SOCKET_PORT_RANGE": "auto", "NCCL_CUMEM_ENABLE": "0", "NCCL_DEBUG": "WARN", "TOKENIZERS_PARALLELISM": "true", "VLLM_ALLOW_RUNTIME_LORA_UPDATING": "true", "VLLM_DISABLE_COMPILE_CACHE": "1", "VLLM_LOGGING_LEVEL": "WARN"}, "py_executable": "nsys profile --cuda-graph-trace=graph --cuda-memory-usage=true --trace=cuda,nvtx,cublas,ucx -o /tmp/ray/session_2026-02-27_00-30-26_175126_10593/logs/nsight/'worker_process_%p' python", "override_worker_entrypoint": null, "java_jars": []}
9
+ 2026-02-27 00:30:59,669 INFO runtime_env_agent.py:378 -- Creating runtime env: {"_nsight":{"capture-range":"cudaProfilerApi","capture-range-end":"repeat-shutdown:24","cuda-graph-trace":"graph","cuda-memory-usage":"true","kill":"none","trace":"cuda,nvtx,osrt"},"env_vars":{"CUDA_DEVICE_MAX_CONNECTIONS":"1","HCCL_HOST_SOCKET_PORT_RANGE":"auto","HCCL_NPU_SOCKET_PORT_RANGE":"auto","MASTER_ADDR":"10.128.0.163","MASTER_PORT":"50347","NCCL_CUMEM_ENABLE":"0","NCCL_DEBUG":"WARN","RANK":"0","RAY_LOCAL_WORLD_SIZE":"1","TOKENIZERS_PARALLELISM":"true","VLLM_ALLOW_RUNTIME_LORA_UPDATING":"true","VLLM_DISABLE_COMPILE_CACHE":"1","VLLM_LOGGING_LEVEL":"WARN","WG_BACKEND":"ray","WG_PREFIX":"B8Qm6M","WORLD_SIZE":"1"}} with timeout 600 seconds.
10
+ 2026-02-27 00:31:00,166 INFO runtime_env_agent.py:428 -- Successfully created runtime env: {"_nsight":{"capture-range":"cudaProfilerApi","capture-range-end":"repeat-shutdown:24","cuda-graph-trace":"graph","cuda-memory-usage":"true","kill":"none","trace":"cuda,nvtx,osrt"},"env_vars":{"CUDA_DEVICE_MAX_CONNECTIONS":"1","HCCL_HOST_SOCKET_PORT_RANGE":"auto","HCCL_NPU_SOCKET_PORT_RANGE":"auto","MASTER_ADDR":"10.128.0.163","MASTER_PORT":"50347","NCCL_CUMEM_ENABLE":"0","NCCL_DEBUG":"WARN","RANK":"0","RAY_LOCAL_WORLD_SIZE":"1","TOKENIZERS_PARALLELISM":"true","VLLM_ALLOW_RUNTIME_LORA_UPDATING":"true","VLLM_DISABLE_COMPILE_CACHE":"1","VLLM_LOGGING_LEVEL":"WARN","WG_BACKEND":"ray","WG_PREFIX":"B8Qm6M","WORLD_SIZE":"1"}}, context: {"command_prefix": [], "env_vars": {"CUDA_DEVICE_MAX_CONNECTIONS": "1", "HCCL_HOST_SOCKET_PORT_RANGE": "auto", "HCCL_NPU_SOCKET_PORT_RANGE": "auto", "MASTER_ADDR": "10.128.0.163", "MASTER_PORT": "50347", "NCCL_CUMEM_ENABLE": "0", "NCCL_DEBUG": "WARN", "RANK": "0", "RAY_LOCAL_WORLD_SIZE": "1", "TOKENIZERS_PARALLELISM": "true", "VLLM_ALLOW_RUNTIME_LORA_UPDATING": "true", "VLLM_DISABLE_COMPILE_CACHE": "1", "VLLM_LOGGING_LEVEL": "WARN", "WG_BACKEND": "ray", "WG_PREFIX": "B8Qm6M", "WORLD_SIZE": "1"}, "py_executable": "nsys profile --capture-range=cudaProfilerApi --capture-range-end=repeat-shutdown:24 --cuda-graph-trace=graph --cuda-memory-usage=true --kill=none --trace=cuda,nvtx,osrt -o /tmp/ray/session_2026-02-27_00-30-26_175126_10593/logs/nsight/'worker_process_%p' python", "override_worker_entrypoint": null, "java_jars": []}
11
+ 2026-02-27 00:31:29,864 INFO runtime_env_agent.py:467 -- Runtime env already created successfully. Env: {"_nsight":{"cuda-graph-trace":"graph","cuda-memory-usage":"true","trace":"cuda,nvtx,cublas,ucx"},"env_vars":{"CUDA_DEVICE_MAX_CONNECTIONS":"1","HCCL_HOST_SOCKET_PORT_RANGE":"auto","HCCL_NPU_SOCKET_PORT_RANGE":"auto","NCCL_CUMEM_ENABLE":"0","NCCL_DEBUG":"WARN","TOKENIZERS_PARALLELISM":"true","VLLM_ALLOW_RUNTIME_LORA_UPDATING":"true","VLLM_DISABLE_COMPILE_CACHE":"1","VLLM_LOGGING_LEVEL":"WARN"}}, context: {"command_prefix": [], "env_vars": {"CUDA_DEVICE_MAX_CONNECTIONS": "1", "HCCL_HOST_SOCKET_PORT_RANGE": "auto", "HCCL_NPU_SOCKET_PORT_RANGE": "auto", "NCCL_CUMEM_ENABLE": "0", "NCCL_DEBUG": "WARN", "TOKENIZERS_PARALLELISM": "true", "VLLM_ALLOW_RUNTIME_LORA_UPDATING": "true", "VLLM_DISABLE_COMPILE_CACHE": "1", "VLLM_LOGGING_LEVEL": "WARN"}, "py_executable": "nsys profile --cuda-graph-trace=graph --cuda-memory-usage=true --trace=cuda,nvtx,cublas,ucx -o /tmp/ray/session_2026-02-27_00-30-26_175126_10593/logs/nsight/'worker_process_%p' python", "override_worker_entrypoint": null, "java_jars": []}
12
+ 2026-02-27 00:31:29,875 INFO runtime_env_agent.py:467 -- Runtime env already created successfully. Env: {"_nsight":{"cuda-graph-trace":"graph","cuda-memory-usage":"true","trace":"cuda,nvtx,cublas,ucx"},"env_vars":{"CUDA_DEVICE_MAX_CONNECTIONS":"1","HCCL_HOST_SOCKET_PORT_RANGE":"auto","HCCL_NPU_SOCKET_PORT_RANGE":"auto","NCCL_CUMEM_ENABLE":"0","NCCL_DEBUG":"WARN","TOKENIZERS_PARALLELISM":"true","VLLM_ALLOW_RUNTIME_LORA_UPDATING":"true","VLLM_DISABLE_COMPILE_CACHE":"1","VLLM_LOGGING_LEVEL":"WARN"}}, context: {"command_prefix": [], "env_vars": {"CUDA_DEVICE_MAX_CONNECTIONS": "1", "HCCL_HOST_SOCKET_PORT_RANGE": "auto", "HCCL_NPU_SOCKET_PORT_RANGE": "auto", "NCCL_CUMEM_ENABLE": "0", "NCCL_DEBUG": "WARN", "TOKENIZERS_PARALLELISM": "true", "VLLM_ALLOW_RUNTIME_LORA_UPDATING": "true", "VLLM_DISABLE_COMPILE_CACHE": "1", "VLLM_LOGGING_LEVEL": "WARN"}, "py_executable": "nsys profile --cuda-graph-trace=graph --cuda-memory-usage=true --trace=cuda,nvtx,cublas,ucx -o /tmp/ray/session_2026-02-27_00-30-26_175126_10593/logs/nsight/'worker_process_%p' python", "override_worker_entrypoint": null, "java_jars": []}
13
+ 2026-02-27 00:31:29,888 INFO runtime_env_agent.py:467 -- Runtime env already created successfully. Env: {"_nsight":{"cuda-graph-trace":"graph","cuda-memory-usage":"true","trace":"cuda,nvtx,cublas,ucx"},"env_vars":{"CUDA_DEVICE_MAX_CONNECTIONS":"1","HCCL_HOST_SOCKET_PORT_RANGE":"auto","HCCL_NPU_SOCKET_PORT_RANGE":"auto","NCCL_CUMEM_ENABLE":"0","NCCL_DEBUG":"WARN","TOKENIZERS_PARALLELISM":"true","VLLM_ALLOW_RUNTIME_LORA_UPDATING":"true","VLLM_DISABLE_COMPILE_CACHE":"1","VLLM_LOGGING_LEVEL":"WARN"}}, context: {"command_prefix": [], "env_vars": {"CUDA_DEVICE_MAX_CONNECTIONS": "1", "HCCL_HOST_SOCKET_PORT_RANGE": "auto", "HCCL_NPU_SOCKET_PORT_RANGE": "auto", "NCCL_CUMEM_ENABLE": "0", "NCCL_DEBUG": "WARN", "TOKENIZERS_PARALLELISM": "true", "VLLM_ALLOW_RUNTIME_LORA_UPDATING": "true", "VLLM_DISABLE_COMPILE_CACHE": "1", "VLLM_LOGGING_LEVEL": "WARN"}, "py_executable": "nsys profile --cuda-graph-trace=graph --cuda-memory-usage=true --trace=cuda,nvtx,cublas,ucx -o /tmp/ray/session_2026-02-27_00-30-26_175126_10593/logs/nsight/'worker_process_%p' python", "override_worker_entrypoint": null, "java_jars": []}
14
+ 2026-02-27 00:31:29,903 INFO runtime_env_agent.py:467 -- Runtime env already created successfully. Env: {"_nsight":{"cuda-graph-trace":"graph","cuda-memory-usage":"true","trace":"cuda,nvtx,cublas,ucx"},"env_vars":{"CUDA_DEVICE_MAX_CONNECTIONS":"1","HCCL_HOST_SOCKET_PORT_RANGE":"auto","HCCL_NPU_SOCKET_PORT_RANGE":"auto","NCCL_CUMEM_ENABLE":"0","NCCL_DEBUG":"WARN","TOKENIZERS_PARALLELISM":"true","VLLM_ALLOW_RUNTIME_LORA_UPDATING":"true","VLLM_DISABLE_COMPILE_CACHE":"1","VLLM_LOGGING_LEVEL":"WARN"}}, context: {"command_prefix": [], "env_vars": {"CUDA_DEVICE_MAX_CONNECTIONS": "1", "HCCL_HOST_SOCKET_PORT_RANGE": "auto", "HCCL_NPU_SOCKET_PORT_RANGE": "auto", "NCCL_CUMEM_ENABLE": "0", "NCCL_DEBUG": "WARN", "TOKENIZERS_PARALLELISM": "true", "VLLM_ALLOW_RUNTIME_LORA_UPDATING": "true", "VLLM_DISABLE_COMPILE_CACHE": "1", "VLLM_LOGGING_LEVEL": "WARN"}, "py_executable": "nsys profile --cuda-graph-trace=graph --cuda-memory-usage=true --trace=cuda,nvtx,cublas,ucx -o /tmp/ray/session_2026-02-27_00-30-26_175126_10593/logs/nsight/'worker_process_%p' python", "override_worker_entrypoint": null, "java_jars": []}
15
+ 2026-02-27 00:31:29,919 INFO runtime_env_agent.py:467 -- Runtime env already created successfully. Env: {"_nsight":{"cuda-graph-trace":"graph","cuda-memory-usage":"true","trace":"cuda,nvtx,cublas,ucx"},"env_vars":{"CUDA_DEVICE_MAX_CONNECTIONS":"1","HCCL_HOST_SOCKET_PORT_RANGE":"auto","HCCL_NPU_SOCKET_PORT_RANGE":"auto","NCCL_CUMEM_ENABLE":"0","NCCL_DEBUG":"WARN","TOKENIZERS_PARALLELISM":"true","VLLM_ALLOW_RUNTIME_LORA_UPDATING":"true","VLLM_DISABLE_COMPILE_CACHE":"1","VLLM_LOGGING_LEVEL":"WARN"}}, context: {"command_prefix": [], "env_vars": {"CUDA_DEVICE_MAX_CONNECTIONS": "1", "HCCL_HOST_SOCKET_PORT_RANGE": "auto", "HCCL_NPU_SOCKET_PORT_RANGE": "auto", "NCCL_CUMEM_ENABLE": "0", "NCCL_DEBUG": "WARN", "TOKENIZERS_PARALLELISM": "true", "VLLM_ALLOW_RUNTIME_LORA_UPDATING": "true", "VLLM_DISABLE_COMPILE_CACHE": "1", "VLLM_LOGGING_LEVEL": "WARN"}, "py_executable": "nsys profile --cuda-graph-trace=graph --cuda-memory-usage=true --trace=cuda,nvtx,cublas,ucx -o /tmp/ray/session_2026-02-27_00-30-26_175126_10593/logs/nsight/'worker_process_%p' python", "override_worker_entrypoint": null, "java_jars": []}
16
+ 2026-02-27 00:31:29,938 INFO runtime_env_agent.py:467 -- Runtime env already created successfully. Env: {"_nsight":{"cuda-graph-trace":"graph","cuda-memory-usage":"true","trace":"cuda,nvtx,cublas,ucx"},"env_vars":{"CUDA_DEVICE_MAX_CONNECTIONS":"1","HCCL_HOST_SOCKET_PORT_RANGE":"auto","HCCL_NPU_SOCKET_PORT_RANGE":"auto","NCCL_CUMEM_ENABLE":"0","NCCL_DEBUG":"WARN","TOKENIZERS_PARALLELISM":"true","VLLM_ALLOW_RUNTIME_LORA_UPDATING":"true","VLLM_DISABLE_COMPILE_CACHE":"1","VLLM_LOGGING_LEVEL":"WARN"}}, context: {"command_prefix": [], "env_vars": {"CUDA_DEVICE_MAX_CONNECTIONS": "1", "HCCL_HOST_SOCKET_PORT_RANGE": "auto", "HCCL_NPU_SOCKET_PORT_RANGE": "auto", "NCCL_CUMEM_ENABLE": "0", "NCCL_DEBUG": "WARN", "TOKENIZERS_PARALLELISM": "true", "VLLM_ALLOW_RUNTIME_LORA_UPDATING": "true", "VLLM_DISABLE_COMPILE_CACHE": "1", "VLLM_LOGGING_LEVEL": "WARN"}, "py_executable": "nsys profile --cuda-graph-trace=graph --cuda-memory-usage=true --trace=cuda,nvtx,cublas,ucx -o /tmp/ray/session_2026-02-27_00-30-26_175126_10593/logs/nsight/'worker_process_%p' python", "override_worker_entrypoint": null, "java_jars": []}
17
+ 2026-02-27 00:31:29,961 INFO runtime_env_agent.py:467 -- Runtime env already created successfully. Env: {"_nsight":{"cuda-graph-trace":"graph","cuda-memory-usage":"true","trace":"cuda,nvtx,cublas,ucx"},"env_vars":{"CUDA_DEVICE_MAX_CONNECTIONS":"1","HCCL_HOST_SOCKET_PORT_RANGE":"auto","HCCL_NPU_SOCKET_PORT_RANGE":"auto","NCCL_CUMEM_ENABLE":"0","NCCL_DEBUG":"WARN","TOKENIZERS_PARALLELISM":"true","VLLM_ALLOW_RUNTIME_LORA_UPDATING":"true","VLLM_DISABLE_COMPILE_CACHE":"1","VLLM_LOGGING_LEVEL":"WARN"}}, context: {"command_prefix": [], "env_vars": {"CUDA_DEVICE_MAX_CONNECTIONS": "1", "HCCL_HOST_SOCKET_PORT_RANGE": "auto", "HCCL_NPU_SOCKET_PORT_RANGE": "auto", "NCCL_CUMEM_ENABLE": "0", "NCCL_DEBUG": "WARN", "TOKENIZERS_PARALLELISM": "true", "VLLM_ALLOW_RUNTIME_LORA_UPDATING": "true", "VLLM_DISABLE_COMPILE_CACHE": "1", "VLLM_LOGGING_LEVEL": "WARN"}, "py_executable": "nsys profile --cuda-graph-trace=graph --cuda-memory-usage=true --trace=cuda,nvtx,cublas,ucx -o /tmp/ray/session_2026-02-27_00-30-26_175126_10593/logs/nsight/'worker_process_%p' python", "override_worker_entrypoint": null, "java_jars": []}
18
+ 2026-02-27 00:31:29,985 INFO runtime_env_agent.py:467 -- Runtime env already created successfully. Env: {"_nsight":{"cuda-graph-trace":"graph","cuda-memory-usage":"true","trace":"cuda,nvtx,cublas,ucx"},"env_vars":{"CUDA_DEVICE_MAX_CONNECTIONS":"1","HCCL_HOST_SOCKET_PORT_RANGE":"auto","HCCL_NPU_SOCKET_PORT_RANGE":"auto","NCCL_CUMEM_ENABLE":"0","NCCL_DEBUG":"WARN","TOKENIZERS_PARALLELISM":"true","VLLM_ALLOW_RUNTIME_LORA_UPDATING":"true","VLLM_DISABLE_COMPILE_CACHE":"1","VLLM_LOGGING_LEVEL":"WARN"}}, context: {"command_prefix": [], "env_vars": {"CUDA_DEVICE_MAX_CONNECTIONS": "1", "HCCL_HOST_SOCKET_PORT_RANGE": "auto", "HCCL_NPU_SOCKET_PORT_RANGE": "auto", "NCCL_CUMEM_ENABLE": "0", "NCCL_DEBUG": "WARN", "TOKENIZERS_PARALLELISM": "true", "VLLM_ALLOW_RUNTIME_LORA_UPDATING": "true", "VLLM_DISABLE_COMPILE_CACHE": "1", "VLLM_LOGGING_LEVEL": "WARN"}, "py_executable": "nsys profile --cuda-graph-trace=graph --cuda-memory-usage=true --trace=cuda,nvtx,cublas,ucx -o /tmp/ray/session_2026-02-27_00-30-26_175126_10593/logs/nsight/'worker_process_%p' python", "override_worker_entrypoint": null, "java_jars": []}
19
+ 2026-02-27 00:31:38,164 INFO runtime_env_agent.py:378 -- Creating runtime env: {"_nsight":{"cuda-graph-trace":"graph","cuda-memory-usage":"true","trace":"cuda,nvtx,cublas,ucx"},"env_vars":{"CUDA_DEVICE_MAX_CONNECTIONS":"1","HCCL_HOST_SOCKET_PORT_RANGE":"auto","HCCL_NPU_SOCKET_PORT_RANGE":"auto","NCCL_CUMEM_ENABLE":"0","NCCL_DEBUG":"WARN","RAY_EXPERIMENTAL_NOSET_CUDA_VISIBLE_DEVICES":"1","TOKENIZERS_PARALLELISM":"true","VLLM_ALLOW_RUNTIME_LORA_UPDATING":"true","VLLM_DISABLE_COMPILE_CACHE":"1","VLLM_LOGGING_LEVEL":"WARN"}} with timeout 600 seconds.
20
+ 2026-02-27 00:31:41,790 INFO runtime_env_agent.py:428 -- Successfully created runtime env: {"_nsight":{"cuda-graph-trace":"graph","cuda-memory-usage":"true","trace":"cuda,nvtx,cublas,ucx"},"env_vars":{"CUDA_DEVICE_MAX_CONNECTIONS":"1","HCCL_HOST_SOCKET_PORT_RANGE":"auto","HCCL_NPU_SOCKET_PORT_RANGE":"auto","NCCL_CUMEM_ENABLE":"0","NCCL_DEBUG":"WARN","RAY_EXPERIMENTAL_NOSET_CUDA_VISIBLE_DEVICES":"1","TOKENIZERS_PARALLELISM":"true","VLLM_ALLOW_RUNTIME_LORA_UPDATING":"true","VLLM_DISABLE_COMPILE_CACHE":"1","VLLM_LOGGING_LEVEL":"WARN"}}, context: {"command_prefix": [], "env_vars": {"CUDA_DEVICE_MAX_CONNECTIONS": "1", "HCCL_HOST_SOCKET_PORT_RANGE": "auto", "HCCL_NPU_SOCKET_PORT_RANGE": "auto", "NCCL_CUMEM_ENABLE": "0", "NCCL_DEBUG": "WARN", "RAY_EXPERIMENTAL_NOSET_CUDA_VISIBLE_DEVICES": "1", "TOKENIZERS_PARALLELISM": "true", "VLLM_ALLOW_RUNTIME_LORA_UPDATING": "true", "VLLM_DISABLE_COMPILE_CACHE": "1", "VLLM_LOGGING_LEVEL": "WARN"}, "py_executable": "nsys profile --cuda-graph-trace=graph --cuda-memory-usage=true --trace=cuda,nvtx,cublas,ucx -o /tmp/ray/session_2026-02-27_00-30-26_175126_10593/logs/nsight/'worker_process_%p' python", "override_worker_entrypoint": null, "java_jars": []}
21
+ 2026-02-27 00:32:12,454 INFO runtime_env_agent.py:540 -- Got request from raylet to decrease reference for runtime env: {"env_vars": {"CUDA_DEVICE_MAX_CONNECTIONS": "1", "HCCL_HOST_SOCKET_PORT_RANGE": "auto", "HCCL_NPU_SOCKET_PORT_RANGE": "auto", "NCCL_CUMEM_ENABLE": "0", "NCCL_DEBUG": "WARN", "TOKENIZERS_PARALLELISM": "true", "VLLM_ALLOW_RUNTIME_LORA_UPDATING": "true", "VLLM_DISABLE_COMPILE_CACHE": "1", "VLLM_LOGGING_LEVEL": "WARN"}}.
22
+ 2026-02-27 00:32:12,454 INFO runtime_env_agent.py:133 -- Unused runtime env {"env_vars": {"CUDA_DEVICE_MAX_CONNECTIONS": "1", "HCCL_HOST_SOCKET_PORT_RANGE": "auto", "HCCL_NPU_SOCKET_PORT_RANGE": "auto", "NCCL_CUMEM_ENABLE": "0", "NCCL_DEBUG": "WARN", "TOKENIZERS_PARALLELISM": "true", "VLLM_ALLOW_RUNTIME_LORA_UPDATING": "true", "VLLM_DISABLE_COMPILE_CACHE": "1", "VLLM_LOGGING_LEVEL": "WARN"}}.
23
+ 2026-02-27 00:32:12,455 INFO runtime_env_agent.py:277 -- Runtime env {"env_vars": {"CUDA_DEVICE_MAX_CONNECTIONS": "1", "HCCL_HOST_SOCKET_PORT_RANGE": "auto", "HCCL_NPU_SOCKET_PORT_RANGE": "auto", "NCCL_CUMEM_ENABLE": "0", "NCCL_DEBUG": "WARN", "TOKENIZERS_PARALLELISM": "true", "VLLM_ALLOW_RUNTIME_LORA_UPDATING": "true", "VLLM_DISABLE_COMPILE_CACHE": "1", "VLLM_LOGGING_LEVEL": "WARN"}} removed from env-level cache.
24
+ 2026-02-27 00:32:12,456 INFO runtime_env_agent.py:540 -- Got request from raylet to decrease reference for runtime env: {"_nsight":{"capture-range":"cudaProfilerApi","capture-range-end":"repeat-shutdown:24","cuda-graph-trace":"graph","cuda-memory-usage":"true","kill":"none","trace":"cuda,nvtx,osrt"},"env_vars":{"CUDA_DEVICE_MAX_CONNECTIONS":"1","HCCL_HOST_SOCKET_PORT_RANGE":"auto","HCCL_NPU_SOCKET_PORT_RANGE":"auto","MASTER_ADDR":"10.128.0.163","MASTER_PORT":"50347","NCCL_CUMEM_ENABLE":"0","NCCL_DEBUG":"WARN","RANK":"0","RAY_LOCAL_WORLD_SIZE":"1","TOKENIZERS_PARALLELISM":"true","VLLM_ALLOW_RUNTIME_LORA_UPDATING":"true","VLLM_DISABLE_COMPILE_CACHE":"1","VLLM_LOGGING_LEVEL":"WARN","WG_BACKEND":"ray","WG_PREFIX":"B8Qm6M","WORLD_SIZE":"1"}}.
25
+ 2026-02-27 00:32:12,457 INFO runtime_env_agent.py:133 -- Unused runtime env {"_nsight":{"capture-range":"cudaProfilerApi","capture-range-end":"repeat-shutdown:24","cuda-graph-trace":"graph","cuda-memory-usage":"true","kill":"none","trace":"cuda,nvtx,osrt"},"env_vars":{"CUDA_DEVICE_MAX_CONNECTIONS":"1","HCCL_HOST_SOCKET_PORT_RANGE":"auto","HCCL_NPU_SOCKET_PORT_RANGE":"auto","MASTER_ADDR":"10.128.0.163","MASTER_PORT":"50347","NCCL_CUMEM_ENABLE":"0","NCCL_DEBUG":"WARN","RANK":"0","RAY_LOCAL_WORLD_SIZE":"1","TOKENIZERS_PARALLELISM":"true","VLLM_ALLOW_RUNTIME_LORA_UPDATING":"true","VLLM_DISABLE_COMPILE_CACHE":"1","VLLM_LOGGING_LEVEL":"WARN","WG_BACKEND":"ray","WG_PREFIX":"B8Qm6M","WORLD_SIZE":"1"}}.
26
+ 2026-02-27 00:32:12,457 INFO runtime_env_agent.py:277 -- Runtime env {"_nsight":{"capture-range":"cudaProfilerApi","capture-range-end":"repeat-shutdown:24","cuda-graph-trace":"graph","cuda-memory-usage":"true","kill":"none","trace":"cuda,nvtx,osrt"},"env_vars":{"CUDA_DEVICE_MAX_CONNECTIONS":"1","HCCL_HOST_SOCKET_PORT_RANGE":"auto","HCCL_NPU_SOCKET_PORT_RANGE":"auto","MASTER_ADDR":"10.128.0.163","MASTER_PORT":"50347","NCCL_CUMEM_ENABLE":"0","NCCL_DEBUG":"WARN","RANK":"0","RAY_LOCAL_WORLD_SIZE":"1","TOKENIZERS_PARALLELISM":"true","VLLM_ALLOW_RUNTIME_LORA_UPDATING":"true","VLLM_DISABLE_COMPILE_CACHE":"1","VLLM_LOGGING_LEVEL":"WARN","WG_BACKEND":"ray","WG_PREFIX":"B8Qm6M","WORLD_SIZE":"1"}} removed from env-level cache.
27
+ 2026-02-27 00:32:12,470 INFO runtime_env_agent.py:540 -- Got request from raylet to decrease reference for runtime env: {"_nsight":{"cuda-graph-trace":"graph","cuda-memory-usage":"true","trace":"cuda,nvtx,cublas,ucx"},"env_vars":{"CUDA_DEVICE_MAX_CONNECTIONS":"1","HCCL_HOST_SOCKET_PORT_RANGE":"auto","HCCL_NPU_SOCKET_PORT_RANGE":"auto","NCCL_CUMEM_ENABLE":"0","NCCL_DEBUG":"WARN","TOKENIZERS_PARALLELISM":"true","VLLM_ALLOW_RUNTIME_LORA_UPDATING":"true","VLLM_DISABLE_COMPILE_CACHE":"1","VLLM_LOGGING_LEVEL":"WARN"}}.
28
+ 2026-02-27 00:32:12,476 INFO runtime_env_agent.py:540 -- Got request from raylet to decrease reference for runtime env: {"_nsight":{"cuda-graph-trace":"graph","cuda-memory-usage":"true","trace":"cuda,nvtx,cublas,ucx"},"env_vars":{"CUDA_DEVICE_MAX_CONNECTIONS":"1","HCCL_HOST_SOCKET_PORT_RANGE":"auto","HCCL_NPU_SOCKET_PORT_RANGE":"auto","NCCL_CUMEM_ENABLE":"0","NCCL_DEBUG":"WARN","TOKENIZERS_PARALLELISM":"true","VLLM_ALLOW_RUNTIME_LORA_UPDATING":"true","VLLM_DISABLE_COMPILE_CACHE":"1","VLLM_LOGGING_LEVEL":"WARN"}}.
29
+ 2026-02-27 00:32:12,478 INFO runtime_env_agent.py:540 -- Got request from raylet to decrease reference for runtime env: {"_nsight":{"cuda-graph-trace":"graph","cuda-memory-usage":"true","trace":"cuda,nvtx,cublas,ucx"},"env_vars":{"CUDA_DEVICE_MAX_CONNECTIONS":"1","HCCL_HOST_SOCKET_PORT_RANGE":"auto","HCCL_NPU_SOCKET_PORT_RANGE":"auto","NCCL_CUMEM_ENABLE":"0","NCCL_DEBUG":"WARN","TOKENIZERS_PARALLELISM":"true","VLLM_ALLOW_RUNTIME_LORA_UPDATING":"true","VLLM_DISABLE_COMPILE_CACHE":"1","VLLM_LOGGING_LEVEL":"WARN"}}.
30
+ 2026-02-27 00:32:12,486 INFO runtime_env_agent.py:540 -- Got request from raylet to decrease reference for runtime env: {"_nsight":{"cuda-graph-trace":"graph","cuda-memory-usage":"true","trace":"cuda,nvtx,cublas,ucx"},"env_vars":{"CUDA_DEVICE_MAX_CONNECTIONS":"1","HCCL_HOST_SOCKET_PORT_RANGE":"auto","HCCL_NPU_SOCKET_PORT_RANGE":"auto","NCCL_CUMEM_ENABLE":"0","NCCL_DEBUG":"WARN","TOKENIZERS_PARALLELISM":"true","VLLM_ALLOW_RUNTIME_LORA_UPDATING":"true","VLLM_DISABLE_COMPILE_CACHE":"1","VLLM_LOGGING_LEVEL":"WARN"}}.
31
+ 2026-02-27 00:32:12,509 INFO runtime_env_agent.py:540 -- Got request from raylet to decrease reference for runtime env: {"_nsight":{"cuda-graph-trace":"graph","cuda-memory-usage":"true","trace":"cuda,nvtx,cublas,ucx"},"env_vars":{"CUDA_DEVICE_MAX_CONNECTIONS":"1","HCCL_HOST_SOCKET_PORT_RANGE":"auto","HCCL_NPU_SOCKET_PORT_RANGE":"auto","NCCL_CUMEM_ENABLE":"0","NCCL_DEBUG":"WARN","TOKENIZERS_PARALLELISM":"true","VLLM_ALLOW_RUNTIME_LORA_UPDATING":"true","VLLM_DISABLE_COMPILE_CACHE":"1","VLLM_LOGGING_LEVEL":"WARN"}}.
32
+ 2026-02-27 00:32:12,518 INFO runtime_env_agent.py:540 -- Got request from raylet to decrease reference for runtime env: {"_nsight":{"cuda-graph-trace":"graph","cuda-memory-usage":"true","trace":"cuda,nvtx,cublas,ucx"},"env_vars":{"CUDA_DEVICE_MAX_CONNECTIONS":"1","HCCL_HOST_SOCKET_PORT_RANGE":"auto","HCCL_NPU_SOCKET_PORT_RANGE":"auto","NCCL_CUMEM_ENABLE":"0","NCCL_DEBUG":"WARN","TOKENIZERS_PARALLELISM":"true","VLLM_ALLOW_RUNTIME_LORA_UPDATING":"true","VLLM_DISABLE_COMPILE_CACHE":"1","VLLM_LOGGING_LEVEL":"WARN"}}.
33
+ 2026-02-27 00:32:12,519 INFO runtime_env_agent.py:540 -- Got request from raylet to decrease reference for runtime env: {"_nsight":{"cuda-graph-trace":"graph","cuda-memory-usage":"true","trace":"cuda,nvtx,cublas,ucx"},"env_vars":{"CUDA_DEVICE_MAX_CONNECTIONS":"1","HCCL_HOST_SOCKET_PORT_RANGE":"auto","HCCL_NPU_SOCKET_PORT_RANGE":"auto","NCCL_CUMEM_ENABLE":"0","NCCL_DEBUG":"WARN","TOKENIZERS_PARALLELISM":"true","VLLM_ALLOW_RUNTIME_LORA_UPDATING":"true","VLLM_DISABLE_COMPILE_CACHE":"1","VLLM_LOGGING_LEVEL":"WARN"}}.
34
+ 2026-02-27 00:32:12,523 INFO runtime_env_agent.py:540 -- Got request from raylet to decrease reference for runtime env: {"_nsight":{"cuda-graph-trace":"graph","cuda-memory-usage":"true","trace":"cuda,nvtx,cublas,ucx"},"env_vars":{"CUDA_DEVICE_MAX_CONNECTIONS":"1","HCCL_HOST_SOCKET_PORT_RANGE":"auto","HCCL_NPU_SOCKET_PORT_RANGE":"auto","NCCL_CUMEM_ENABLE":"0","NCCL_DEBUG":"WARN","RAY_EXPERIMENTAL_NOSET_CUDA_VISIBLE_DEVICES":"1","TOKENIZERS_PARALLELISM":"true","VLLM_ALLOW_RUNTIME_LORA_UPDATING":"true","VLLM_DISABLE_COMPILE_CACHE":"1","VLLM_LOGGING_LEVEL":"WARN"}}.
35
+ 2026-02-27 00:32:12,524 INFO runtime_env_agent.py:133 -- Unused runtime env {"_nsight":{"cuda-graph-trace":"graph","cuda-memory-usage":"true","trace":"cuda,nvtx,cublas,ucx"},"env_vars":{"CUDA_DEVICE_MAX_CONNECTIONS":"1","HCCL_HOST_SOCKET_PORT_RANGE":"auto","HCCL_NPU_SOCKET_PORT_RANGE":"auto","NCCL_CUMEM_ENABLE":"0","NCCL_DEBUG":"WARN","RAY_EXPERIMENTAL_NOSET_CUDA_VISIBLE_DEVICES":"1","TOKENIZERS_PARALLELISM":"true","VLLM_ALLOW_RUNTIME_LORA_UPDATING":"true","VLLM_DISABLE_COMPILE_CACHE":"1","VLLM_LOGGING_LEVEL":"WARN"}}.
36
+ 2026-02-27 00:32:12,524 INFO runtime_env_agent.py:277 -- Runtime env {"_nsight":{"cuda-graph-trace":"graph","cuda-memory-usage":"true","trace":"cuda,nvtx,cublas,ucx"},"env_vars":{"CUDA_DEVICE_MAX_CONNECTIONS":"1","HCCL_HOST_SOCKET_PORT_RANGE":"auto","HCCL_NPU_SOCKET_PORT_RANGE":"auto","NCCL_CUMEM_ENABLE":"0","NCCL_DEBUG":"WARN","RAY_EXPERIMENTAL_NOSET_CUDA_VISIBLE_DEVICES":"1","TOKENIZERS_PARALLELISM":"true","VLLM_ALLOW_RUNTIME_LORA_UPDATING":"true","VLLM_DISABLE_COMPILE_CACHE":"1","VLLM_LOGGING_LEVEL":"WARN"}} removed from env-level cache.
37
+ 2026-02-27 00:32:12,524 INFO runtime_env_agent.py:540 -- Got request from raylet to decrease reference for runtime env: {"_nsight":{"cuda-graph-trace":"graph","cuda-memory-usage":"true","trace":"cuda,nvtx,cublas,ucx"},"env_vars":{"CUDA_DEVICE_MAX_CONNECTIONS":"1","HCCL_HOST_SOCKET_PORT_RANGE":"auto","HCCL_NPU_SOCKET_PORT_RANGE":"auto","NCCL_CUMEM_ENABLE":"0","NCCL_DEBUG":"WARN","TOKENIZERS_PARALLELISM":"true","VLLM_ALLOW_RUNTIME_LORA_UPDATING":"true","VLLM_DISABLE_COMPILE_CACHE":"1","VLLM_LOGGING_LEVEL":"WARN"}}.
38
+ 2026-02-27 00:32:12,525 INFO runtime_env_agent.py:540 -- Got request from raylet to decrease reference for runtime env: {"_nsight":{"cuda-graph-trace":"graph","cuda-memory-usage":"true","trace":"cuda,nvtx,cublas,ucx"},"env_vars":{"CUDA_DEVICE_MAX_CONNECTIONS":"1","HCCL_HOST_SOCKET_PORT_RANGE":"auto","HCCL_NPU_SOCKET_PORT_RANGE":"auto","NCCL_CUMEM_ENABLE":"0","NCCL_DEBUG":"WARN","TOKENIZERS_PARALLELISM":"true","VLLM_ALLOW_RUNTIME_LORA_UPDATING":"true","VLLM_DISABLE_COMPILE_CACHE":"1","VLLM_LOGGING_LEVEL":"WARN"}}.
runtime_env_agent.out ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ ======== Running on http://10.128.0.163:57999 ========
2
+ (Press CTRL+C to quit)
worker-100e3eeb4a57ce034285a311628f834885904c1e1ea9caa911a3c4da-ffffffff-11375.err ADDED
File without changes
worker-100e3eeb4a57ce034285a311628f834885904c1e1ea9caa911a3c4da-ffffffff-11375.out ADDED
File without changes
worker-1535028fd440028216a02042c55e0b58baec34df171b54a8306f4bc8-ffffffff-11376.out ADDED
File without changes
worker-15c410d5d6a75625cb50c80927d18090e899b8edc49402fe08e50ee6-01000000-12567.err ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ :job_id:01000000
2
+ :actor_name:RewardLoopWorker
worker-2cda7ffb1fdfeaaf98e6be62760ae2627c565d43cb10409f83c0a748-ffffffff-11372.err ADDED
File without changes
worker-33b9d0a21a51ca22dda2aa2142cb264d1ee4f9d53a55dc567b49496c-01000000-12500.err ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ :job_id:01000000
2
+ :actor_name:RewardLoopWorker
worker-658f00c930b44d143152233262d8e94af875b52448898614a4b579ba-ffffffff-11377.err ADDED
File without changes
worker-772052e39bd349442253d65d82fc94825e9e58c75098b1b473bedce2-01000000-12586.out ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ :job_id:01000000
2
+ :actor_name:RewardLoopWorker
worker-8d27d27b6a5c820150d6a54cd27fe296fd0409567d2b4685b9a84fc8-01000000-11896.err ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ :job_id:01000000
2
+ W0227 00:30:48.007000 11896 torch/utils/cpp_extension.py:117] No CUDA runtime is found, using CUDA_HOME='/usr/local/cuda'
3
+ WARNING:2026-02-27 00:30:48,712:fused_indices_to_multihot has reached end of life. Please migrate to a non-experimental function.
4
+ :actor_name:TaskRunner
5
+ /workspace/verl/verl/workers/config/critic.py:204: UserWarning: using model in Critic Config is deprecated, please use model_config instead
6
+ super().__post_init__()
7
+ /workspace/verl/verl/utils/tokenizer.py:109: UserWarning: Failed to create processor: Unsupported processor type: Qwen2TokenizerFast. This may affect multimodal processing
8
+ warnings.warn(f"Failed to create processor: {e}. This may affect multimodal processing", stacklevel=1)
worker-9224fcd6abcfd04deeca6990e3ac522c58f6eec637ba09c0e927aaef-01000000-12481.out ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ :job_id:01000000
2
+ :actor_name:RewardLoopWorker
worker-af6e4d2eae80c226c783dd6717832e015ec8fc0144d801649c12abfe-01000000-12563.err ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ :job_id:01000000
2
+ :actor_name:RewardLoopWorker
worker-b809b75ac50a13f3d02e083041fe7ba32c1445fa33db5801d3c6cfe5-01000000-12477.err ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ :job_id:01000000
2
+ :actor_name:RewardLoopWorker
worker-bb0f50c5405699ae07f957ec3f7c03f2bdf40be03f6e39b39232dc16-ffffffff-11378.err ADDED
File without changes
worker-f46103e29121f0b748164b47d1653310da7f304c2c8c8df73871f0e5-01000000-12507.err ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ :job_id:01000000
2
+ :actor_name:RewardLoopWorker
worker-f46103e29121f0b748164b47d1653310da7f304c2c8c8df73871f0e5-01000000-12507.out ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ :job_id:01000000
2
+ :actor_name:RewardLoopWorker
worker-fc14e0d4e4b6acb4ecead813c2d960587eefa7859aac6d8e19aeec98-ffffffff-11374.out ADDED
File without changes