Package evaluation of Schedulers on Julia 1.13.0-DEV.794 (d7c70bcbab*) started at 2025-07-02T07:02:47.752 ################################################################################ # Set-up # Installing PkgEval dependencies (TestEnv)... Set-up completed after 8.24s ################################################################################ # Installation # Installing Schedulers... Resolving package versions... Updating `~/.julia/environments/v1.13/Project.toml` [fcb5363a] + Schedulers v0.11.3 Updating `~/.julia/environments/v1.13/Manifest.toml` [682c06a0] + JSON v0.21.4 [69de0a69] + Parsers v2.8.3 [aea7be01] + PrecompileTools v1.3.2 [21216c6a] + Preferences v1.4.3 [fcb5363a] + Schedulers v0.11.3 [10745b16] + Statistics v1.11.1 [56f22d72] + Artifacts v1.11.0 [ade2ca70] + Dates v1.11.0 [8ba89e20] + Distributed v1.11.0 [8f399da3] + Libdl v1.11.0 [37e2e46d] + LinearAlgebra v1.12.0 [56ddb016] + Logging v1.11.0 [a63ad114] + Mmap v1.11.0 [de0858da] + Printf v1.11.0 [9a3f8284] + Random v1.11.0 [ea8e919c] + SHA v0.7.0 [9e88b42a] + Serialization v1.11.0 [6462fe0b] + Sockets v1.11.0 [fa267f1f] + TOML v1.0.3 [cf7118a7] + UUIDs v1.11.0 [4ec0a83e] + Unicode v1.11.0 [e66e0078] + CompilerSupportLibraries_jll v1.3.0+1 [4536629a] + OpenBLAS_jll v0.3.29+0 [8e850b90] + libblastrampoline_jll v5.13.1+0 Installation completed after 1.06s ################################################################################ # Precompilation # Precompiling PkgEval dependencies... Precompiling package dependencies... Precompilation completed after 15.91s ################################################################################ # Testing # Testing Schedulers Status `/tmp/jl_mXaXSE/Project.toml` [fcb5363a] Schedulers v0.11.3 [8ba89e20] Distributed v1.11.0 [56ddb016] Logging v1.11.0 [9a3f8284] Random v1.11.0 [9e88b42a] Serialization v1.11.0 [8dfed614] Test v1.11.0 Status `/tmp/jl_mXaXSE/Manifest.toml` [682c06a0] JSON v0.21.4 [69de0a69] Parsers v2.8.3 [aea7be01] PrecompileTools v1.3.2 [21216c6a] Preferences v1.4.3 [fcb5363a] Schedulers v0.11.3 [10745b16] Statistics v1.11.1 [56f22d72] Artifacts v1.11.0 [2a0f44e3] Base64 v1.11.0 [ade2ca70] Dates v1.11.0 [8ba89e20] Distributed v1.11.0 [b77e0a4c] InteractiveUtils v1.11.0 [ac6e5ff7] JuliaSyntaxHighlighting v1.12.0 [8f399da3] Libdl v1.11.0 [37e2e46d] LinearAlgebra v1.12.0 [56ddb016] Logging v1.11.0 [d6f4376e] Markdown v1.11.0 [a63ad114] Mmap v1.11.0 [de0858da] Printf v1.11.0 [9a3f8284] Random v1.11.0 [ea8e919c] SHA v0.7.0 [9e88b42a] Serialization v1.11.0 [6462fe0b] Sockets v1.11.0 [f489334b] StyledStrings v1.11.0 [fa267f1f] TOML v1.0.3 [8dfed614] Test v1.11.0 [cf7118a7] UUIDs v1.11.0 [4ec0a83e] Unicode v1.11.0 [e66e0078] CompilerSupportLibraries_jll v1.3.0+1 [4536629a] OpenBLAS_jll v0.3.29+0 [8e850b90] libblastrampoline_jll v5.13.1+0 Testing Running tests... [ Info: running task 1 on process 2 (Schedulers-primary-eA3Kzwpz); 5 workers total; 9 tasks left in task-pool. [ Info: running task 2 on process 3 (Schedulers-primary-eA3Kzwpz); 5 workers total; 5 tasks left in task-pool. [ Info: running task 3 on process 4 (Schedulers-primary-eA3Kzwpz); 5 workers total; 5 tasks left in task-pool. [ Info: running task 4 on process 5 (Schedulers-primary-eA3Kzwpz); 5 workers total; 5 tasks left in task-pool. [ Info: running task 5 on process 6 (Schedulers-primary-eA3Kzwpz); 5 workers total; 5 tasks left in task-pool. [ Info: running task 6 on process 2 (Schedulers-primary-eA3Kzwpz); 5 workers total; 4 tasks left in task-pool. [ Info: running task 7 on process 3 (Schedulers-primary-eA3Kzwpz); 5 workers total; 3 tasks left in task-pool. [ Info: running task 8 on process 4 (Schedulers-primary-eA3Kzwpz); 5 workers total; 0 tasks left in task-pool. [ Info: running task 9 on process 5 (Schedulers-primary-eA3Kzwpz); 5 workers total; 0 tasks left in task-pool. [ Info: running task 10 on process 6 (Schedulers-primary-eA3Kzwpz); 5 workers total; 0 tasks left in task-pool. Test Summary: | Pass Total Time pmap, stable cluster test | 16 16 2m28.4s [ Info: running task 1 on process 7 (Schedulers-primary-eA3Kzwpz); 10 workers total; 99 tasks left in task-pool. [ Info: running task 2 on process 8 (Schedulers-primary-eA3Kzwpz); 10 workers total; 98 tasks left in task-pool. ┌ Warning: addprocs/rmprocs taking longer than expected, cancelling. └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:776 [ Info: running task 3 on process 7 (Schedulers-primary-eA3Kzwpz); 10 workers total; 97 tasks left in task-pool. [ Info: running task 4 on process 8 (Schedulers-primary-eA3Kzwpz); 10 workers total; 96 tasks left in task-pool. ┌ Warning: problem adding or removing processes └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:737 ┌ Warning: TaskFailedException │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1211 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#197 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch(t::Task) │ @ Base ./task.jl:525 │ [4] loop(eloop::Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, journal::Dict{String, Any}, journal_task_callback::Function, tsk_map::Task, tsk_reduce::Task) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:735 │ [5] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1046 │ [6] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1041 │ [7] top-level scope │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:46 │ [8] macro expansion │ @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1855 [inlined] │ [9] macro expansion │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:55 [inlined] │ [10] include(mapexpr::Function, mod::Module, _path::String) │ @ Base ./Base.jl:310 │ [11] top-level scope │ @ none:6 │ [12] eval(m::Module, e::Any) │ @ Core ./boot.jl:489 │ [13] exec_options(opts::Base.JLOptions) │ @ Base ./client.jl:290 │ [14] _start() │ @ Base ./client.jl:557 │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1211 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#197 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 [ Info: running task 5 on process 7 (Schedulers-primary-eA3Kzwpz); 10 workers total; 95 tasks left in task-pool. [ Info: running task 6 on process 8 (Schedulers-primary-eA3Kzwpz); 10 workers total; 94 tasks left in task-pool. ┌ Warning: problem adding or removing processes └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:737 ┌ Warning: TaskFailedException │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1211 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#197 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch(t::Task) │ @ Base ./task.jl:525 │ [4] loop(eloop::Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, journal::Dict{String, Any}, journal_task_callback::Function, tsk_map::Task, tsk_reduce::Task) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:735 │ [5] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1046 │ [6] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1041 │ [7] top-level scope │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:46 │ [8] macro expansion │ @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1855 [inlined] │ [9] macro expansion │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:55 [inlined] │ [10] include(mapexpr::Function, mod::Module, _path::String) │ @ Base ./Base.jl:310 │ [11] top-level scope │ @ none:6 │ [12] eval(m::Module, e::Any) │ @ Core ./boot.jl:489 │ [13] exec_options(opts::Base.JLOptions) │ @ Base ./client.jl:290 │ [14] _start() │ @ Base ./client.jl:557 │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1211 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#197 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 [ Info: running task 7 on process 7 (Schedulers-primary-eA3Kzwpz); 10 workers total; 93 tasks left in task-pool. [ Info: running task 8 on process 8 (Schedulers-primary-eA3Kzwpz); 10 workers total; 92 tasks left in task-pool. ┌ Warning: problem adding or removing processes └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:737 ┌ Warning: TaskFailedException │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1211 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#197 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch(t::Task) │ @ Base ./task.jl:525 │ [4] loop(eloop::Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, journal::Dict{String, Any}, journal_task_callback::Function, tsk_map::Task, tsk_reduce::Task) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:735 │ [5] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1046 │ [6] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1041 │ [7] top-level scope │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:46 │ [8] macro expansion │ @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1855 [inlined] │ [9] macro expansion │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:55 [inlined] │ [10] include(mapexpr::Function, mod::Module, _path::String) │ @ Base ./Base.jl:310 │ [11] top-level scope │ @ none:6 │ [12] eval(m::Module, e::Any) │ @ Core ./boot.jl:489 │ [13] exec_options(opts::Base.JLOptions) │ @ Base ./client.jl:290 │ [14] _start() │ @ Base ./client.jl:557 │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1211 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#197 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 ┌ Warning: problem adding or removing processes └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:737 ┌ Warning: TaskFailedException │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1211 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#197 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch(t::Task) │ @ Base ./task.jl:525 │ [4] loop(eloop::Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, journal::Dict{String, Any}, journal_task_callback::Function, tsk_map::Task, tsk_reduce::Task) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:735 │ [5] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1046 │ [6] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1041 │ [7] top-level scope │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:46 │ [8] macro expansion │ @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1855 [inlined] │ [9] macro expansion │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:55 [inlined] │ [10] include(mapexpr::Function, mod::Module, _path::String) │ @ Base ./Base.jl:310 │ [11] top-level scope │ @ none:6 │ [12] eval(m::Module, e::Any) │ @ Core ./boot.jl:489 │ [13] exec_options(opts::Base.JLOptions) │ @ Base ./client.jl:290 │ [14] _start() │ @ Base ./client.jl:557 │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1211 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#197 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 [ Info: running task 9 on process 7 (Schedulers-primary-eA3Kzwpz); 10 workers total; 91 tasks left in task-pool. [ Info: running task 10 on process 8 (Schedulers-primary-eA3Kzwpz); 10 workers total; 90 tasks left in task-pool. ┌ Warning: problem adding or removing processes └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:737 ┌ Warning: TaskFailedException │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1211 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#197 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch(t::Task) │ @ Base ./task.jl:525 │ [4] loop(eloop::Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, journal::Dict{String, Any}, journal_task_callback::Function, tsk_map::Task, tsk_reduce::Task) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:735 │ [5] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1046 │ [6] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1041 │ [7] top-level scope │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:46 │ [8] macro expansion │ @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1855 [inlined] │ [9] macro expansion │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:55 [inlined] │ [10] include(mapexpr::Function, mod::Module, _path::String) │ @ Base ./Base.jl:310 │ [11] top-level scope │ @ none:6 │ [12] eval(m::Module, e::Any) │ @ Core ./boot.jl:489 │ [13] exec_options(opts::Base.JLOptions) │ @ Base ./client.jl:290 │ [14] _start() │ @ Base ./client.jl:557 │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1211 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#197 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 ┌ Warning: problem adding or removing processes └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:737 ┌ Warning: TaskFailedException │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1211 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#197 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch(t::Task) │ @ Base ./task.jl:525 │ [4] loop(eloop::Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, journal::Dict{String, Any}, journal_task_callback::Function, tsk_map::Task, tsk_reduce::Task) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:735 │ [5] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1046 │ [6] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1041 │ [7] top-level scope │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:46 │ [8] macro expansion │ @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1855 [inlined] │ [9] macro expansion │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:55 [inlined] │ [10] include(mapexpr::Function, mod::Module, _path::String) │ @ Base ./Base.jl:310 │ [11] top-level scope │ @ none:6 │ [12] eval(m::Module, e::Any) │ @ Core ./boot.jl:489 │ [13] exec_options(opts::Base.JLOptions) │ @ Base ./client.jl:290 │ [14] _start() │ @ Base ./client.jl:557 │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1211 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#197 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 [ Info: running task 11 on process 7 (Schedulers-primary-eA3Kzwpz); 10 workers total; 89 tasks left in task-pool. [ Info: running task 12 on process 8 (Schedulers-primary-eA3Kzwpz); 10 workers total; 88 tasks left in task-pool. ┌ Warning: problem adding or removing processes └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:737 ┌ Warning: TaskFailedException │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1211 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#197 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch(t::Task) │ @ Base ./task.jl:525 │ [4] loop(eloop::Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, journal::Dict{String, Any}, journal_task_callback::Function, tsk_map::Task, tsk_reduce::Task) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:735 │ [5] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1046 │ [6] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1041 │ [7] top-level scope │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:46 │ [8] macro expansion │ @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1855 [inlined] │ [9] macro expansion │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:55 [inlined] │ [10] include(mapexpr::Function, mod::Module, _path::String) │ @ Base ./Base.jl:310 │ [11] top-level scope │ @ none:6 │ [12] eval(m::Module, e::Any) │ @ Core ./boot.jl:489 │ [13] exec_options(opts::Base.JLOptions) │ @ Base ./client.jl:290 │ [14] _start() │ @ Base ./client.jl:557 │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1211 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#197 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 [ Info: running task 13 on process 7 (Schedulers-primary-eA3Kzwpz); 10 workers total; 87 tasks left in task-pool. [ Info: running task 14 on process 8 (Schedulers-primary-eA3Kzwpz); 10 workers total; 86 tasks left in task-pool. ┌ Warning: problem adding or removing processes └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:737 ┌ Warning: TaskFailedException │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1211 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#197 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch(t::Task) │ @ Base ./task.jl:525 │ [4] loop(eloop::Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, journal::Dict{String, Any}, journal_task_callback::Function, tsk_map::Task, tsk_reduce::Task) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:735 │ [5] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1046 │ [6] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1041 │ [7] top-level scope │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:46 │ [8] macro expansion │ @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1855 [inlined] │ [9] macro expansion │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:55 [inlined] │ [10] include(mapexpr::Function, mod::Module, _path::String) │ @ Base ./Base.jl:310 │ [11] top-level scope │ @ none:6 │ [12] eval(m::Module, e::Any) │ @ Core ./boot.jl:489 │ [13] exec_options(opts::Base.JLOptions) │ @ Base ./client.jl:290 │ [14] _start() │ @ Base ./client.jl:557 │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1211 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#197 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 From worker 11: ErrorException("Process(1) - Invalid connection credentials sent by remote.")CapturedException(ErrorException("Process(1) - Invalid connection credentials sent by remote."), Any[(error(s::String) at Master process (id 1) could not connect within 120.0 seconds. From worker 11: exiting. From worker 10: ErrorException("Process(1) - Invalid connection credentials sent by remote.")CapturedException(ErrorException("Process(1) - Invalid connection credentials sent by remote."), Any[(error(s::String) at Master process (id 1) could not connect within 120.0 seconds. From worker 10: exiting. From worker 14: ErrorException("Process(1) - Invalid connection credentials sent by remote.")CapturedException(ErrorException("Process(1) - Invalid connection credentials sent by remote."), Any[(error(s::String) at Master process (id 1) could not connect within 120.0 seconds. From worker 14: exiting. ┌ Warning: problem adding or removing processes └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:737 ┌ Warning: TaskFailedException │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1211 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#197 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch(t::Task) │ @ Base ./task.jl:525 │ [4] loop(eloop::Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, journal::Dict{String, Any}, journal_task_callback::Function, tsk_map::Task, tsk_reduce::Task) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:735 │ [5] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1046 │ [6] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1041 │ [7] top-level scope │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:46 │ [8] macro expansion │ @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1855 [inlined] │ [9] macro expansion │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:55 [inlined] │ [10] include(mapexpr::Function, mod::Module, _path::String) │ @ Base ./Base.jl:310 │ [11] top-level scope │ @ none:6 │ [12] eval(m::Module, e::Any) │ @ Core ./boot.jl:489 │ [13] exec_options(opts::Base.JLOptions) │ @ Base ./client.jl:290 │ [14] _start() │ @ Base ./client.jl:557 │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1211 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#197 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 From worker 11: error.jl: From worker 10: error.jl: From worker 14: error.jl: Worker 11 terminated. [ Info: running task 15 on process 7 (Schedulers-primary-eA3Kzwpz); 7 workers total; 85 tasks left in task-pool. Worker 10 terminated.[ Info: running task 16 on process 8 (Schedulers-primary-eA3Kzwpz); 7 workers total; 84 tasks left in task-pool. UNHANDLED TASK ERROR: Version read failed. Connection closed by peer. Stacktrace: [1] error(s::String) @ Base ./error.jl:44 [2] process_hdr(s::Sockets.TCPSocket, validate_cookie::Bool) @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:276 [3] message_handler_loop(r_stream::Sockets.TCPSocket, w_stream::Sockets.TCPSocket, incoming::Bool) @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:158 [4] process_tcp_streams(r_stream::Sockets.TCPSocket, w_stream::Sockets.TCPSocket, incoming::Bool) @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:133 [5] (::Distributed.var"#process_messages##0#process_messages##1"{Sockets.TCPSocket, Sockets.TCPSocket, Bool})() @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:121 Worker 14 terminated.UNHANDLED TASK ERROR: Version read failed. Connection closed by peer. Stacktrace: [1] error(s::String) @ Base ./error.jl:44 [2] process_hdr(s::Sockets.TCPSocket, validate_cookie::Bool) @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:276 [3] message_handler_loop(r_stream::Sockets.TCPSocket, w_stream::Sockets.TCPSocket, incoming::Bool) @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:158 [4] process_tcp_streams(r_stream::Sockets.TCPSocket, w_stream::Sockets.TCPSocket, incoming::Bool) @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:133 [5] (::Distributed.var"#process_messages##0#process_messages##1"{Sockets.TCPSocket, Sockets.TCPSocket, Bool})() @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:121 ┌ Warning: problem adding or removing processes └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:737 ┌ Warning: TaskFailedException │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1211 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#197 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch(t::Task) │ @ Base ./task.jl:525 │ [4] loop(eloop::Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, journal::Dict{String, Any}, journal_task_callback::Function, tsk_map::Task, tsk_reduce::Task) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:735 │ [5] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1046 │ [6] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1041 │ [7] top-level scope │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:46 │ [8] macro expansion │ @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1855 [inlined] │ [9] macro expansion │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:55 [inlined] │ [10] include(mapexpr::Function, mod::Module, _path::String) │ @ Base ./Base.jl:310 │ [11] top-level scope │ @ none:6 │ [12] eval(m::Module, e::Any) │ @ Core ./boot.jl:489 │ [13] exec_options(opts::Base.JLOptions) │ @ Base ./client.jl:290 │ [14] _start() │ @ Base ./client.jl:557 │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1211 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#197 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 UNHANDLED TASK ERROR: Version read failed. Connection closed by peer. Stacktrace: [1] error(s::String) @ Base ./error.jl:44 [2] process_hdr(s::Sockets.TCPSocket, validate_cookie::Bool) @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:276 [3] message_handler_loop(r_stream::Sockets.TCPSocket, w_stream::Sockets.TCPSocket, incoming::Bool) @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:158 [4] process_tcp_streams(r_stream::Sockets.TCPSocket, w_stream::Sockets.TCPSocket, incoming::Bool) @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:133 [5] (::Distributed.var"#process_messages##0#process_messages##1"{Sockets.TCPSocket, Sockets.TCPSocket, Bool})() @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:121 [ Info: running task 17 on process 9 (Schedulers-primary-eA3Kzwpz); 7 workers total; 83 tasks left in task-pool. [ Info: running task 18 on process 12 (Schedulers-primary-eA3Kzwpz); 7 workers total; 79 tasks left in task-pool. [ Info: running task 19 on process 13 (Schedulers-primary-eA3Kzwpz); 7 workers total; 79 tasks left in task-pool. [ Info: running task 20 on process 15 (Schedulers-primary-eA3Kzwpz); 7 workers total; 79 tasks left in task-pool. [ Info: running task 21 on process 16 (Schedulers-primary-eA3Kzwpz); 7 workers total; 79 tasks left in task-pool. [ Info: running task 22 on process 7 (Schedulers-primary-eA3Kzwpz); 7 workers total; 78 tasks left in task-pool. [ Info: running task 23 on process 8 (Schedulers-primary-eA3Kzwpz); 7 workers total; 77 tasks left in task-pool. [ Info: running task 24 on process 9 (Schedulers-primary-eA3Kzwpz); 10 workers total; 76 tasks left in task-pool. [ Info: running task 25 on process 12 (Schedulers-primary-eA3Kzwpz); 10 workers total; 75 tasks left in task-pool. [ Info: running task 26 on process 13 (Schedulers-primary-eA3Kzwpz); 10 workers total; 74 tasks left in task-pool. [ Info: running task 27 on process 15 (Schedulers-primary-eA3Kzwpz); 10 workers total; 73 tasks left in task-pool. [ Info: running task 28 on process 16 (Schedulers-primary-eA3Kzwpz); 10 workers total; 72 tasks left in task-pool. [ Info: running task 29 on process 7 (Schedulers-primary-eA3Kzwpz); 10 workers total; 71 tasks left in task-pool. [ Info: running task 30 on process 8 (Schedulers-primary-eA3Kzwpz); 10 workers total; 70 tasks left in task-pool. [ Info: running task 31 on process 9 (Schedulers-primary-eA3Kzwpz); 10 workers total; 69 tasks left in task-pool. [ Info: running task 32 on process 12 (Schedulers-primary-eA3Kzwpz); 10 workers total; 68 tasks left in task-pool. [ Info: running task 33 on process 15 (Schedulers-primary-eA3Kzwpz); 10 workers total; 67 tasks left in task-pool. [ Info: running task 34 on process 16 (Schedulers-primary-eA3Kzwpz); 10 workers total; 66 tasks left in task-pool. [ Info: running task 35 on process 13 (Schedulers-primary-eA3Kzwpz); 10 workers total; 65 tasks left in task-pool. [ Info: running task 36 on process 7 (Schedulers-primary-eA3Kzwpz); 10 workers total; 64 tasks left in task-pool. [ Info: running task 37 on process 8 (Schedulers-primary-eA3Kzwpz); 10 workers total; 63 tasks left in task-pool. [ Info: running task 38 on process 9 (Schedulers-primary-eA3Kzwpz); 10 workers total; 62 tasks left in task-pool. [ Info: running task 39 on process 16 (Schedulers-primary-eA3Kzwpz); 10 workers total; 61 tasks left in task-pool. [ Info: running task 40 on process 12 (Schedulers-primary-eA3Kzwpz); 10 workers total; 60 tasks left in task-pool. [ Info: running task 41 on process 15 (Schedulers-primary-eA3Kzwpz); 10 workers total; 59 tasks left in task-pool. [ Info: running task 42 on process 13 (Schedulers-primary-eA3Kzwpz); 10 workers total; 58 tasks left in task-pool. [ Info: running task 43 on process 7 (Schedulers-primary-eA3Kzwpz); 10 workers total; 57 tasks left in task-pool. [ Info: running task 44 on process 8 (Schedulers-primary-eA3Kzwpz); 10 workers total; 56 tasks left in task-pool. [ Info: running task 45 on process 9 (Schedulers-primary-eA3Kzwpz); 10 workers total; 55 tasks left in task-pool. [ Info: running task 46 on process 16 (Schedulers-primary-eA3Kzwpz); 10 workers total; 54 tasks left in task-pool. [ Info: running task 47 on process 12 (Schedulers-primary-eA3Kzwpz); 10 workers total; 53 tasks left in task-pool. [ Info: running task 48 on process 15 (Schedulers-primary-eA3Kzwpz); 10 workers total; 52 tasks left in task-pool. [ Info: running task 49 on process 13 (Schedulers-primary-eA3Kzwpz); 10 workers total; 51 tasks left in task-pool. [ Info: running task 50 on process 8 (Schedulers-primary-eA3Kzwpz); 10 workers total; 50 tasks left in task-pool. [ Info: running task 51 on process 7 (Schedulers-primary-eA3Kzwpz); 10 workers total; 49 tasks left in task-pool. [ Info: running task 52 on process 9 (Schedulers-primary-eA3Kzwpz); 10 workers total; 48 tasks left in task-pool. [ Info: running task 53 on process 16 (Schedulers-primary-eA3Kzwpz); 10 workers total; 47 tasks left in task-pool. [ Info: running task 54 on process 12 (Schedulers-primary-eA3Kzwpz); 10 workers total; 46 tasks left in task-pool. [ Info: running task 55 on process 15 (Schedulers-primary-eA3Kzwpz); 10 workers total; 45 tasks left in task-pool. [ Info: running task 56 on process 17 (Schedulers-primary-eA3Kzwpz); 10 workers total; 44 tasks left in task-pool. [ Info: running task 57 on process 18 (Schedulers-primary-eA3Kzwpz); 10 workers total; 42 tasks left in task-pool. [ Info: running task 58 on process 19 (Schedulers-primary-eA3Kzwpz); 10 workers total; 42 tasks left in task-pool. [ Info: running task 59 on process 13 (Schedulers-primary-eA3Kzwpz); 10 workers total; 41 tasks left in task-pool. [ Info: running task 60 on process 8 (Schedulers-primary-eA3Kzwpz); 10 workers total; 40 tasks left in task-pool. [ Info: running task 61 on process 7 (Schedulers-primary-eA3Kzwpz); 10 workers total; 39 tasks left in task-pool. [ Info: running task 62 on process 9 (Schedulers-primary-eA3Kzwpz); 10 workers total; 38 tasks left in task-pool. [ Info: running task 63 on process 16 (Schedulers-primary-eA3Kzwpz); 10 workers total; 37 tasks left in task-pool. [ Info: running task 64 on process 12 (Schedulers-primary-eA3Kzwpz); 10 workers total; 36 tasks left in task-pool. [ Info: running task 65 on process 15 (Schedulers-primary-eA3Kzwpz); 10 workers total; 35 tasks left in task-pool. [ Info: running task 66 on process 13 (Schedulers-primary-eA3Kzwpz); 10 workers total; 34 tasks left in task-pool. [ Info: running task 67 on process 8 (Schedulers-primary-eA3Kzwpz); 10 workers total; 33 tasks left in task-pool. [ Info: running task 68 on process 17 (Schedulers-primary-eA3Kzwpz); 10 workers total; 32 tasks left in task-pool. [ Info: running task 69 on process 18 (Schedulers-primary-eA3Kzwpz); 10 workers total; 31 tasks left in task-pool. [ Info: running task 70 on process 19 (Schedulers-primary-eA3Kzwpz); 10 workers total; 30 tasks left in task-pool. [ Info: running task 71 on process 7 (Schedulers-primary-eA3Kzwpz); 10 workers total; 29 tasks left in task-pool. [ Info: running task 72 on process 12 (Schedulers-primary-eA3Kzwpz); 10 workers total; 28 tasks left in task-pool. [ Info: running task 73 on process 9 (Schedulers-primary-eA3Kzwpz); 10 workers total; 27 tasks left in task-pool. [ Info: running task 74 on process 16 (Schedulers-primary-eA3Kzwpz); 10 workers total; 26 tasks left in task-pool. [ Info: running task 75 on process 15 (Schedulers-primary-eA3Kzwpz); 10 workers total; 25 tasks left in task-pool. [ Info: running task 76 on process 13 (Schedulers-primary-eA3Kzwpz); 10 workers total; 24 tasks left in task-pool. [ Info: running task 77 on process 8 (Schedulers-primary-eA3Kzwpz); 10 workers total; 22 tasks left in task-pool. [ Info: running task 78 on process 17 (Schedulers-primary-eA3Kzwpz); 10 workers total; 22 tasks left in task-pool. [ Info: running task 79 on process 18 (Schedulers-primary-eA3Kzwpz); 10 workers total; 21 tasks left in task-pool. [ Info: running task 80 on process 19 (Schedulers-primary-eA3Kzwpz); 10 workers total; 20 tasks left in task-pool. [ Info: running task 81 on process 7 (Schedulers-primary-eA3Kzwpz); 10 workers total; 19 tasks left in task-pool. [ Info: running task 82 on process 12 (Schedulers-primary-eA3Kzwpz); 10 workers total; 18 tasks left in task-pool. [ Info: running task 83 on process 9 (Schedulers-primary-eA3Kzwpz); 10 workers total; 16 tasks left in task-pool. [ Info: running task 84 on process 16 (Schedulers-primary-eA3Kzwpz); 10 workers total; 16 tasks left in task-pool. [ Info: running task 85 on process 15 (Schedulers-primary-eA3Kzwpz); 10 workers total; 15 tasks left in task-pool. [ Info: running task 86 on process 13 (Schedulers-primary-eA3Kzwpz); 10 workers total; 14 tasks left in task-pool. [ Info: running task 87 on process 8 (Schedulers-primary-eA3Kzwpz); 10 workers total; 9 tasks left in task-pool. [ Info: running task 88 on process 17 (Schedulers-primary-eA3Kzwpz); 10 workers total; 9 tasks left in task-pool. [ Info: running task 89 on process 7 (Schedulers-primary-eA3Kzwpz); 10 workers total; 9 tasks left in task-pool. [ Info: running task 90 on process 18 (Schedulers-primary-eA3Kzwpz); 10 workers total; 9 tasks left in task-pool. [ Info: running task 91 on process 19 (Schedulers-primary-eA3Kzwpz); 10 workers total; 9 tasks left in task-pool. [ Info: running task 92 on process 12 (Schedulers-primary-eA3Kzwpz); 10 workers total; 8 tasks left in task-pool. [ Info: running task 93 on process 9 (Schedulers-primary-eA3Kzwpz); 10 workers total; 6 tasks left in task-pool. [ Info: running task 94 on process 16 (Schedulers-primary-eA3Kzwpz); 10 workers total; 6 tasks left in task-pool. [ Info: running task 95 on process 15 (Schedulers-primary-eA3Kzwpz); 10 workers total; 5 tasks left in task-pool. [ Info: running task 96 on process 13 (Schedulers-primary-eA3Kzwpz); 10 workers total; 4 tasks left in task-pool. [ Info: running task 97 on process 8 (Schedulers-primary-eA3Kzwpz); 10 workers total; 0 tasks left in task-pool. [ Info: running task 98 on process 17 (Schedulers-primary-eA3Kzwpz); 10 workers total; 0 tasks left in task-pool. [ Info: running task 99 on process 7 (Schedulers-primary-eA3Kzwpz); 10 workers total; 0 tasks left in task-pool. [ Info: running task 100 on process 18 (Schedulers-primary-eA3Kzwpz); 10 workers total; 0 tasks left in task-pool. Test Summary: | Pass Total Time pmap, growing cluster test | 111 111 5m28.0s [ Info: running task 1 on process 20 (Schedulers-primary-eA3Kzwpz); 8 workers total; 99 tasks left in task-pool. [ Info: running task 2 on process 27 (Schedulers-primary-eA3Kzwpz); 8 workers total; 98 tasks left in task-pool. [ Info: running task 3 on process 21 (Schedulers-primary-eA3Kzwpz); 8 workers total; 97 tasks left in task-pool. [ Info: running task 4 on process 22 (Schedulers-primary-eA3Kzwpz); 8 workers total; 92 tasks left in task-pool. [ Info: running task 5 on process 24 (Schedulers-primary-eA3Kzwpz); 8 workers total; 92 tasks left in task-pool. [ Info: running task 6 on process 25 (Schedulers-primary-eA3Kzwpz); 8 workers total; 92 tasks left in task-pool. [ Info: running task 7 on process 26 (Schedulers-primary-eA3Kzwpz); 8 workers total; 92 tasks left in task-pool. [ Info: running task 8 on process 28 (Schedulers-primary-eA3Kzwpz); 8 workers total; 92 tasks left in task-pool. [ Info: running task 9 on process 20 (Schedulers-primary-eA3Kzwpz); 10 workers total; 91 tasks left in task-pool. [ Info: running task 10 on process 27 (Schedulers-primary-eA3Kzwpz); 10 workers total; 90 tasks left in task-pool. [ Info: running task 11 on process 21 (Schedulers-primary-eA3Kzwpz); 10 workers total; 89 tasks left in task-pool. [ Info: running task 12 on process 22 (Schedulers-primary-eA3Kzwpz); 10 workers total; 84 tasks left in task-pool. [ Info: running task 13 on process 24 (Schedulers-primary-eA3Kzwpz); 10 workers total; 84 tasks left in task-pool. [ Info: running task 14 on process 25 (Schedulers-primary-eA3Kzwpz); 10 workers total; 84 tasks left in task-pool. [ Info: running task 15 on process 26 (Schedulers-primary-eA3Kzwpz); 10 workers total; 84 tasks left in task-pool. [ Info: running task 16 on process 28 (Schedulers-primary-eA3Kzwpz); 10 workers total; 84 tasks left in task-pool. [ Info: running task 17 on process 20 (Schedulers-primary-eA3Kzwpz); 10 workers total; 83 tasks left in task-pool. [ Info: running task 18 on process 27 (Schedulers-primary-eA3Kzwpz); 10 workers total; 82 tasks left in task-pool. [ Info: running task 19 on process 21 (Schedulers-primary-eA3Kzwpz); 10 workers total; 81 tasks left in task-pool. [ Info: running task 20 on process 25 (Schedulers-primary-eA3Kzwpz); 10 workers total; 78 tasks left in task-pool. [ Info: running task 21 on process 26 (Schedulers-primary-eA3Kzwpz); 10 workers total; 78 tasks left in task-pool. [ Info: running task 22 on process 28 (Schedulers-primary-eA3Kzwpz); 10 workers total; 78 tasks left in task-pool. [ Info: running task 23 on process 22 (Schedulers-primary-eA3Kzwpz); 10 workers total; 77 tasks left in task-pool. [ Info: running task 24 on process 24 (Schedulers-primary-eA3Kzwpz); 10 workers total; 76 tasks left in task-pool. [ Info: running task 25 on process 30 (Schedulers-primary-eA3Kzwpz); 10 workers total; 75 tasks left in task-pool. [ Info: running task 26 on process 31 (Schedulers-primary-eA3Kzwpz); 10 workers total; 74 tasks left in task-pool. [ Info: running task 27 on process 20 (Schedulers-primary-eA3Kzwpz); 10 workers total; 73 tasks left in task-pool. [ Info: running task 28 on process 27 (Schedulers-primary-eA3Kzwpz); 10 workers total; 72 tasks left in task-pool. [ Info: running task 29 on process 25 (Schedulers-primary-eA3Kzwpz); 10 workers total; 71 tasks left in task-pool. [ Info: running task 30 on process 26 (Schedulers-primary-eA3Kzwpz); 10 workers total; 69 tasks left in task-pool. [ Info: running task 31 on process 28 (Schedulers-primary-eA3Kzwpz); 10 workers total; 69 tasks left in task-pool. [ Info: running task 32 on process 21 (Schedulers-primary-eA3Kzwpz); 10 workers total; 68 tasks left in task-pool. [ Info: running task 33 on process 22 (Schedulers-primary-eA3Kzwpz); 10 workers total; 66 tasks left in task-pool. [ Info: running task 34 on process 24 (Schedulers-primary-eA3Kzwpz); 10 workers total; 66 tasks left in task-pool. [ Info: running task 35 on process 20 (Schedulers-primary-eA3Kzwpz); 10 workers total; 65 tasks left in task-pool. [ Info: running task 36 on process 30 (Schedulers-primary-eA3Kzwpz); 10 workers total; 64 tasks left in task-pool. [ Info: running task 37 on process 31 (Schedulers-primary-eA3Kzwpz); 10 workers total; 63 tasks left in task-pool. [ Info: running task 38 on process 27 (Schedulers-primary-eA3Kzwpz); 10 workers total; 62 tasks left in task-pool. [ Info: running task 39 on process 25 (Schedulers-primary-eA3Kzwpz); 10 workers total; 59 tasks left in task-pool. [ Info: running task 40 on process 26 (Schedulers-primary-eA3Kzwpz); 10 workers total; 59 tasks left in task-pool. [ Info: running task 41 on process 28 (Schedulers-primary-eA3Kzwpz); 10 workers total; 59 tasks left in task-pool. [ Info: running task 42 on process 21 (Schedulers-primary-eA3Kzwpz); 10 workers total; 58 tasks left in task-pool. [ Info: running task 43 on process 22 (Schedulers-primary-eA3Kzwpz); 10 workers total; 56 tasks left in task-pool. [ Info: running task 44 on process 24 (Schedulers-primary-eA3Kzwpz); 10 workers total; 56 tasks left in task-pool. [ Info: running task 45 on process 20 (Schedulers-primary-eA3Kzwpz); 10 workers total; 55 tasks left in task-pool. [ Info: running task 46 on process 30 (Schedulers-primary-eA3Kzwpz); 10 workers total; 54 tasks left in task-pool. [ Info: running task 47 on process 27 (Schedulers-primary-eA3Kzwpz); 10 workers total; 52 tasks left in task-pool. [ Info: running task 48 on process 31 (Schedulers-primary-eA3Kzwpz); 10 workers total; 52 tasks left in task-pool. [ Info: running task 49 on process 25 (Schedulers-primary-eA3Kzwpz); 10 workers total; 51 tasks left in task-pool. [ Info: running task 50 on process 26 (Schedulers-primary-eA3Kzwpz); 10 workers total; 50 tasks left in task-pool. [ Info: running task 51 on process 28 (Schedulers-primary-eA3Kzwpz); 10 workers total; 49 tasks left in task-pool. [ Info: running task 52 on process 21 (Schedulers-primary-eA3Kzwpz); 10 workers total; 48 tasks left in task-pool. [ Info: running task 53 on process 22 (Schedulers-primary-eA3Kzwpz); 10 workers total; 46 tasks left in task-pool. [ Info: running task 54 on process 24 (Schedulers-primary-eA3Kzwpz); 10 workers total; 46 tasks left in task-pool. [ Info: running task 55 on process 20 (Schedulers-primary-eA3Kzwpz); 10 workers total; 45 tasks left in task-pool. [ Info: running task 56 on process 30 (Schedulers-primary-eA3Kzwpz); 10 workers total; 44 tasks left in task-pool. [ Info: running task 57 on process 27 (Schedulers-primary-eA3Kzwpz); 10 workers total; 42 tasks left in task-pool. [ Info: running task 58 on process 31 (Schedulers-primary-eA3Kzwpz); 10 workers total; 42 tasks left in task-pool. [ Info: running task 59 on process 25 (Schedulers-primary-eA3Kzwpz); 10 workers total; 41 tasks left in task-pool. [ Info: running task 60 on process 26 (Schedulers-primary-eA3Kzwpz); 10 workers total; 39 tasks left in task-pool. [ Info: running task 61 on process 28 (Schedulers-primary-eA3Kzwpz); 10 workers total; 39 tasks left in task-pool. [ Info: running task 62 on process 22 (Schedulers-primary-eA3Kzwpz); 10 workers total; 38 tasks left in task-pool. [ Info: running task 63 on process 24 (Schedulers-primary-eA3Kzwpz); 10 workers total; 37 tasks left in task-pool. [ Info: running task 64 on process 21 (Schedulers-primary-eA3Kzwpz); 10 workers total; 36 tasks left in task-pool. [ Info: running task 65 on process 20 (Schedulers-primary-eA3Kzwpz); 10 workers total; 35 tasks left in task-pool. [ Info: running task 66 on process 30 (Schedulers-primary-eA3Kzwpz); 10 workers total; 34 tasks left in task-pool. [ Info: running task 67 on process 27 (Schedulers-primary-eA3Kzwpz); 10 workers total; 32 tasks left in task-pool. [ Info: running task 68 on process 31 (Schedulers-primary-eA3Kzwpz); 10 workers total; 32 tasks left in task-pool. [ Info: running task 69 on process 22 (Schedulers-primary-eA3Kzwpz); 10 workers total; 31 tasks left in task-pool. [ Info: running task 70 on process 24 (Schedulers-primary-eA3Kzwpz); 10 workers total; 30 tasks left in task-pool. [ Info: running task 71 on process 25 (Schedulers-primary-eA3Kzwpz); 10 workers total; 29 tasks left in task-pool. [ Info: running task 72 on process 26 (Schedulers-primary-eA3Kzwpz); 10 workers total; 28 tasks left in task-pool. [ Info: running task 73 on process 28 (Schedulers-primary-eA3Kzwpz); 10 workers total; 27 tasks left in task-pool. [ Info: running task 74 on process 21 (Schedulers-primary-eA3Kzwpz); 10 workers total; 26 tasks left in task-pool. [ Info: running task 75 on process 20 (Schedulers-primary-eA3Kzwpz); 10 workers total; 25 tasks left in task-pool. [ Info: running task 76 on process 30 (Schedulers-primary-eA3Kzwpz); 10 workers total; 24 tasks left in task-pool. [ Info: running task 77 on process 27 (Schedulers-primary-eA3Kzwpz); 10 workers total; 23 tasks left in task-pool. [ Info: running task 78 on process 31 (Schedulers-primary-eA3Kzwpz); 10 workers total; 22 tasks left in task-pool. [ Info: running task 79 on process 26 (Schedulers-primary-eA3Kzwpz); 10 workers total; 21 tasks left in task-pool. [ Info: running task 80 on process 28 (Schedulers-primary-eA3Kzwpz); 10 workers total; 20 tasks left in task-pool. [ Info: running task 81 on process 22 (Schedulers-primary-eA3Kzwpz); 10 workers total; 19 tasks left in task-pool. [ Info: running task 82 on process 24 (Schedulers-primary-eA3Kzwpz); 10 workers total; 18 tasks left in task-pool. [ Info: running task 83 on process 25 (Schedulers-primary-eA3Kzwpz); 10 workers total; 17 tasks left in task-pool. [ Info: running task 84 on process 21 (Schedulers-primary-eA3Kzwpz); 10 workers total; 16 tasks left in task-pool. [ Info: running task 85 on process 20 (Schedulers-primary-eA3Kzwpz); 10 workers total; 15 tasks left in task-pool. [ Info: running task 86 on process 30 (Schedulers-primary-eA3Kzwpz); 10 workers total; 14 tasks left in task-pool. [ Info: running task 87 on process 27 (Schedulers-primary-eA3Kzwpz); 10 workers total; 12 tasks left in task-pool. [ Info: running task 88 on process 31 (Schedulers-primary-eA3Kzwpz); 10 workers total; 12 tasks left in task-pool. [ Info: running task 89 on process 26 (Schedulers-primary-eA3Kzwpz); 10 workers total; 11 tasks left in task-pool. [ Info: running task 90 on process 28 (Schedulers-primary-eA3Kzwpz); 10 workers total; 8 tasks left in task-pool. [ Info: running task 91 on process 22 (Schedulers-primary-eA3Kzwpz); 10 workers total; 8 tasks left in task-pool. [ Info: running task 92 on process 24 (Schedulers-primary-eA3Kzwpz); 10 workers total; 8 tasks left in task-pool. [ Info: running task 93 on process 25 (Schedulers-primary-eA3Kzwpz); 10 workers total; 7 tasks left in task-pool. [ Info: running task 94 on process 21 (Schedulers-primary-eA3Kzwpz); 10 workers total; 6 tasks left in task-pool. [ Info: running task 95 on process 20 (Schedulers-primary-eA3Kzwpz); 10 workers total; 5 tasks left in task-pool. [ Info: running task 96 on process 30 (Schedulers-primary-eA3Kzwpz); 10 workers total; 4 tasks left in task-pool. [ Info: running task 97 on process 31 (Schedulers-primary-eA3Kzwpz); 10 workers total; 3 tasks left in task-pool. [ Info: running task 98 on process 26 (Schedulers-primary-eA3Kzwpz); 10 workers total; 2 tasks left in task-pool. [ Info: running task 99 on process 28 (Schedulers-primary-eA3Kzwpz); 10 workers total; 0 tasks left in task-pool. [ Info: running task 100 on process 27 (Schedulers-primary-eA3Kzwpz); 10 workers total; 0 tasks left in task-pool. Test Summary: | Pass Total Time pmap, elastic cluster with faults | 111 111 4m57.9s [ Info: running task 1 on process 32 (Schedulers-primary-eA3Kzwpz); 10 workers total; 104 tasks left in task-pool. [ Info: running task 2 on process 33 (Schedulers-primary-eA3Kzwpz); 10 workers total; 95 tasks left in task-pool. [ Info: running task 3 on process 34 (Schedulers-primary-eA3Kzwpz); 10 workers total; 95 tasks left in task-pool. [ Info: running task 4 on process 35 (Schedulers-primary-eA3Kzwpz); 10 workers total; 95 tasks left in task-pool. [ Info: running task 5 on process 36 (Schedulers-primary-eA3Kzwpz); 10 workers total; 95 tasks left in task-pool. [ Info: running task 6 on process 37 (Schedulers-primary-eA3Kzwpz); 10 workers total; 95 tasks left in task-pool. [ Info: running task 7 on process 38 (Schedulers-primary-eA3Kzwpz); 10 workers total; 95 tasks left in task-pool. [ Info: running task 8 on process 39 (Schedulers-primary-eA3Kzwpz); 10 workers total; 95 tasks left in task-pool. [ Info: running task 9 on process 40 (Schedulers-primary-eA3Kzwpz); 10 workers total; 95 tasks left in task-pool. [ Info: running task 10 on process 41 (Schedulers-primary-eA3Kzwpz); 10 workers total; 95 tasks left in task-pool. [ Info: running task 11 on process 32 (Schedulers-primary-eA3Kzwpz); 10 workers total; 94 tasks left in task-pool. [ Info: running task 12 on process 33 (Schedulers-primary-eA3Kzwpz); 10 workers total; 93 tasks left in task-pool. [ Info: running task 13 on process 34 (Schedulers-primary-eA3Kzwpz); 10 workers total; 92 tasks left in task-pool. [ Info: running task 14 on process 35 (Schedulers-primary-eA3Kzwpz); 10 workers total; 85 tasks left in task-pool. [ Info: running task 15 on process 36 (Schedulers-primary-eA3Kzwpz); 10 workers total; 85 tasks left in task-pool. [ Info: running task 16 on process 37 (Schedulers-primary-eA3Kzwpz); 10 workers total; 85 tasks left in task-pool. [ Info: running task 17 on process 38 (Schedulers-primary-eA3Kzwpz); 10 workers total; 85 tasks left in task-pool. [ Info: running task 18 on process 39 (Schedulers-primary-eA3Kzwpz); 10 workers total; 85 tasks left in task-pool. [ Info: running task 19 on process 40 (Schedulers-primary-eA3Kzwpz); 10 workers total; 85 tasks left in task-pool. [ Info: running task 20 on process 41 (Schedulers-primary-eA3Kzwpz); 10 workers total; 85 tasks left in task-pool. [ Info: running task 21 on process 32 (Schedulers-primary-eA3Kzwpz); 10 workers total; 84 tasks left in task-pool. [ Info: running task 22 on process 33 (Schedulers-primary-eA3Kzwpz); 10 workers total; 83 tasks left in task-pool. [ Info: running task 23 on process 34 (Schedulers-primary-eA3Kzwpz); 10 workers total; 82 tasks left in task-pool. [ Info: running task 24 on process 35 (Schedulers-primary-eA3Kzwpz); 10 workers total; 75 tasks left in task-pool. [ Info: running task 25 on process 36 (Schedulers-primary-eA3Kzwpz); 10 workers total; 75 tasks left in task-pool. [ Info: running task 26 on process 37 (Schedulers-primary-eA3Kzwpz); 10 workers total; 75 tasks left in task-pool. [ Info: running task 27 on process 38 (Schedulers-primary-eA3Kzwpz); 10 workers total; 75 tasks left in task-pool. [ Info: running task 28 on process 39 (Schedulers-primary-eA3Kzwpz); 10 workers total; 75 tasks left in task-pool. [ Info: running task 29 on process 40 (Schedulers-primary-eA3Kzwpz); 10 workers total; 75 tasks left in task-pool. [ Info: running task 30 on process 41 (Schedulers-primary-eA3Kzwpz); 10 workers total; 75 tasks left in task-pool. [ Info: running task 31 on process 32 (Schedulers-primary-eA3Kzwpz); 10 workers total; 74 tasks left in task-pool. [ Info: running task 32 on process 33 (Schedulers-primary-eA3Kzwpz); 10 workers total; 73 tasks left in task-pool. [ Info: running task 33 on process 34 (Schedulers-primary-eA3Kzwpz); 10 workers total; 72 tasks left in task-pool. [ Info: running task 34 on process 35 (Schedulers-primary-eA3Kzwpz); 10 workers total; 65 tasks left in task-pool. [ Info: running task 35 on process 36 (Schedulers-primary-eA3Kzwpz); 10 workers total; 65 tasks left in task-pool. [ Info: running task 36 on process 37 (Schedulers-primary-eA3Kzwpz); 10 workers total; 65 tasks left in task-pool. [ Info: running task 37 on process 38 (Schedulers-primary-eA3Kzwpz); 10 workers total; 65 tasks left in task-pool. [ Info: running task 38 on process 39 (Schedulers-primary-eA3Kzwpz); 10 workers total; 65 tasks left in task-pool. [ Info: running task 39 on process 40 (Schedulers-primary-eA3Kzwpz); 10 workers total; 65 tasks left in task-pool. [ Info: running task 40 on process 41 (Schedulers-primary-eA3Kzwpz); 10 workers total; 65 tasks left in task-pool. [ Info: running task 41 on process 32 (Schedulers-primary-eA3Kzwpz); 10 workers total; 64 tasks left in task-pool. [ Info: running task 42 on process 36 (Schedulers-primary-eA3Kzwpz); 10 workers total; 63 tasks left in task-pool. [ Info: running task 43 on process 37 (Schedulers-primary-eA3Kzwpz); 10 workers total; 60 tasks left in task-pool. [ Info: running task 44 on process 38 (Schedulers-primary-eA3Kzwpz); 10 workers total; 60 tasks left in task-pool. [ Info: running task 45 on process 39 (Schedulers-primary-eA3Kzwpz); 10 workers total; 60 tasks left in task-pool. [ Info: running task 46 on process 33 (Schedulers-primary-eA3Kzwpz); 10 workers total; 59 tasks left in task-pool. [ Info: running task 47 on process 34 (Schedulers-primary-eA3Kzwpz); 10 workers total; 58 tasks left in task-pool. [ Info: running task 48 on process 35 (Schedulers-primary-eA3Kzwpz); 10 workers total; 57 tasks left in task-pool. [ Info: running task 49 on process 40 (Schedulers-primary-eA3Kzwpz); 10 workers total; 56 tasks left in task-pool. [ Info: running task 50 on process 41 (Schedulers-primary-eA3Kzwpz); 10 workers total; 55 tasks left in task-pool. [ Info: running task 51 on process 32 (Schedulers-primary-eA3Kzwpz); 10 workers total; 54 tasks left in task-pool. [ Info: running task 52 on process 33 (Schedulers-primary-eA3Kzwpz); 10 workers total; 53 tasks left in task-pool. [ Info: running task 53 on process 36 (Schedulers-primary-eA3Kzwpz); 10 workers total; 52 tasks left in task-pool. [ Info: running task 54 on process 37 (Schedulers-primary-eA3Kzwpz); 10 workers total; 49 tasks left in task-pool. [ Info: running task 55 on process 38 (Schedulers-primary-eA3Kzwpz); 10 workers total; 49 tasks left in task-pool. [ Info: running task 56 on process 39 (Schedulers-primary-eA3Kzwpz); 10 workers total; 49 tasks left in task-pool. [ Info: running task 57 on process 34 (Schedulers-primary-eA3Kzwpz); 10 workers total; 48 tasks left in task-pool. [ Info: running task 58 on process 35 (Schedulers-primary-eA3Kzwpz); 10 workers total; 46 tasks left in task-pool. [ Info: running task 59 on process 40 (Schedulers-primary-eA3Kzwpz); 10 workers total; 46 tasks left in task-pool. [ Info: running task 60 on process 41 (Schedulers-primary-eA3Kzwpz); 10 workers total; 45 tasks left in task-pool. [ Info: running task 61 on process 32 (Schedulers-primary-eA3Kzwpz); 10 workers total; 44 tasks left in task-pool. [ Info: running task 62 on process 33 (Schedulers-primary-eA3Kzwpz); 10 workers total; 43 tasks left in task-pool. [ Info: running task 63 on process 36 (Schedulers-primary-eA3Kzwpz); 10 workers total; 42 tasks left in task-pool. [ Info: running task 64 on process 37 (Schedulers-primary-eA3Kzwpz); 10 workers total; 41 tasks left in task-pool. [ Info: running task 65 on process 38 (Schedulers-primary-eA3Kzwpz); 10 workers total; 40 tasks left in task-pool. [ Info: running task 66 on process 39 (Schedulers-primary-eA3Kzwpz); 10 workers total; 39 tasks left in task-pool. [ Info: running task 67 on process 34 (Schedulers-primary-eA3Kzwpz); 10 workers total; 38 tasks left in task-pool. [ Info: running task 68 on process 35 (Schedulers-primary-eA3Kzwpz); 10 workers total; 36 tasks left in task-pool. [ Info: running task 69 on process 40 (Schedulers-primary-eA3Kzwpz); 10 workers total; 36 tasks left in task-pool. [ Info: running task 70 on process 41 (Schedulers-primary-eA3Kzwpz); 10 workers total; 35 tasks left in task-pool. [ Info: running task 71 on process 32 (Schedulers-primary-eA3Kzwpz); 10 workers total; 34 tasks left in task-pool. [ Info: running task 72 on process 33 (Schedulers-primary-eA3Kzwpz); 10 workers total; 33 tasks left in task-pool. [ Info: running task 73 on process 36 (Schedulers-primary-eA3Kzwpz); 10 workers total; 31 tasks left in task-pool. [ Info: running task 74 on process 37 (Schedulers-primary-eA3Kzwpz); 10 workers total; 31 tasks left in task-pool. [ Info: running task 75 on process 38 (Schedulers-primary-eA3Kzwpz); 10 workers total; 30 tasks left in task-pool. [ Info: running task 76 on process 39 (Schedulers-primary-eA3Kzwpz); 10 workers total; 29 tasks left in task-pool. [ Info: running task 77 on process 34 (Schedulers-primary-eA3Kzwpz); 10 workers total; 28 tasks left in task-pool. [ Info: running task 78 on process 35 (Schedulers-primary-eA3Kzwpz); 10 workers total; 25 tasks left in task-pool. [ Info: running task 79 on process 40 (Schedulers-primary-eA3Kzwpz); 10 workers total; 25 tasks left in task-pool. [ Info: running task 80 on process 41 (Schedulers-primary-eA3Kzwpz); 10 workers total; 25 tasks left in task-pool. [ Info: running task 81 on process 32 (Schedulers-primary-eA3Kzwpz); 10 workers total; 24 tasks left in task-pool. [ Info: running task 82 on process 33 (Schedulers-primary-eA3Kzwpz); 10 workers total; 23 tasks left in task-pool. [ Info: running task 83 on process 36 (Schedulers-primary-eA3Kzwpz); 10 workers total; 21 tasks left in task-pool. [ Info: running task 84 on process 37 (Schedulers-primary-eA3Kzwpz); 10 workers total; 21 tasks left in task-pool. [ Info: running task 85 on process 38 (Schedulers-primary-eA3Kzwpz); 10 workers total; 20 tasks left in task-pool. [ Info: running task 86 on process 39 (Schedulers-primary-eA3Kzwpz); 10 workers total; 19 tasks left in task-pool. [ Info: running task 87 on process 34 (Schedulers-primary-eA3Kzwpz); 10 workers total; 18 tasks left in task-pool. [ Info: running task 88 on process 35 (Schedulers-primary-eA3Kzwpz); 10 workers total; 15 tasks left in task-pool. [ Info: running task 89 on process 40 (Schedulers-primary-eA3Kzwpz); 10 workers total; 15 tasks left in task-pool. [ Info: running task 90 on process 41 (Schedulers-primary-eA3Kzwpz); 10 workers total; 15 tasks left in task-pool. [ Info: running task 91 on process 32 (Schedulers-primary-eA3Kzwpz); 10 workers total; 14 tasks left in task-pool. [ Info: running task 92 on process 33 (Schedulers-primary-eA3Kzwpz); 10 workers total; 13 tasks left in task-pool. [ Info: running task 93 on process 36 (Schedulers-primary-eA3Kzwpz); 10 workers total; 11 tasks left in task-pool. [ Info: running task 94 on process 37 (Schedulers-primary-eA3Kzwpz); 10 workers total; 11 tasks left in task-pool. [ Info: running task 95 on process 38 (Schedulers-primary-eA3Kzwpz); 10 workers total; 10 tasks left in task-pool. [ Info: running task 96 on process 39 (Schedulers-primary-eA3Kzwpz); 10 workers total; 9 tasks left in task-pool. [ Info: running task 97 on process 34 (Schedulers-primary-eA3Kzwpz); 10 workers total; 8 tasks left in task-pool. [ Info: running task 98 on process 35 (Schedulers-primary-eA3Kzwpz); 10 workers total; 5 tasks left in task-pool. [ Info: running task 99 on process 40 (Schedulers-primary-eA3Kzwpz); 10 workers total; 5 tasks left in task-pool. [ Info: running task 100 on process 41 (Schedulers-primary-eA3Kzwpz); 10 workers total; 5 tasks left in task-pool. [ Info: running task 101 on process 32 (Schedulers-primary-eA3Kzwpz); 10 workers total; 4 tasks left in task-pool. [ Info: running task 102 on process 33 (Schedulers-primary-eA3Kzwpz); 10 workers total; 3 tasks left in task-pool. [ Info: running task 103 on process 36 (Schedulers-primary-eA3Kzwpz); 10 workers total; 1 tasks left in task-pool. [ Info: running task 104 on process 37 (Schedulers-primary-eA3Kzwpz); 10 workers total; 1 tasks left in task-pool. [ Info: running task 105 on process 38 (Schedulers-primary-eA3Kzwpz); 10 workers total; 0 tasks left in task-pool. Test Summary: | Pass Total Time pmap with shrinking cluster | 116 116 4m59.1s [ Info: running task 1 on process 42 (Schedulers-primary-eA3Kzwpz); 5 workers total; 99 tasks left in task-pool. [ Info: running task 2 on process 43 (Schedulers-primary-eA3Kzwpz); 5 workers total; 98 tasks left in task-pool. [ Info: running task 3 on process 42 (Schedulers-primary-eA3Kzwpz); 5 workers total; 97 tasks left in task-pool. [ Info: running task 4 on process 43 (Schedulers-primary-eA3Kzwpz); 5 workers total; 96 tasks left in task-pool. [ Info: running task 5 on process 42 (Schedulers-primary-eA3Kzwpz); 5 workers total; 95 tasks left in task-pool. [ Info: running task 6 on process 43 (Schedulers-primary-eA3Kzwpz); 5 workers total; 94 tasks left in task-pool. [ Info: running task 7 on process 42 (Schedulers-primary-eA3Kzwpz); 5 workers total; 93 tasks left in task-pool. [ Info: running task 8 on process 43 (Schedulers-primary-eA3Kzwpz); 5 workers total; 92 tasks left in task-pool. [ Info: running task 9 on process 42 (Schedulers-primary-eA3Kzwpz); 5 workers total; 91 tasks left in task-pool. [ Info: running task 10 on process 43 (Schedulers-primary-eA3Kzwpz); 5 workers total; 90 tasks left in task-pool. [ Info: running task 11 on process 42 (Schedulers-primary-eA3Kzwpz); 10 workers total; 89 tasks left in task-pool. [ Info: running task 12 on process 43 (Schedulers-primary-eA3Kzwpz); 10 workers total; 88 tasks left in task-pool. [ Info: running task 13 on process 42 (Schedulers-primary-eA3Kzwpz); 10 workers total; 87 tasks left in task-pool. [ Info: running task 14 on process 43 (Schedulers-primary-eA3Kzwpz); 10 workers total; 86 tasks left in task-pool. [ Info: running task 15 on process 44 (Schedulers-primary-eA3Kzwpz); 10 workers total; 85 tasks left in task-pool. [ Info: running task 16 on process 45 (Schedulers-primary-eA3Kzwpz); 10 workers total; 83 tasks left in task-pool. [ Info: running task 17 on process 46 (Schedulers-primary-eA3Kzwpz); 10 workers total; 83 tasks left in task-pool. [ Info: running task 18 on process 42 (Schedulers-primary-eA3Kzwpz); 10 workers total; 82 tasks left in task-pool. [ Info: running task 19 on process 43 (Schedulers-primary-eA3Kzwpz); 10 workers total; 81 tasks left in task-pool. [ Info: running task 20 on process 44 (Schedulers-primary-eA3Kzwpz); 10 workers total; 80 tasks left in task-pool. [ Info: running task 21 on process 45 (Schedulers-primary-eA3Kzwpz); 10 workers total; 78 tasks left in task-pool. [ Info: running task 22 on process 46 (Schedulers-primary-eA3Kzwpz); 10 workers total; 78 tasks left in task-pool. [ Info: running task 23 on process 42 (Schedulers-primary-eA3Kzwpz); 10 workers total; 77 tasks left in task-pool. [ Info: running task 24 on process 43 (Schedulers-primary-eA3Kzwpz); 10 workers total; 76 tasks left in task-pool. ┌ Warning: addprocs/rmprocs taking longer than expected, cancelling. └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:776 ┌ Warning: problem adding or removing processes └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:737 ┌ Warning: TaskFailedException │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1211 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#197 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch(t::Task) │ @ Base ./task.jl:525 │ [4] loop(eloop::Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, journal::Dict{String, Any}, journal_task_callback::Function, tsk_map::Task, tsk_reduce::Task) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:735 │ [5] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1046 │ [6] epmap │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1041 [inlined] │ [7] (::var"#11#12"{SchedulerOptions{String}, String})() │ @ Main ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:166 │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1211 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#197 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 [ Info: running task 25 on process 44 (Schedulers-primary-eA3Kzwpz); 10 workers total; 75 tasks left in task-pool. [ Info: running task 26 on process 45 (Schedulers-primary-eA3Kzwpz); 10 workers total; 73 tasks left in task-pool. [ Info: running task 27 on process 46 (Schedulers-primary-eA3Kzwpz); 10 workers total; 73 tasks left in task-pool. ┌ Warning: problem adding or removing processes └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:737 ┌ Warning: TaskFailedException │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1211 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#197 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch(t::Task) │ @ Base ./task.jl:525 │ [4] loop(eloop::Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, journal::Dict{String, Any}, journal_task_callback::Function, tsk_map::Task, tsk_reduce::Task) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:735 │ [5] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1046 │ [6] epmap │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1041 [inlined] │ [7] (::var"#11#12"{SchedulerOptions{String}, String})() │ @ Main ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:166 │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1211 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#197 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 [ Info: running task 28 on process 42 (Schedulers-primary-eA3Kzwpz); 10 workers total; 72 tasks left in task-pool. [ Info: running task 29 on process 43 (Schedulers-primary-eA3Kzwpz); 10 workers total; 71 tasks left in task-pool. ┌ Warning: problem adding or removing processes └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:737 ┌ Warning: TaskFailedException │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1211 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#197 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch(t::Task) │ @ Base ./task.jl:525 │ [4] loop(eloop::Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, journal::Dict{String, Any}, journal_task_callback::Function, tsk_map::Task, tsk_reduce::Task) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:735 │ [5] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1046 │ [6] epmap │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1041 [inlined] │ [7] (::var"#11#12"{SchedulerOptions{String}, String})() │ @ Main ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:166 │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1211 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#197 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 ┌ Warning: problem adding or removing processes └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:737 ┌ Warning: TaskFailedException │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1211 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#197 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch(t::Task) │ @ Base ./task.jl:525 │ [4] loop(eloop::Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, journal::Dict{String, Any}, journal_task_callback::Function, tsk_map::Task, tsk_reduce::Task) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:735 │ [5] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1046 │ [6] epmap │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1041 [inlined] │ [7] (::var"#11#12"{SchedulerOptions{String}, String})() │ @ Main ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:166 │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1211 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#197 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 [ Info: running task 30 on process 44 (Schedulers-primary-eA3Kzwpz); 10 workers total; 70 tasks left in task-pool. [ Info: running task 31 on process 45 (Schedulers-primary-eA3Kzwpz); 10 workers total; 68 tasks left in task-pool. [ Info: running task 32 on process 46 (Schedulers-primary-eA3Kzwpz); 10 workers total; 68 tasks left in task-pool. ┌ Warning: problem adding or removing processes └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:737 ┌ Warning: TaskFailedException │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1211 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#197 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch(t::Task) │ @ Base ./task.jl:525 │ [4] loop(eloop::Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, journal::Dict{String, Any}, journal_task_callback::Function, tsk_map::Task, tsk_reduce::Task) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:735 │ [5] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1046 │ [6] epmap │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1041 [inlined] │ [7] (::var"#11#12"{SchedulerOptions{String}, String})() │ @ Main ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:166 │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1211 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#197 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 [ Info: running task 33 on process 42 (Schedulers-primary-eA3Kzwpz); 10 workers total; 67 tasks left in task-pool. [ Info: running task 34 on process 43 (Schedulers-primary-eA3Kzwpz); 10 workers total; 66 tasks left in task-pool. ┌ Warning: problem adding or removing processes └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:737 ┌ Warning: TaskFailedException │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1211 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#197 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch(t::Task) │ @ Base ./task.jl:525 │ [4] loop(eloop::Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, journal::Dict{String, Any}, journal_task_callback::Function, tsk_map::Task, tsk_reduce::Task) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:735 │ [5] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1046 │ [6] epmap │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1041 [inlined] │ [7] (::var"#11#12"{SchedulerOptions{String}, String})() │ @ Main ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:166 │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1211 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#197 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 ┌ Warning: problem adding or removing processes └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:737 ┌ Warning: TaskFailedException │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1211 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#197 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch(t::Task) │ @ Base ./task.jl:525 │ [4] loop(eloop::Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, journal::Dict{String, Any}, journal_task_callback::Function, tsk_map::Task, tsk_reduce::Task) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:735 │ [5] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1046 │ [6] epmap │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1041 [inlined] │ [7] (::var"#11#12"{SchedulerOptions{String}, String})() │ @ Main ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:166 │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1211 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#197 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 [ Info: running task 35 on process 44 (Schedulers-primary-eA3Kzwpz); 10 workers total; 65 tasks left in task-pool. [ Info: running task 36 on process 45 (Schedulers-primary-eA3Kzwpz); 10 workers total; 63 tasks left in task-pool. [ Info: running task 37 on process 46 (Schedulers-primary-eA3Kzwpz); 10 workers total; 63 tasks left in task-pool. ┌ Warning: problem adding or removing processes └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:737 ┌ Warning: TaskFailedException │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1211 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#197 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch(t::Task) │ @ Base ./task.jl:525 │ [4] loop(eloop::Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, journal::Dict{String, Any}, journal_task_callback::Function, tsk_map::Task, tsk_reduce::Task) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:735 │ [5] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1046 │ [6] epmap │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1041 [inlined] │ [7] (::var"#11#12"{SchedulerOptions{String}, String})() │ @ Main ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:166 │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1211 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#197 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 [ Info: running task 38 on process 42 (Schedulers-primary-eA3Kzwpz); 10 workers total; 62 tasks left in task-pool. [ Info: running task 39 on process 43 (Schedulers-primary-eA3Kzwpz); 10 workers total; 61 tasks left in task-pool. ┌ Warning: problem adding or removing processes └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:737 ┌ Warning: TaskFailedException │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1211 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#197 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch(t::Task) │ @ Base ./task.jl:525 │ [4] loop(eloop::Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, journal::Dict{String, Any}, journal_task_callback::Function, tsk_map::Task, tsk_reduce::Task) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:735 │ [5] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1046 │ [6] epmap │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1041 [inlined] │ [7] (::var"#11#12"{SchedulerOptions{String}, String})() │ @ Main ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:166 │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1211 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#197 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 ┌ Warning: problem adding or removing processes └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:737 ┌ Warning: TaskFailedException │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1211 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#197 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch(t::Task) │ @ Base ./task.jl:525 │ [4] loop(eloop::Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, journal::Dict{String, Any}, journal_task_callback::Function, tsk_map::Task, tsk_reduce::Task) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:735 │ [5] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1046 │ [6] epmap │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1041 [inlined] │ [7] (::var"#11#12"{SchedulerOptions{String}, String})() │ @ Main ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:166 │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1211 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#197 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 [ Info: running task 40 on process 44 (Schedulers-primary-eA3Kzwpz); 10 workers total; 60 tasks left in task-pool. [ Info: running task 41 on process 45 (Schedulers-primary-eA3Kzwpz); 10 workers total; 58 tasks left in task-pool. [ Info: running task 42 on process 46 (Schedulers-primary-eA3Kzwpz); 10 workers total; 58 tasks left in task-pool. ┌ Warning: problem adding or removing processes └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:737 ┌ Warning: TaskFailedException │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1211 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#197 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch(t::Task) │ @ Base ./task.jl:525 │ [4] loop(eloop::Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, journal::Dict{String, Any}, journal_task_callback::Function, tsk_map::Task, tsk_reduce::Task) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:735 │ [5] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1046 │ [6] epmap │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1041 [inlined] │ [7] (::var"#11#12"{SchedulerOptions{String}, String})() │ @ Main ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:166 │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1211 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#197 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 [ Info: running task 43 on process 47 (Schedulers-primary-eA3Kzwpz); 10 workers total; 57 tasks left in task-pool. [ Info: running task 44 on process 48 (Schedulers-primary-eA3Kzwpz); 10 workers total; 56 tasks left in task-pool. ┌ Warning: problem adding or removing processes └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:737 ┌ Warning: TaskFailedException │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1211 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#197 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch(t::Task) │ @ Base ./task.jl:525 │ [4] loop(eloop::Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, journal::Dict{String, Any}, journal_task_callback::Function, tsk_map::Task, tsk_reduce::Task) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:735 │ [5] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1046 │ [6] epmap │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1041 [inlined] │ [7] (::var"#11#12"{SchedulerOptions{String}, String})() │ @ Main ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:166 │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1211 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#197 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 From worker 50: ErrorException("Process(1) - Invalid connection credentials sent by remote.")CapturedException(ErrorException("Process(1) - Invalid connection credentials sent by remote."), Any[(error(s::String) at error.jl:44, 1), (process_hdr(s::Sockets.TCPSocket, validate_cookie::Bool) at process_messages.jl:266, 1), (message_handler_loop(r_stream::Sockets.TCPSocket, w_stream::Sockets.TCPSocket, incoming::Bool) at process_messages.jl:158, 1), (process_tcp_streams(r_stream::Sockets.TCPSocket, w_stream::Sockets.TCPSocket, incoming::Bool) at process_messages.jl:133, 1), ((::Distributed.var"#process_messages##0#process_messages##1"{Sockets.TCPSocket, Sockets.TCPSocket, Bool})() at process_messages.jl:121, 1)]) From worker 50: Process(1) - Unknown remote, closing connection. Worker 50 terminated. From worker 49: ErrorException("Process(1) - Invalid connection credentials sent by remote.")CapturedException(ErrorException("Process(1) - Invalid connection credentials sent by remote."), Any[(error(s::String) at error.jl:44, 1), (process_hdr(s::Sockets.TCPSocket, validate_cookie::Bool) at process_messages.jl:266, 1), (message_handler_loop(r_stream::Sockets.TCPSocket, w_stream::Sockets.TCPSocket, incoming::Bool) at process_messages.jl:158, 1), (process_tcp_streams(r_stream::Sockets.TCPSocket, w_stream::Sockets.TCPSocket, incoming::Bool) at process_messages.jl:133, 1), ((::Distributed.var"#process_messages##0#process_messages##1"{Sockets.TCPSocket, Sockets.TCPSocket, Bool})() at process_messages.jl:121, 1)]) [ Info: running task 45 on process 42 (Schedulers-primary-eA3Kzwpz); 7 workers total; 55 tasks left in task-pool. From worker 49: Process(1) - Unknown remote, closing connection.[ Info: running task 46 on process 43 (Schedulers-primary-eA3Kzwpz); 7 workers total; 54 tasks left in task-pool. UNHANDLED TASK ERROR: Version read failed. Connection closed by peer. Stacktrace: [1] error(s::String) @ Base ./error.jl:44 [2] process_hdr(s::Sockets.TCPSocket, validate_cookie::Bool) @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:276 [3] message_handler_loop(r_stream::Sockets.TCPSocket, w_stream::Sockets.TCPSocket, incoming::Bool) @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:158 [4] process_tcp_streams(r_stream::Sockets.TCPSocket, w_stream::Sockets.TCPSocket, incoming::Bool) @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:133 [5] (::Distributed.var"#process_messages##0#process_messages##1"{Sockets.TCPSocket, Sockets.TCPSocket, Bool})() @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:121 From worker 51: ErrorException("Process(1) - Invalid connection credentials sent by remote.")CapturedException(ErrorException("Process(1) - Invalid connection credentials sent by remote."), Any[(error(s::String) at error.jl:44, 1), (process_hdr(s::Sockets.TCPSocket, validate_cookie::Bool) at process_messages.jl:266, 1), (message_handler_loop(r_stream::Sockets.TCPSocket, w_stream::Sockets.TCPSocket, incoming::Bool) at process_messages.jl:158, 1), (process_tcp_streams(r_stream::Sockets.TCPSocket, w_stream::Sockets.TCPSocket, incoming::Bool) at process_messages.jl:133, 1), ((::Distributed.var"#process_messages##0#process_messages##1"{Sockets.TCPSocket, Sockets.TCPSocket, Bool})() at process_messages.jl:121, 1)]) ┌ Warning: problem adding or removing processes └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:737 From worker 51: Process(1) - Unknown remote, closing connection.┌ Warning: TaskFailedException │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1211 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#197 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch(t::Task) │ @ Base ./task.jl:525 │ [4] loop(eloop::Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, journal::Dict{String, Any}, journal_task_callback::Function, tsk_map::Task, tsk_reduce::Task) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:735 │ [5] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1046 │ [6] epmap │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1041 [inlined] │ [7] (::var"#11#12"{SchedulerOptions{String}, String})() │ @ Main ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:166 │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1211 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#197 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 Worker 51 terminated. Worker 49 terminated.UNHANDLED TASK ERROR: IOError: read: connection reset by peer (ECONNRESET) Stacktrace: [1] wait_readnb(x::Sockets.TCPSocket, nb::Int64) @ Base ./stream.jl:410 [2] (::Base.var"#wait_locked#readbytes!##0")(s::Sockets.TCPSocket, buf::IOBuffer, nb::Int64) @ Base ./stream.jl:934 [3] readbytes!(s::Sockets.TCPSocket, a::Vector{UInt8}, nb::Int64) @ Base ./stream.jl:940 [4] read @ ./io.jl:1184 [inlined] [5] process_hdr(s::Sockets.TCPSocket, validate_cookie::Bool) @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:274 [6] message_handler_loop(r_stream::Sockets.TCPSocket, w_stream::Sockets.TCPSocket, incoming::Bool) @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:158 [7] process_tcp_streams(r_stream::Sockets.TCPSocket, w_stream::Sockets.TCPSocket, incoming::Bool) @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:133 [8] (::Distributed.var"#process_messages##0#process_messages##1"{Sockets.TCPSocket, Sockets.TCPSocket, Bool})() @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:121 UNHANDLED TASK ERROR: IOError: read: connection reset by peer (ECONNRESET) Stacktrace: [1] wait_readnb(x::Sockets.TCPSocket, nb::Int64) @ Base ./stream.jl:410 [2] (::Base.var"#wait_locked#readbytes!##0")(s::Sockets.TCPSocket, buf::IOBuffer, nb::Int64) @ Base ./stream.jl:934 [3] readbytes!(s::Sockets.TCPSocket, a::Vector{UInt8}, nb::Int64) @ Base ./stream.jl:940 [4] read @ ./io.jl:1184 [inlined] [5] process_hdr(s::Sockets.TCPSocket, validate_cookie::Bool) @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:274 [6] message_handler_loop(r_stream::Sockets.TCPSocket, w_stream::Sockets.TCPSocket, incoming::Bool) @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:158 [7] process_tcp_streams(r_stream::Sockets.TCPSocket, w_stream::Sockets.TCPSocket, incoming::Bool) @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:133 [8] (::Distributed.var"#process_messages##0#process_messages##1"{Sockets.TCPSocket, Sockets.TCPSocket, Bool})() @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:121 [ Info: running task 47 on process 44 (Schedulers-primary-eA3Kzwpz); 7 workers total; 53 tasks left in task-pool. [ Info: running task 48 on process 45 (Schedulers-primary-eA3Kzwpz); 7 workers total; 51 tasks left in task-pool. [ Info: running task 49 on process 46 (Schedulers-primary-eA3Kzwpz); 7 workers total; 51 tasks left in task-pool. [ Info: running task 50 on process 47 (Schedulers-primary-eA3Kzwpz); 7 workers total; 50 tasks left in task-pool. [ Info: running task 51 on process 48 (Schedulers-primary-eA3Kzwpz); 7 workers total; 49 tasks left in task-pool. [ Info: running task 52 on process 42 (Schedulers-primary-eA3Kzwpz); 7 workers total; 48 tasks left in task-pool. [ Info: running task 53 on process 43 (Schedulers-primary-eA3Kzwpz); 7 workers total; 47 tasks left in task-pool. [ Info: running task 54 on process 44 (Schedulers-primary-eA3Kzwpz); 10 workers total; 46 tasks left in task-pool. [ Info: running task 55 on process 45 (Schedulers-primary-eA3Kzwpz); 10 workers total; 44 tasks left in task-pool. [ Info: running task 56 on process 46 (Schedulers-primary-eA3Kzwpz); 10 workers total; 44 tasks left in task-pool. [ Info: running task 57 on process 47 (Schedulers-primary-eA3Kzwpz); 10 workers total; 43 tasks left in task-pool. [ Info: running task 58 on process 48 (Schedulers-primary-eA3Kzwpz); 10 workers total; 42 tasks left in task-pool. [ Info: running task 59 on process 42 (Schedulers-primary-eA3Kzwpz); 10 workers total; 41 tasks left in task-pool. [ Info: running task 60 on process 43 (Schedulers-primary-eA3Kzwpz); 10 workers total; 40 tasks left in task-pool. [ Info: running task 61 on process 44 (Schedulers-primary-eA3Kzwpz); 10 workers total; 39 tasks left in task-pool. [ Info: running task 62 on process 45 (Schedulers-primary-eA3Kzwpz); 10 workers total; 37 tasks left in task-pool. [ Info: running task 63 on process 46 (Schedulers-primary-eA3Kzwpz); 10 workers total; 37 tasks left in task-pool. [ Info: running task 64 on process 47 (Schedulers-primary-eA3Kzwpz); 10 workers total; 36 tasks left in task-pool. [ Info: running task 65 on process 48 (Schedulers-primary-eA3Kzwpz); 10 workers total; 35 tasks left in task-pool. [ Info: running task 66 on process 42 (Schedulers-primary-eA3Kzwpz); 10 workers total; 33 tasks left in task-pool. [ Info: running task 67 on process 43 (Schedulers-primary-eA3Kzwpz); 10 workers total; 33 tasks left in task-pool. From worker 51: Master process (id 1) could not connect within 120.0 seconds. From worker 51: exiting. From worker 50: Master process (id 1) could not connect within 120.0 seconds. From worker 50: exiting. From worker 49: Master process (id 1) could not connect within 120.0 seconds. From worker 49: exiting. [ Info: running task 68 on process 44 (Schedulers-primary-eA3Kzwpz); 10 workers total; 32 tasks left in task-pool. [ Info: running task 69 on process 46 (Schedulers-primary-eA3Kzwpz); 10 workers total; 31 tasks left in task-pool. [ Info: running task 70 on process 45 (Schedulers-primary-eA3Kzwpz); 10 workers total; 30 tasks left in task-pool. [ Info: running task 71 on process 47 (Schedulers-primary-eA3Kzwpz); 10 workers total; 29 tasks left in task-pool. [ Info: running task 72 on process 48 (Schedulers-primary-eA3Kzwpz); 10 workers total; 28 tasks left in task-pool. [ Info: running task 73 on process 42 (Schedulers-primary-eA3Kzwpz); 10 workers total; 26 tasks left in task-pool. [ Info: running task 74 on process 43 (Schedulers-primary-eA3Kzwpz); 10 workers total; 26 tasks left in task-pool. [ Info: running task 75 on process 44 (Schedulers-primary-eA3Kzwpz); 10 workers total; 25 tasks left in task-pool. [ Info: running task 76 on process 46 (Schedulers-primary-eA3Kzwpz); 10 workers total; 24 tasks left in task-pool. [ Info: running task 77 on process 45 (Schedulers-primary-eA3Kzwpz); 10 workers total; 23 tasks left in task-pool. [ Info: running task 78 on process 47 (Schedulers-primary-eA3Kzwpz); 10 workers total; 22 tasks left in task-pool. [ Info: running task 79 on process 48 (Schedulers-primary-eA3Kzwpz); 10 workers total; 21 tasks left in task-pool. [ Info: running task 80 on process 42 (Schedulers-primary-eA3Kzwpz); 10 workers total; 19 tasks left in task-pool. [ Info: running task 81 on process 43 (Schedulers-primary-eA3Kzwpz); 10 workers total; 19 tasks left in task-pool. [ Info: running task 82 on process 44 (Schedulers-primary-eA3Kzwpz); 10 workers total; 18 tasks left in task-pool. [ Info: running task 83 on process 46 (Schedulers-primary-eA3Kzwpz); 10 workers total; 17 tasks left in task-pool. [ Info: running task 84 on process 45 (Schedulers-primary-eA3Kzwpz); 10 workers total; 16 tasks left in task-pool. [ Info: running task 85 on process 52 (Schedulers-primary-eA3Kzwpz); 10 workers total; 15 tasks left in task-pool. [ Info: running task 86 on process 53 (Schedulers-primary-eA3Kzwpz); 10 workers total; 13 tasks left in task-pool. [ Info: running task 87 on process 54 (Schedulers-primary-eA3Kzwpz); 10 workers total; 13 tasks left in task-pool. [ Info: running task 88 on process 47 (Schedulers-primary-eA3Kzwpz); 10 workers total; 12 tasks left in task-pool. [ Info: running task 89 on process 48 (Schedulers-primary-eA3Kzwpz); 10 workers total; 11 tasks left in task-pool. [ Info: running task 90 on process 42 (Schedulers-primary-eA3Kzwpz); 10 workers total; 9 tasks left in task-pool. [ Info: running task 91 on process 43 (Schedulers-primary-eA3Kzwpz); 10 workers total; 9 tasks left in task-pool. [ Info: running task 92 on process 44 (Schedulers-primary-eA3Kzwpz); 10 workers total; 8 tasks left in task-pool. [ Info: running task 93 on process 46 (Schedulers-primary-eA3Kzwpz); 10 workers total; 7 tasks left in task-pool. [ Info: running task 94 on process 45 (Schedulers-primary-eA3Kzwpz); 10 workers total; 6 tasks left in task-pool. [ Info: running task 95 on process 52 (Schedulers-primary-eA3Kzwpz); 10 workers total; 5 tasks left in task-pool. [ Info: running task 96 on process 53 (Schedulers-primary-eA3Kzwpz); 10 workers total; 3 tasks left in task-pool. [ Info: running task 97 on process 54 (Schedulers-primary-eA3Kzwpz); 10 workers total; 3 tasks left in task-pool. [ Info: running task 98 on process 47 (Schedulers-primary-eA3Kzwpz); 10 workers total; 2 tasks left in task-pool. [ Info: running task 99 on process 48 (Schedulers-primary-eA3Kzwpz); 10 workers total; 1 tasks left in task-pool. [ Info: running task 100 on process 42 (Schedulers-primary-eA3Kzwpz); 10 workers total; 0 tasks left in task-pool. Test Summary: | Pass Total Time pmap with interactive growing cluster | 111 111 4m45.7s [ Info: nworkers()=1, i=1 ┌ Warning: problem adding or removing processes └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:737 ┌ Warning: TaskFailedException │ │ nested task error: MethodError: no method matching addprocs(::Int64, ::String) │ The function `addprocs` exists, but no method is defined for this combination of argument types. │ │ Closest candidates are: │ addprocs(::Integer; restrict, kwargs...) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ addprocs(; ...) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ addprocs(!Matched::ClusterManager; kwargs...) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 │ ... │ │ Stacktrace: │ [1] (::var"#myaddprocs#myaddprocs##0")(n::Int64) │ @ Main ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:208 │ [2] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{var"#myaddprocs#myaddprocs##0", Schedulers.var"#14#15", typeof(nworkers), Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#94#95"{Int64}, Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch(t::Task) │ @ Base ./task.jl:525 │ [4] loop(eloop::Schedulers.ElasticLoop{var"#myaddprocs#myaddprocs##0", Schedulers.var"#14#15", typeof(nworkers), Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#94#95"{Int64}, Int64, Nothing}, journal::Dict{String, Any}, journal_task_callback::Function, tsk_map::Task, tsk_reduce::Task) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:735 │ [5] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1046 │ [6] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1041 │ [7] top-level scope │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:195 │ [8] macro expansion │ @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1855 [inlined] │ [9] macro expansion │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:218 [inlined] │ [10] include(mapexpr::Function, mod::Module, _path::String) │ @ Base ./Base.jl:310 │ [11] top-level scope │ @ none:6 │ [12] eval(m::Module, e::Any) │ @ Core ./boot.jl:489 │ [13] exec_options(opts::Base.JLOptions) │ @ Base ./client.jl:290 │ [14] _start() │ @ Base ./client.jl:557 │ │ nested task error: MethodError: no method matching addprocs(::Int64, ::String) │ The function `addprocs` exists, but no method is defined for this combination of argument types. │ │ Closest candidates are: │ addprocs(::Integer; restrict, kwargs...) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ addprocs(; ...) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ addprocs(!Matched::ClusterManager; kwargs...) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 │ ... │ │ Stacktrace: │ [1] (::var"#myaddprocs#myaddprocs##0")(n::Int64) │ @ Main ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:208 │ [2] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{var"#myaddprocs#myaddprocs##0", Schedulers.var"#14#15", typeof(nworkers), Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#94#95"{Int64}, Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 [ Info: nworkers()=1, i=2 [ Info: nworkers()=1, i=3 [ Info: running task 1 on process 55 (Schedulers-primary-eA3Kzwpz); 2 workers total; 99 tasks left in task-pool. [ Info: running task 2 on process 55 (Schedulers-primary-eA3Kzwpz); 2 workers total; 98 tasks left in task-pool. [ Info: running task 3 on process 55 (Schedulers-primary-eA3Kzwpz); 2 workers total; 97 tasks left in task-pool. [ Info: nworkers()=2, i=4 [ Info: running task 4 on process 55 (Schedulers-primary-eA3Kzwpz); 2 workers total; 96 tasks left in task-pool. [ Info: running task 5 on process 55 (Schedulers-primary-eA3Kzwpz); 2 workers total; 95 tasks left in task-pool. [ Info: running task 6 on process 55 (Schedulers-primary-eA3Kzwpz); 2 workers total; 94 tasks left in task-pool. [ Info: running task 7 on process 55 (Schedulers-primary-eA3Kzwpz); 2 workers total; 93 tasks left in task-pool. [ Info: running task 8 on process 55 (Schedulers-primary-eA3Kzwpz); 2 workers total; 92 tasks left in task-pool. [ Info: running task 9 on process 55 (Schedulers-primary-eA3Kzwpz); 3 workers total; 91 tasks left in task-pool. [ Info: running task 10 on process 56 (Schedulers-primary-eA3Kzwpz); 3 workers total; 90 tasks left in task-pool. [ Info: running task 11 on process 55 (Schedulers-primary-eA3Kzwpz); 3 workers total; 89 tasks left in task-pool. [ Info: running task 12 on process 56 (Schedulers-primary-eA3Kzwpz); 3 workers total; 88 tasks left in task-pool. [ Info: running task 13 on process 55 (Schedulers-primary-eA3Kzwpz); 3 workers total; 87 tasks left in task-pool. [ Info: running task 14 on process 56 (Schedulers-primary-eA3Kzwpz); 3 workers total; 86 tasks left in task-pool. [ Info: running task 15 on process 55 (Schedulers-primary-eA3Kzwpz); 3 workers total; 85 tasks left in task-pool. [ Info: nworkers()=3, i=5 [ Info: running task 16 on process 56 (Schedulers-primary-eA3Kzwpz); 3 workers total; 84 tasks left in task-pool. [ Info: running task 17 on process 55 (Schedulers-primary-eA3Kzwpz); 3 workers total; 83 tasks left in task-pool. [ Info: running task 18 on process 56 (Schedulers-primary-eA3Kzwpz); 3 workers total; 82 tasks left in task-pool. [ Info: running task 19 on process 55 (Schedulers-primary-eA3Kzwpz); 3 workers total; 81 tasks left in task-pool. [ Info: running task 20 on process 56 (Schedulers-primary-eA3Kzwpz); 3 workers total; 80 tasks left in task-pool. [ Info: running task 21 on process 55 (Schedulers-primary-eA3Kzwpz); 3 workers total; 79 tasks left in task-pool. [ Info: running task 22 on process 56 (Schedulers-primary-eA3Kzwpz); 3 workers total; 78 tasks left in task-pool. [ Info: running task 23 on process 57 (Schedulers-primary-eA3Kzwpz); 3 workers total; 77 tasks left in task-pool. [ Info: running task 24 on process 55 (Schedulers-primary-eA3Kzwpz); 3 workers total; 76 tasks left in task-pool. [ Info: running task 25 on process 56 (Schedulers-primary-eA3Kzwpz); 3 workers total; 75 tasks left in task-pool. [ Info: running task 26 on process 57 (Schedulers-primary-eA3Kzwpz); 3 workers total; 74 tasks left in task-pool. [ Info: running task 27 on process 55 (Schedulers-primary-eA3Kzwpz); 3 workers total; 73 tasks left in task-pool. [ Info: running task 28 on process 56 (Schedulers-primary-eA3Kzwpz); 3 workers total; 72 tasks left in task-pool. [ Info: running task 29 on process 57 (Schedulers-primary-eA3Kzwpz); 3 workers total; 71 tasks left in task-pool. [ Info: running task 30 on process 55 (Schedulers-primary-eA3Kzwpz); 3 workers total; 70 tasks left in task-pool. [ Info: running task 31 on process 56 (Schedulers-primary-eA3Kzwpz); 3 workers total; 69 tasks left in task-pool. [ Info: running task 32 on process 57 (Schedulers-primary-eA3Kzwpz); 3 workers total; 68 tasks left in task-pool. [ Info: running task 33 on process 55 (Schedulers-primary-eA3Kzwpz); 3 workers total; 67 tasks left in task-pool. [ Info: running task 34 on process 56 (Schedulers-primary-eA3Kzwpz); 3 workers total; 66 tasks left in task-pool. [ Info: running task 35 on process 57 (Schedulers-primary-eA3Kzwpz); 3 workers total; 65 tasks left in task-pool. [ Info: running task 36 on process 55 (Schedulers-primary-eA3Kzwpz); 3 workers total; 64 tasks left in task-pool. [ Info: running task 37 on process 56 (Schedulers-primary-eA3Kzwpz); 3 workers total; 63 tasks left in task-pool. [ Info: running task 38 on process 57 (Schedulers-primary-eA3Kzwpz); 3 workers total; 62 tasks left in task-pool. [ Info: running task 39 on process 55 (Schedulers-primary-eA3Kzwpz); 3 workers total; 61 tasks left in task-pool. [ Info: running task 40 on process 56 (Schedulers-primary-eA3Kzwpz); 3 workers total; 60 tasks left in task-pool. [ Info: running task 41 on process 57 (Schedulers-primary-eA3Kzwpz); 3 workers total; 59 tasks left in task-pool. [ Info: running task 42 on process 55 (Schedulers-primary-eA3Kzwpz); 3 workers total; 58 tasks left in task-pool. [ Info: running task 43 on process 56 (Schedulers-primary-eA3Kzwpz); 3 workers total; 57 tasks left in task-pool. [ Info: running task 44 on process 57 (Schedulers-primary-eA3Kzwpz); 3 workers total; 56 tasks left in task-pool. [ Info: running task 45 on process 55 (Schedulers-primary-eA3Kzwpz); 3 workers total; 55 tasks left in task-pool. [ Info: running task 46 on process 56 (Schedulers-primary-eA3Kzwpz); 3 workers total; 54 tasks left in task-pool. [ Info: running task 47 on process 57 (Schedulers-primary-eA3Kzwpz); 3 workers total; 53 tasks left in task-pool. [ Info: running task 48 on process 55 (Schedulers-primary-eA3Kzwpz); 3 workers total; 52 tasks left in task-pool. [ Info: running task 49 on process 56 (Schedulers-primary-eA3Kzwpz); 3 workers total; 51 tasks left in task-pool. [ Info: running task 50 on process 57 (Schedulers-primary-eA3Kzwpz); 3 workers total; 50 tasks left in task-pool. [ Info: running task 51 on process 55 (Schedulers-primary-eA3Kzwpz); 3 workers total; 49 tasks left in task-pool. [ Info: running task 52 on process 56 (Schedulers-primary-eA3Kzwpz); 3 workers total; 48 tasks left in task-pool. [ Info: running task 53 on process 57 (Schedulers-primary-eA3Kzwpz); 3 workers total; 47 tasks left in task-pool. [ Info: running task 54 on process 55 (Schedulers-primary-eA3Kzwpz); 3 workers total; 46 tasks left in task-pool. [ Info: running task 55 on process 56 (Schedulers-primary-eA3Kzwpz); 3 workers total; 45 tasks left in task-pool. [ Info: running task 56 on process 57 (Schedulers-primary-eA3Kzwpz); 3 workers total; 44 tasks left in task-pool. [ Info: running task 57 on process 55 (Schedulers-primary-eA3Kzwpz); 3 workers total; 43 tasks left in task-pool. [ Info: running task 58 on process 56 (Schedulers-primary-eA3Kzwpz); 3 workers total; 42 tasks left in task-pool. [ Info: running task 59 on process 57 (Schedulers-primary-eA3Kzwpz); 3 workers total; 41 tasks left in task-pool. [ Info: running task 60 on process 55 (Schedulers-primary-eA3Kzwpz); 3 workers total; 40 tasks left in task-pool. [ Info: running task 61 on process 56 (Schedulers-primary-eA3Kzwpz); 3 workers total; 39 tasks left in task-pool. [ Info: running task 62 on process 57 (Schedulers-primary-eA3Kzwpz); 3 workers total; 38 tasks left in task-pool. [ Info: running task 63 on process 55 (Schedulers-primary-eA3Kzwpz); 3 workers total; 37 tasks left in task-pool. [ Info: running task 64 on process 56 (Schedulers-primary-eA3Kzwpz); 3 workers total; 36 tasks left in task-pool. [ Info: running task 65 on process 57 (Schedulers-primary-eA3Kzwpz); 3 workers total; 35 tasks left in task-pool. [ Info: running task 66 on process 55 (Schedulers-primary-eA3Kzwpz); 3 workers total; 34 tasks left in task-pool. [ Info: running task 67 on process 56 (Schedulers-primary-eA3Kzwpz); 3 workers total; 33 tasks left in task-pool. [ Info: running task 68 on process 57 (Schedulers-primary-eA3Kzwpz); 3 workers total; 32 tasks left in task-pool. [ Info: running task 69 on process 55 (Schedulers-primary-eA3Kzwpz); 3 workers total; 31 tasks left in task-pool. [ Info: running task 70 on process 56 (Schedulers-primary-eA3Kzwpz); 3 workers total; 30 tasks left in task-pool. [ Info: running task 71 on process 57 (Schedulers-primary-eA3Kzwpz); 3 workers total; 29 tasks left in task-pool. [ Info: running task 72 on process 55 (Schedulers-primary-eA3Kzwpz); 3 workers total; 28 tasks left in task-pool. [ Info: running task 73 on process 56 (Schedulers-primary-eA3Kzwpz); 3 workers total; 27 tasks left in task-pool. [ Info: running task 74 on process 57 (Schedulers-primary-eA3Kzwpz); 3 workers total; 26 tasks left in task-pool. [ Info: running task 75 on process 55 (Schedulers-primary-eA3Kzwpz); 3 workers total; 25 tasks left in task-pool. [ Info: running task 76 on process 56 (Schedulers-primary-eA3Kzwpz); 3 workers total; 24 tasks left in task-pool. [ Info: running task 77 on process 57 (Schedulers-primary-eA3Kzwpz); 3 workers total; 23 tasks left in task-pool. [ Info: running task 78 on process 55 (Schedulers-primary-eA3Kzwpz); 3 workers total; 22 tasks left in task-pool. [ Info: running task 79 on process 56 (Schedulers-primary-eA3Kzwpz); 3 workers total; 21 tasks left in task-pool. [ Info: running task 80 on process 57 (Schedulers-primary-eA3Kzwpz); 3 workers total; 20 tasks left in task-pool. [ Info: running task 81 on process 55 (Schedulers-primary-eA3Kzwpz); 3 workers total; 19 tasks left in task-pool. [ Info: running task 82 on process 56 (Schedulers-primary-eA3Kzwpz); 3 workers total; 18 tasks left in task-pool. [ Info: running task 83 on process 57 (Schedulers-primary-eA3Kzwpz); 3 workers total; 17 tasks left in task-pool. [ Info: running task 84 on process 55 (Schedulers-primary-eA3Kzwpz); 3 workers total; 16 tasks left in task-pool. [ Info: running task 85 on process 56 (Schedulers-primary-eA3Kzwpz); 3 workers total; 15 tasks left in task-pool. [ Info: running task 86 on process 57 (Schedulers-primary-eA3Kzwpz); 3 workers total; 14 tasks left in task-pool. [ Info: running task 87 on process 55 (Schedulers-primary-eA3Kzwpz); 3 workers total; 13 tasks left in task-pool. [ Info: running task 88 on process 56 (Schedulers-primary-eA3Kzwpz); 3 workers total; 12 tasks left in task-pool. [ Info: running task 89 on process 57 (Schedulers-primary-eA3Kzwpz); 3 workers total; 11 tasks left in task-pool. [ Info: running task 90 on process 55 (Schedulers-primary-eA3Kzwpz); 3 workers total; 10 tasks left in task-pool. [ Info: running task 91 on process 56 (Schedulers-primary-eA3Kzwpz); 3 workers total; 9 tasks left in task-pool. [ Info: running task 92 on process 57 (Schedulers-primary-eA3Kzwpz); 3 workers total; 8 tasks left in task-pool. [ Info: running task 93 on process 55 (Schedulers-primary-eA3Kzwpz); 3 workers total; 7 tasks left in task-pool. [ Info: running task 94 on process 56 (Schedulers-primary-eA3Kzwpz); 3 workers total; 6 tasks left in task-pool. [ Info: running task 95 on process 57 (Schedulers-primary-eA3Kzwpz); 3 workers total; 5 tasks left in task-pool. [ Info: running task 96 on process 55 (Schedulers-primary-eA3Kzwpz); 3 workers total; 4 tasks left in task-pool. [ Info: running task 97 on process 56 (Schedulers-primary-eA3Kzwpz); 3 workers total; 3 tasks left in task-pool. [ Info: running task 98 on process 57 (Schedulers-primary-eA3Kzwpz); 3 workers total; 2 tasks left in task-pool. [ Info: running task 99 on process 55 (Schedulers-primary-eA3Kzwpz); 3 workers total; 1 tasks left in task-pool. [ Info: running task 100 on process 56 (Schedulers-primary-eA3Kzwpz); 3 workers total; 0 tasks left in task-pool. ┌ Warning: addprocs/rmprocs taking longer than expected, cancelling. └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:823 Test Summary: | Pass Total Time pmap with blocking addprocs | 103 103 2m09.8s [ Info: running task 1 on process 58 (Schedulers-primary-eA3Kzwpz); 5 workers total; 19 tasks left in task-pool. [ Info: running task 2 on process 59 (Schedulers-primary-eA3Kzwpz); 5 workers total; 15 tasks left in task-pool. [ Info: running task 3 on process 60 (Schedulers-primary-eA3Kzwpz); 5 workers total; 15 tasks left in task-pool. [ Info: running task 4 on process 61 (Schedulers-primary-eA3Kzwpz); 5 workers total; 15 tasks left in task-pool. [ Info: running task 5 on process 62 (Schedulers-primary-eA3Kzwpz); 5 workers total; 15 tasks left in task-pool. [ Info: running task 6 on process 58 (Schedulers-primary-eA3Kzwpz); 5 workers total; 14 tasks left in task-pool. [ Info: running task 7 on process 59 (Schedulers-primary-eA3Kzwpz); 5 workers total; 13 tasks left in task-pool. [ Info: running task 8 on process 60 (Schedulers-primary-eA3Kzwpz); 5 workers total; 12 tasks left in task-pool. [ Info: running task 9 on process 61 (Schedulers-primary-eA3Kzwpz); 5 workers total; 11 tasks left in task-pool. [ Info: running task 10 on process 62 (Schedulers-primary-eA3Kzwpz); 5 workers total; 10 tasks left in task-pool. [ Info: running task 11 on process 58 (Schedulers-primary-eA3Kzwpz); 5 workers total; 9 tasks left in task-pool. [ Info: running task 12 on process 59 (Schedulers-primary-eA3Kzwpz); 5 workers total; 8 tasks left in task-pool. [ Info: running task 13 on process 60 (Schedulers-primary-eA3Kzwpz); 5 workers total; 7 tasks left in task-pool. [ Info: running task 14 on process 61 (Schedulers-primary-eA3Kzwpz); 5 workers total; 6 tasks left in task-pool. [ Info: running task 15 on process 62 (Schedulers-primary-eA3Kzwpz); 5 workers total; 5 tasks left in task-pool. [ Info: running task 16 on process 58 (Schedulers-primary-eA3Kzwpz); 5 workers total; 4 tasks left in task-pool. [ Info: running task 17 on process 59 (Schedulers-primary-eA3Kzwpz); 5 workers total; 3 tasks left in task-pool. [ Info: running task 18 on process 60 (Schedulers-primary-eA3Kzwpz); 5 workers total; 2 tasks left in task-pool. [ Info: running task 19 on process 61 (Schedulers-primary-eA3Kzwpz); 5 workers total; 1 tasks left in task-pool. [ Info: running task 20 on process 62 (Schedulers-primary-eA3Kzwpz); 5 workers total; 0 tasks left in task-pool. ┌ Warning: caught an exception, there have been 0 failure(s) on process 62 (Schedulers-primary-eA3Kzwpz)... └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1111 ┌ Warning: Schedulers.TimeoutException(62, 17.031870126724243) │ error type: Schedulers.TimeoutException │ 1-element ExceptionStack: │ Schedulers.TimeoutException(62, 17.031870126724243) │ Stacktrace: │ [1] remotecall_func_wait_timeout(::Vector{Float64}, ::Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", typeof(nworkers), Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, ::SchedulerOptions{String}, ::Nothing, ::Function, ::Function, ::Int64, ::var"#foo5c#foo5c##0", ::Int64, ::Int64, ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:343 │ [2] remotecall_func_wait_timeout(::Vector{Float64}, ::Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", typeof(nworkers), Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, ::SchedulerOptions{String}, ::Nothing, ::Function, ::Function, ::Int64, ::Function, ::Int64, ::Int64, ::Vararg{Any}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:338 │ [3] (::Schedulers.var"#107#108"{@Kwargs{}, SchedulerOptions{String}, var"#foo5c#foo5c##0", Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", typeof(nworkers), Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, Dict{String, Any}, Tuple{String}, Vector{Float64}, Int64})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1105 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 [ Info: running task 20 on process 58 (Schedulers-primary-eA3Kzwpz); 5 workers total; 0 tasks left in task-pool. Test Summary: | Pass Total Time pmap with timeout | 21 21 1m58.3s [ Info: running task 1 on process 58 (Schedulers-primary-eA3Kzwpz); 5 workers total; 19 tasks left in task-pool. [ Info: running task 2 on process 59 (Schedulers-primary-eA3Kzwpz); 5 workers total; 16 tasks left in task-pool. [ Info: running task 3 on process 60 (Schedulers-primary-eA3Kzwpz); 5 workers total; 16 tasks left in task-pool. [ Info: running task 4 on process 61 (Schedulers-primary-eA3Kzwpz); 5 workers total; 16 tasks left in task-pool. [ Info: running task 5 on process 58 (Schedulers-primary-eA3Kzwpz); 5 workers total; 15 tasks left in task-pool. [ Info: running task 6 on process 59 (Schedulers-primary-eA3Kzwpz); 5 workers total; 14 tasks left in task-pool. [ Info: running task 7 on process 60 (Schedulers-primary-eA3Kzwpz); 5 workers total; 13 tasks left in task-pool. [ Info: running task 8 on process 61 (Schedulers-primary-eA3Kzwpz); 5 workers total; 12 tasks left in task-pool. [ Info: running task 9 on process 58 (Schedulers-primary-eA3Kzwpz); 5 workers total; 11 tasks left in task-pool. [ Info: running task 10 on process 59 (Schedulers-primary-eA3Kzwpz); 5 workers total; 10 tasks left in task-pool. [ Info: running task 11 on process 60 (Schedulers-primary-eA3Kzwpz); 5 workers total; 9 tasks left in task-pool. [ Info: running task 12 on process 61 (Schedulers-primary-eA3Kzwpz); 5 workers total; 8 tasks left in task-pool. [ Info: running task 13 on process 58 (Schedulers-primary-eA3Kzwpz); 5 workers total; 7 tasks left in task-pool. [ Info: running task 14 on process 59 (Schedulers-primary-eA3Kzwpz); 5 workers total; 6 tasks left in task-pool. [ Info: running task 15 on process 60 (Schedulers-primary-eA3Kzwpz); 5 workers total; 5 tasks left in task-pool. [ Info: running task 16 on process 61 (Schedulers-primary-eA3Kzwpz); 5 workers total; 4 tasks left in task-pool. [ Info: running task 17 on process 58 (Schedulers-primary-eA3Kzwpz); 5 workers total; 3 tasks left in task-pool. [ Info: running task 18 on process 59 (Schedulers-primary-eA3Kzwpz); 5 workers total; 2 tasks left in task-pool. [ Info: running task 19 on process 60 (Schedulers-primary-eA3Kzwpz); 5 workers total; 1 tasks left in task-pool. [ Info: running task 20 on process 61 (Schedulers-primary-eA3Kzwpz); 5 workers total; 0 tasks left in task-pool. ┌ Warning: caught an exception, there have been 0 failure(s) on process 61 (Schedulers-primary-eA3Kzwpz)... └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1111 ┌ Warning: skipping task '20' that timed out └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1114 ┌ Warning: Schedulers.TimeoutException(61, 10.018981218338013) │ error type: Schedulers.TimeoutException │ 1-element ExceptionStack: │ Schedulers.TimeoutException(61, 10.018981218338013) │ Stacktrace: │ [1] remotecall_func_wait_timeout(::Vector{Float64}, ::Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", typeof(nworkers), Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, ::SchedulerOptions{String}, ::Nothing, ::Function, ::Function, ::Int64, ::var"#foo5c#foo5c##1", ::Int64, ::Int64, ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:343 │ [2] remotecall_func_wait_timeout(::Vector{Float64}, ::Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", typeof(nworkers), Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, ::SchedulerOptions{String}, ::Nothing, ::Function, ::Function, ::Int64, ::Function, ::Int64, ::Int64, ::Vararg{Any}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:338 │ [3] (::Schedulers.var"#107#108"{@Kwargs{}, SchedulerOptions{String}, var"#foo5c#foo5c##1", Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", typeof(nworkers), Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, Dict{String, Any}, Tuple{String}, Vector{Float64}, Int64})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1105 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 Test Summary: | Pass Total Time pmap with timeout and skip_tasks_that_timeout=true | 21 21 25.2s ┌ Warning: problem initializing 58, removing 58 from cluster. └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:675 ┌ Warning: no process with id 58 exists │ error type: ErrorException │ 1-element ExceptionStack: │ no process with id 58 exists │ Stacktrace: │ [1] error(s::String) │ @ Base ./error.jl:44 │ [2] worker_from_id(pg::Distributed.ProcessGroup, i::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:1114 │ [3] worker_from_id │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:1106 [inlined] │ [4] remotecall(::Function, ::Int64, ::Type, ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:447 │ [5] remotecall │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:447 [inlined] │ [6] init │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:299 [inlined] │ [7] #20 │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:340 [inlined] │ [8] (::Schedulers.var"#loop##12#loop##13"{Schedulers.ElasticLoop{typeof(addprocs), var"#20#21"{var"#init#init##0"{Dict{Int64, Future}}, Dict{Int64, Future}}, Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, Set{Int64}, Set{Int64}, Int64})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:668 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 [ Info: running task 1 on process 61 (Schedulers-primary-eA3Kzwpz); 2 workers total; 3 tasks left in task-pool. [ Info: running task 2 on process 63 (Schedulers-primary-eA3Kzwpz); 2 workers total; 2 tasks left in task-pool. ┌ Warning: problem initializing 59, removing 59 from cluster. └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:675 ┌ Warning: ProcessExitedException(59) │ error type: ProcessExitedException │ 1-element ExceptionStack: │ ProcessExitedException(59) │ Stacktrace: │ [1] worker_from_id(pg::Distributed.ProcessGroup, i::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:1109 │ [2] worker_from_id │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:1106 [inlined] │ [3] remotecall(::Function, ::Int64, ::Type, ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:447 │ [4] remotecall │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:447 [inlined] │ [5] init │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:299 [inlined] │ [6] #20 │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:340 [inlined] │ [7] (::Schedulers.var"#loop##12#loop##13"{Schedulers.ElasticLoop{typeof(addprocs), var"#20#21"{var"#init#init##0"{Dict{Int64, Future}}, Dict{Int64, Future}}, Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, Set{Int64}, Set{Int64}, Int64})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:668 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 ┌ Warning: problem initializing 60, removing 60 from cluster. └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:675 ┌ Warning: ProcessExitedException(60) │ error type: ProcessExitedException │ 1-element ExceptionStack: │ ProcessExitedException(60) │ Stacktrace: │ [1] worker_from_id(pg::Distributed.ProcessGroup, i::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:1109 │ [2] worker_from_id │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:1106 [inlined] │ [3] remotecall(::Function, ::Int64, ::Type, ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:447 │ [4] remotecall │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:447 [inlined] │ [5] init │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:299 [inlined] │ [6] #20 │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:340 [inlined] │ [7] (::Schedulers.var"#loop##12#loop##13"{Schedulers.ElasticLoop{typeof(addprocs), var"#20#21"{var"#init#init##0"{Dict{Int64, Future}}, Dict{Int64, Future}}, Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, Set{Int64}, Set{Int64}, Int64})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:668 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 From worker 61: [ Info: restart task... From worker 61: ┌ Info: ...restart task, From worker 61: │ _s = From worker 61: │ 1-element Vector{Int64}: From worker 61: └ 1 From worker 61: [ Info: _s on pid=61 is 1 From worker 63: [ Info: restart task... From worker 63: ┌ Info: ...restart task, From worker 63: │ _s = From worker 63: │ 1-element Vector{Int64}: From worker 63: └ 1 From worker 63: [ Info: _s on pid=63 is 1 From worker 61: [ Info: _s on pid=61 is 2 From worker 63: [ Info: _s on pid=63 is 2 From worker 61: [ Info: checkpoint task... From worker 61: [ Info: ...checkpoint task with state=2. ┌ Warning: caught an exception, there have been 0 failure(s) on process 61 (Schedulers-primary-eA3Kzwpz)... └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1111 ┌ Warning: TaskFailedException │ │ nested task error: On worker 61: │ Schedulers.PreemptException() │ Stacktrace: │ [1] #default_threadpool_checkpoint_call#25 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:278 │ [2] default_threadpool_checkpoint_call │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:250 │ [3] #handle_msg##8 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:303 │ [4] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:70 │ [5] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:79 │ [6] #schedule_call##2 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:88 │ Stacktrace: │ [1] remotecall_wait(::Function, ::Distributed.Worker, ::Future, ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:507 │ [2] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:497 [inlined] │ [3] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:519 [inlined] │ [4] (::Schedulers.var"#43#44"{@Kwargs{}, Future, var"#22#23"{var"#checkpoint_task#checkpoint_task##0", Dict{Int64, Future}}, var"#24#25"{var"#restart_task!#restart_task!##0", Dict{Int64, Future}}, Int64, var"#30#31"{var"#foo5d#foo5d##0", String, Dict{Int64, Future}}, Int64, Tuple{Int64}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:339 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] remotecall_func_wait_timeout(tsk_times::Vector{Float64}, eloop::Schedulers.ElasticLoop{typeof(addprocs), var"#20#21"{var"#init#init##0"{Dict{Int64, Future}}, Dict{Int64, Future}}, Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, options::SchedulerOptions{String}, preempt_channel_future::Future, checkpoint_task::Function, restart_task::Function, tsk::Int64, f::var"#30#31"{var"#foo5d#foo5d##0", String, Dict{Int64, Future}}, pid::Int64, args::Int64; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:349 │ [5] remotecall_func_wait_timeout(tsk_times::Vector{Float64}, eloop::Schedulers.ElasticLoop{typeof(addprocs), var"#20#21"{var"#init#init##0"{Dict{Int64, Future}}, Dict{Int64, Future}}, Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, options::SchedulerOptions{String}, preempt_channel_future::Future, checkpoint_task::Function, restart_task::Function, tsk::Int64, f::Function, pid::Int64, args::Int64) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:338 │ [6] (::Schedulers.var"#107#108"{@Kwargs{}, SchedulerOptions{String}, var"#30#31"{var"#foo5d#foo5d##0", String, Dict{Int64, Future}}, Schedulers.ElasticLoop{typeof(addprocs), var"#20#21"{var"#init#init##0"{Dict{Int64, Future}}, Dict{Int64, Future}}, Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, Dict{String, Any}, Tuple{}, Vector{Float64}, Int64})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1105 │ │ nested task error: On worker 61: │ Schedulers.PreemptException() │ Stacktrace: │ [1] #default_threadpool_checkpoint_call#25 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:278 │ [2] default_threadpool_checkpoint_call │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:250 │ [3] #handle_msg##8 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:303 │ [4] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:70 │ [5] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:79 │ [6] #schedule_call##2 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:88 │ Stacktrace: │ [1] remotecall_wait(::Function, ::Distributed.Worker, ::Future, ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:507 │ [2] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:497 [inlined] │ [3] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:519 [inlined] │ [4] (::Schedulers.var"#43#44"{@Kwargs{}, Future, var"#22#23"{var"#checkpoint_task#checkpoint_task##0", Dict{Int64, Future}}, var"#24#25"{var"#restart_task!#restart_task!##0", Dict{Int64, Future}}, Int64, var"#30#31"{var"#foo5d#foo5d##0", String, Dict{Int64, Future}}, Int64, Tuple{Int64}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:339 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 ┌ Warning: too many failures on process with id=61 (Schedulers-primary-eA3Kzwpz), removing from process list └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:393 From worker 63: [ Info: _s on pid=63 is 3 From worker 63: [ Info: _s on pid=63 is 4 From worker 63: [ Info: _s on pid=63 is 5 From worker 63: [ Info: _s on pid=63 is 6 [ Info: running task 3 on process 64 (Schedulers-primary-eA3Kzwpz); 2 workers total; 2 tasks left in task-pool. From worker 63: [ Info: _s on pid=63 is 7 From worker 64: [ Info: restart task... From worker 64: ┌ Info: ...restart task, From worker 64: │ _s = From worker 64: │ 1-element Vector{Int64}: From worker 64: └ 1 From worker 64: [ Info: _s on pid=64 is 1 From worker 63: [ Info: _s on pid=63 is 8 From worker 64: [ Info: _s on pid=64 is 2 From worker 63: [ Info: _s on pid=63 is 9 From worker 64: [ Info: _s on pid=64 is 3 From worker 63: [ Info: _s on pid=63 is 10 From worker 64: [ Info: _s on pid=64 is 4 [ Info: running task 4 on process 63 (Schedulers-primary-eA3Kzwpz); 2 workers total; 1 tasks left in task-pool. From worker 63: [ Info: restart task... From worker 63: ┌ Info: ...restart task, From worker 63: │ _s = From worker 63: │ 1-element Vector{Int64}: From worker 63: └ 1 From worker 63: [ Info: _s on pid=63 is 1 From worker 64: [ Info: _s on pid=64 is 5 From worker 63: [ Info: _s on pid=63 is 2 From worker 64: [ Info: _s on pid=64 is 6 From worker 63: [ Info: _s on pid=63 is 3 From worker 64: [ Info: _s on pid=64 is 7 From worker 63: [ Info: _s on pid=63 is 4 From worker 64: [ Info: _s on pid=64 is 8 From worker 63: [ Info: _s on pid=63 is 5 From worker 64: [ Info: _s on pid=64 is 9 From worker 63: [ Info: _s on pid=63 is 6 From worker 64: [ Info: _s on pid=64 is 10 From worker 63: [ Info: _s on pid=63 is 7 [ Info: running task 1 on process 64 (Schedulers-primary-eA3Kzwpz); 2 workers total; 0 tasks left in task-pool. From worker 64: [ Info: restart task... From worker 64: ┌ Info: ...restart task, From worker 64: │ _s = From worker 64: │ 1-element Vector{Int64}: From worker 64: └ 2 From worker 64: [ Info: _s on pid=64 is 2 From worker 63: [ Info: _s on pid=63 is 8 From worker 64: [ Info: _s on pid=64 is 3 From worker 63: [ Info: _s on pid=63 is 9 From worker 64: [ Info: _s on pid=64 is 4 From worker 63: [ Info: _s on pid=63 is 10 From worker 64: [ Info: _s on pid=64 is 5 From worker 64: [ Info: _s on pid=64 is 6 From worker 64: [ Info: _s on pid=64 is 7 From worker 64: [ Info: _s on pid=64 is 8 From worker 64: [ Info: _s on pid=64 is 9 From worker 64: [ Info: _s on pid=64 is 10 Test Summary: | Pass Total Time pmap with task checkpoint and restart | 1 1 2m26.5s [ Info: running task 1 on process 65 (Schedulers-primary-eA3Kzwpz); 5 workers total; 99 tasks left in task-pool. [ Info: running task 2 on process 66 (Schedulers-primary-eA3Kzwpz); 5 workers total; 98 tasks left in task-pool. [ Info: running task 3 on process 67 (Schedulers-primary-eA3Kzwpz); 5 workers total; 97 tasks left in task-pool. [ Info: running task 4 on process 68 (Schedulers-primary-eA3Kzwpz); 5 workers total; 96 tasks left in task-pool. [ Info: running task 5 on process 69 (Schedulers-primary-eA3Kzwpz); 5 workers total; 95 tasks left in task-pool. [ Info: running task 6 on process 65 (Schedulers-primary-eA3Kzwpz); 5 workers total; 94 tasks left in task-pool. [ Info: running task 7 on process 66 (Schedulers-primary-eA3Kzwpz); 5 workers total; 93 tasks left in task-pool. [ Info: running task 8 on process 67 (Schedulers-primary-eA3Kzwpz); 5 workers total; 92 tasks left in task-pool. [ Info: running task 9 on process 68 (Schedulers-primary-eA3Kzwpz); 5 workers total; 90 tasks left in task-pool. [ Info: running task 10 on process 69 (Schedulers-primary-eA3Kzwpz); 5 workers total; 90 tasks left in task-pool. [ Info: running task 11 on process 65 (Schedulers-primary-eA3Kzwpz); 5 workers total; 89 tasks left in task-pool. [ Info: running task 12 on process 66 (Schedulers-primary-eA3Kzwpz); 5 workers total; 88 tasks left in task-pool. [ Info: running task 13 on process 67 (Schedulers-primary-eA3Kzwpz); 5 workers total; 87 tasks left in task-pool. [ Info: running task 14 on process 68 (Schedulers-primary-eA3Kzwpz); 5 workers total; 86 tasks left in task-pool. [ Info: running task 15 on process 69 (Schedulers-primary-eA3Kzwpz); 5 workers total; 85 tasks left in task-pool. [ Info: running task 16 on process 65 (Schedulers-primary-eA3Kzwpz); 5 workers total; 84 tasks left in task-pool. [ Info: running task 17 on process 66 (Schedulers-primary-eA3Kzwpz); 5 workers total; 83 tasks left in task-pool. [ Info: running task 18 on process 67 (Schedulers-primary-eA3Kzwpz); 5 workers total; 82 tasks left in task-pool. [ Info: running task 19 on process 68 (Schedulers-primary-eA3Kzwpz); 5 workers total; 81 tasks left in task-pool. [ Info: running task 20 on process 69 (Schedulers-primary-eA3Kzwpz); 5 workers total; 80 tasks left in task-pool. [ Info: running task 21 on process 65 (Schedulers-primary-eA3Kzwpz); 5 workers total; 79 tasks left in task-pool. [ Info: running task 22 on process 66 (Schedulers-primary-eA3Kzwpz); 5 workers total; 78 tasks left in task-pool. [ Info: running task 23 on process 67 (Schedulers-primary-eA3Kzwpz); 5 workers total; 77 tasks left in task-pool. [ Info: running task 24 on process 68 (Schedulers-primary-eA3Kzwpz); 5 workers total; 76 tasks left in task-pool. [ Info: running task 25 on process 69 (Schedulers-primary-eA3Kzwpz); 5 workers total; 75 tasks left in task-pool. [ Info: running task 26 on process 65 (Schedulers-primary-eA3Kzwpz); 5 workers total; 74 tasks left in task-pool. [ Info: running task 27 on process 66 (Schedulers-primary-eA3Kzwpz); 5 workers total; 73 tasks left in task-pool. [ Info: running task 28 on process 67 (Schedulers-primary-eA3Kzwpz); 5 workers total; 72 tasks left in task-pool. [ Info: running task 29 on process 68 (Schedulers-primary-eA3Kzwpz); 5 workers total; 71 tasks left in task-pool. [ Info: running task 30 on process 69 (Schedulers-primary-eA3Kzwpz); 5 workers total; 70 tasks left in task-pool. [ Info: running task 31 on process 65 (Schedulers-primary-eA3Kzwpz); 5 workers total; 69 tasks left in task-pool. [ Info: running task 32 on process 66 (Schedulers-primary-eA3Kzwpz); 5 workers total; 68 tasks left in task-pool. [ Info: running task 33 on process 67 (Schedulers-primary-eA3Kzwpz); 5 workers total; 67 tasks left in task-pool. [ Info: running task 34 on process 68 (Schedulers-primary-eA3Kzwpz); 5 workers total; 66 tasks left in task-pool. [ Info: running task 35 on process 69 (Schedulers-primary-eA3Kzwpz); 5 workers total; 65 tasks left in task-pool. [ Info: running task 36 on process 65 (Schedulers-primary-eA3Kzwpz); 5 workers total; 64 tasks left in task-pool. [ Info: running task 37 on process 66 (Schedulers-primary-eA3Kzwpz); 5 workers total; 63 tasks left in task-pool. [ Info: running task 38 on process 67 (Schedulers-primary-eA3Kzwpz); 5 workers total; 62 tasks left in task-pool. [ Info: running task 39 on process 68 (Schedulers-primary-eA3Kzwpz); 5 workers total; 61 tasks left in task-pool. [ Info: running task 40 on process 69 (Schedulers-primary-eA3Kzwpz); 5 workers total; 60 tasks left in task-pool. [ Info: running task 41 on process 65 (Schedulers-primary-eA3Kzwpz); 5 workers total; 59 tasks left in task-pool. [ Info: running task 42 on process 66 (Schedulers-primary-eA3Kzwpz); 5 workers total; 58 tasks left in task-pool. [ Info: running task 43 on process 67 (Schedulers-primary-eA3Kzwpz); 5 workers total; 57 tasks left in task-pool. [ Info: running task 44 on process 68 (Schedulers-primary-eA3Kzwpz); 5 workers total; 56 tasks left in task-pool. [ Info: running task 45 on process 69 (Schedulers-primary-eA3Kzwpz); 5 workers total; 55 tasks left in task-pool. [ Info: running task 46 on process 65 (Schedulers-primary-eA3Kzwpz); 5 workers total; 54 tasks left in task-pool. [ Info: running task 47 on process 66 (Schedulers-primary-eA3Kzwpz); 5 workers total; 53 tasks left in task-pool. [ Info: running task 48 on process 67 (Schedulers-primary-eA3Kzwpz); 5 workers total; 52 tasks left in task-pool. [ Info: running task 49 on process 68 (Schedulers-primary-eA3Kzwpz); 5 workers total; 51 tasks left in task-pool. [ Info: running task 50 on process 69 (Schedulers-primary-eA3Kzwpz); 5 workers total; 50 tasks left in task-pool. [ Info: running task 51 on process 65 (Schedulers-primary-eA3Kzwpz); 5 workers total; 49 tasks left in task-pool. [ Info: running task 52 on process 66 (Schedulers-primary-eA3Kzwpz); 5 workers total; 48 tasks left in task-pool. [ Info: running task 53 on process 67 (Schedulers-primary-eA3Kzwpz); 5 workers total; 47 tasks left in task-pool. [ Info: running task 54 on process 68 (Schedulers-primary-eA3Kzwpz); 5 workers total; 46 tasks left in task-pool. [ Info: running task 55 on process 69 (Schedulers-primary-eA3Kzwpz); 5 workers total; 45 tasks left in task-pool. [ Info: running task 56 on process 65 (Schedulers-primary-eA3Kzwpz); 5 workers total; 44 tasks left in task-pool. [ Info: running task 57 on process 66 (Schedulers-primary-eA3Kzwpz); 5 workers total; 43 tasks left in task-pool. [ Info: running task 58 on process 67 (Schedulers-primary-eA3Kzwpz); 5 workers total; 42 tasks left in task-pool. [ Info: running task 59 on process 68 (Schedulers-primary-eA3Kzwpz); 5 workers total; 41 tasks left in task-pool. [ Info: running task 60 on process 69 (Schedulers-primary-eA3Kzwpz); 5 workers total; 40 tasks left in task-pool. [ Info: running task 61 on process 65 (Schedulers-primary-eA3Kzwpz); 5 workers total; 39 tasks left in task-pool. [ Info: running task 62 on process 66 (Schedulers-primary-eA3Kzwpz); 5 workers total; 38 tasks left in task-pool. [ Info: running task 63 on process 67 (Schedulers-primary-eA3Kzwpz); 5 workers total; 37 tasks left in task-pool. [ Info: running task 64 on process 68 (Schedulers-primary-eA3Kzwpz); 5 workers total; 36 tasks left in task-pool. [ Info: running task 65 on process 69 (Schedulers-primary-eA3Kzwpz); 5 workers total; 35 tasks left in task-pool. [ Info: running task 66 on process 65 (Schedulers-primary-eA3Kzwpz); 5 workers total; 34 tasks left in task-pool. [ Info: running task 67 on process 66 (Schedulers-primary-eA3Kzwpz); 5 workers total; 33 tasks left in task-pool. [ Info: running task 68 on process 67 (Schedulers-primary-eA3Kzwpz); 5 workers total; 32 tasks left in task-pool. [ Info: running task 69 on process 68 (Schedulers-primary-eA3Kzwpz); 5 workers total; 31 tasks left in task-pool. [ Info: running task 70 on process 69 (Schedulers-primary-eA3Kzwpz); 5 workers total; 30 tasks left in task-pool. [ Info: running task 71 on process 65 (Schedulers-primary-eA3Kzwpz); 5 workers total; 29 tasks left in task-pool. [ Info: running task 72 on process 66 (Schedulers-primary-eA3Kzwpz); 5 workers total; 28 tasks left in task-pool. [ Info: running task 73 on process 67 (Schedulers-primary-eA3Kzwpz); 5 workers total; 27 tasks left in task-pool. [ Info: running task 74 on process 68 (Schedulers-primary-eA3Kzwpz); 5 workers total; 26 tasks left in task-pool. [ Info: running task 75 on process 69 (Schedulers-primary-eA3Kzwpz); 5 workers total; 25 tasks left in task-pool. [ Info: running task 76 on process 65 (Schedulers-primary-eA3Kzwpz); 5 workers total; 24 tasks left in task-pool. [ Info: running task 77 on process 66 (Schedulers-primary-eA3Kzwpz); 5 workers total; 23 tasks left in task-pool. [ Info: running task 78 on process 67 (Schedulers-primary-eA3Kzwpz); 5 workers total; 22 tasks left in task-pool. [ Info: running task 79 on process 68 (Schedulers-primary-eA3Kzwpz); 5 workers total; 21 tasks left in task-pool. [ Info: running task 80 on process 69 (Schedulers-primary-eA3Kzwpz); 5 workers total; 20 tasks left in task-pool. [ Info: running task 81 on process 65 (Schedulers-primary-eA3Kzwpz); 5 workers total; 19 tasks left in task-pool. [ Info: running task 82 on process 66 (Schedulers-primary-eA3Kzwpz); 5 workers total; 18 tasks left in task-pool. [ Info: running task 83 on process 67 (Schedulers-primary-eA3Kzwpz); 5 workers total; 17 tasks left in task-pool. [ Info: running task 84 on process 68 (Schedulers-primary-eA3Kzwpz); 5 workers total; 15 tasks left in task-pool. [ Info: running task 85 on process 69 (Schedulers-primary-eA3Kzwpz); 5 workers total; 15 tasks left in task-pool. [ Info: running task 86 on process 65 (Schedulers-primary-eA3Kzwpz); 5 workers total; 14 tasks left in task-pool. [ Info: running task 87 on process 66 (Schedulers-primary-eA3Kzwpz); 5 workers total; 13 tasks left in task-pool. [ Info: running task 88 on process 67 (Schedulers-primary-eA3Kzwpz); 5 workers total; 12 tasks left in task-pool. [ Info: running task 89 on process 68 (Schedulers-primary-eA3Kzwpz); 5 workers total; 10 tasks left in task-pool. [ Info: running task 90 on process 69 (Schedulers-primary-eA3Kzwpz); 5 workers total; 10 tasks left in task-pool. [ Info: running task 91 on process 65 (Schedulers-primary-eA3Kzwpz); 5 workers total; 9 tasks left in task-pool. [ Info: running task 92 on process 66 (Schedulers-primary-eA3Kzwpz); 5 workers total; 8 tasks left in task-pool. [ Info: running task 93 on process 67 (Schedulers-primary-eA3Kzwpz); 5 workers total; 7 tasks left in task-pool. [ Info: running task 94 on process 68 (Schedulers-primary-eA3Kzwpz); 5 workers total; 6 tasks left in task-pool. [ Info: running task 95 on process 69 (Schedulers-primary-eA3Kzwpz); 5 workers total; 5 tasks left in task-pool. [ Info: running task 96 on process 65 (Schedulers-primary-eA3Kzwpz); 5 workers total; 4 tasks left in task-pool. [ Info: running task 97 on process 66 (Schedulers-primary-eA3Kzwpz); 5 workers total; 3 tasks left in task-pool. [ Info: running task 98 on process 67 (Schedulers-primary-eA3Kzwpz); 5 workers total; 2 tasks left in task-pool. [ Info: running task 99 on process 68 (Schedulers-primary-eA3Kzwpz); 5 workers total; 1 tasks left in task-pool. [ Info: running task 100 on process 69 (Schedulers-primary-eA3Kzwpz); 5 workers total; 0 tasks left in task-pool. [ Info: reducing from 5 checkpoints using process 65 (5 workers, 1 reduce workers). [ Info: reducing from 4 checkpoints using process 66 (5 workers, 2 reduce workers). [ Info: reducing from 2 checkpoints using process 65 (5 workers, 2 reduce workers). [ Info: reducing from 2 checkpoints using process 65 (5 workers, 2 reduce workers). Test Summary: | Pass Total Time pmapreduce, stable cluster test, backwards compatability | 36 36 3m14.7s [ Info: running task 1 on process 70 (Schedulers-primary-eA3Kzwpz); 5 workers total; 99 tasks left in task-pool. [ Info: running task 2 on process 71 (Schedulers-primary-eA3Kzwpz); 5 workers total; 98 tasks left in task-pool. [ Info: running task 3 on process 72 (Schedulers-primary-eA3Kzwpz); 5 workers total; 97 tasks left in task-pool. [ Info: running task 4 on process 73 (Schedulers-primary-eA3Kzwpz); 5 workers total; 95 tasks left in task-pool. [ Info: running task 5 on process 74 (Schedulers-primary-eA3Kzwpz); 5 workers total; 95 tasks left in task-pool. [ Info: running task 6 on process 70 (Schedulers-primary-eA3Kzwpz); 5 workers total; 94 tasks left in task-pool. [ Info: running task 7 on process 71 (Schedulers-primary-eA3Kzwpz); 5 workers total; 93 tasks left in task-pool. [ Info: running task 8 on process 72 (Schedulers-primary-eA3Kzwpz); 5 workers total; 92 tasks left in task-pool. [ Info: running task 9 on process 73 (Schedulers-primary-eA3Kzwpz); 5 workers total; 90 tasks left in task-pool. [ Info: running task 10 on process 74 (Schedulers-primary-eA3Kzwpz); 5 workers total; 90 tasks left in task-pool. [ Info: running task 11 on process 70 (Schedulers-primary-eA3Kzwpz); 5 workers total; 89 tasks left in task-pool. [ Info: running task 12 on process 71 (Schedulers-primary-eA3Kzwpz); 5 workers total; 88 tasks left in task-pool. [ Info: running task 13 on process 72 (Schedulers-primary-eA3Kzwpz); 5 workers total; 87 tasks left in task-pool. [ Info: running task 14 on process 73 (Schedulers-primary-eA3Kzwpz); 5 workers total; 85 tasks left in task-pool. [ Info: running task 15 on process 74 (Schedulers-primary-eA3Kzwpz); 5 workers total; 85 tasks left in task-pool. [ Info: running task 16 on process 70 (Schedulers-primary-eA3Kzwpz); 5 workers total; 84 tasks left in task-pool. [ Info: running task 17 on process 71 (Schedulers-primary-eA3Kzwpz); 5 workers total; 83 tasks left in task-pool. [ Info: running task 18 on process 72 (Schedulers-primary-eA3Kzwpz); 5 workers total; 82 tasks left in task-pool. [ Info: running task 19 on process 73 (Schedulers-primary-eA3Kzwpz); 5 workers total; 80 tasks left in task-pool. [ Info: running task 20 on process 74 (Schedulers-primary-eA3Kzwpz); 5 workers total; 80 tasks left in task-pool. [ Info: running task 21 on process 70 (Schedulers-primary-eA3Kzwpz); 5 workers total; 79 tasks left in task-pool. [ Info: running task 22 on process 71 (Schedulers-primary-eA3Kzwpz); 5 workers total; 78 tasks left in task-pool. [ Info: running task 23 on process 72 (Schedulers-primary-eA3Kzwpz); 5 workers total; 77 tasks left in task-pool. [ Info: running task 24 on process 73 (Schedulers-primary-eA3Kzwpz); 5 workers total; 75 tasks left in task-pool. [ Info: running task 25 on process 74 (Schedulers-primary-eA3Kzwpz); 5 workers total; 75 tasks left in task-pool. [ Info: running task 26 on process 70 (Schedulers-primary-eA3Kzwpz); 5 workers total; 74 tasks left in task-pool. [ Info: running task 27 on process 71 (Schedulers-primary-eA3Kzwpz); 5 workers total; 73 tasks left in task-pool. [ Info: running task 28 on process 72 (Schedulers-primary-eA3Kzwpz); 5 workers total; 72 tasks left in task-pool. [ Info: running task 29 on process 73 (Schedulers-primary-eA3Kzwpz); 5 workers total; 70 tasks left in task-pool. [ Info: running task 30 on process 74 (Schedulers-primary-eA3Kzwpz); 5 workers total; 70 tasks left in task-pool. [ Info: running task 31 on process 70 (Schedulers-primary-eA3Kzwpz); 5 workers total; 69 tasks left in task-pool. [ Info: running task 32 on process 71 (Schedulers-primary-eA3Kzwpz); 5 workers total; 68 tasks left in task-pool. [ Info: running task 33 on process 72 (Schedulers-primary-eA3Kzwpz); 5 workers total; 67 tasks left in task-pool. [ Info: running task 34 on process 73 (Schedulers-primary-eA3Kzwpz); 5 workers total; 65 tasks left in task-pool. [ Info: running task 35 on process 74 (Schedulers-primary-eA3Kzwpz); 5 workers total; 65 tasks left in task-pool. [ Info: running task 36 on process 70 (Schedulers-primary-eA3Kzwpz); 5 workers total; 64 tasks left in task-pool. [ Info: running task 37 on process 71 (Schedulers-primary-eA3Kzwpz); 5 workers total; 63 tasks left in task-pool. [ Info: running task 38 on process 72 (Schedulers-primary-eA3Kzwpz); 5 workers total; 62 tasks left in task-pool. [ Info: running task 39 on process 73 (Schedulers-primary-eA3Kzwpz); 5 workers total; 60 tasks left in task-pool. [ Info: running task 40 on process 74 (Schedulers-primary-eA3Kzwpz); 5 workers total; 60 tasks left in task-pool. [ Info: running task 41 on process 70 (Schedulers-primary-eA3Kzwpz); 5 workers total; 59 tasks left in task-pool. [ Info: running task 42 on process 71 (Schedulers-primary-eA3Kzwpz); 5 workers total; 58 tasks left in task-pool. [ Info: running task 43 on process 72 (Schedulers-primary-eA3Kzwpz); 5 workers total; 57 tasks left in task-pool. [ Info: running task 44 on process 73 (Schedulers-primary-eA3Kzwpz); 5 workers total; 55 tasks left in task-pool. [ Info: running task 45 on process 74 (Schedulers-primary-eA3Kzwpz); 5 workers total; 55 tasks left in task-pool. [ Info: running task 46 on process 70 (Schedulers-primary-eA3Kzwpz); 5 workers total; 54 tasks left in task-pool. [ Info: running task 47 on process 71 (Schedulers-primary-eA3Kzwpz); 5 workers total; 53 tasks left in task-pool. [ Info: running task 48 on process 72 (Schedulers-primary-eA3Kzwpz); 5 workers total; 52 tasks left in task-pool. [ Info: running task 49 on process 73 (Schedulers-primary-eA3Kzwpz); 5 workers total; 50 tasks left in task-pool. [ Info: running task 50 on process 74 (Schedulers-primary-eA3Kzwpz); 5 workers total; 50 tasks left in task-pool. [ Info: running task 51 on process 70 (Schedulers-primary-eA3Kzwpz); 5 workers total; 49 tasks left in task-pool. [ Info: running task 52 on process 71 (Schedulers-primary-eA3Kzwpz); 5 workers total; 48 tasks left in task-pool. [ Info: running task 53 on process 72 (Schedulers-primary-eA3Kzwpz); 5 workers total; 47 tasks left in task-pool. [ Info: running task 54 on process 73 (Schedulers-primary-eA3Kzwpz); 5 workers total; 45 tasks left in task-pool. [ Info: running task 55 on process 74 (Schedulers-primary-eA3Kzwpz); 5 workers total; 45 tasks left in task-pool. [ Info: running task 56 on process 70 (Schedulers-primary-eA3Kzwpz); 5 workers total; 44 tasks left in task-pool. [ Info: running task 57 on process 71 (Schedulers-primary-eA3Kzwpz); 5 workers total; 43 tasks left in task-pool. [ Info: running task 58 on process 72 (Schedulers-primary-eA3Kzwpz); 5 workers total; 42 tasks left in task-pool. [ Info: running task 59 on process 73 (Schedulers-primary-eA3Kzwpz); 5 workers total; 40 tasks left in task-pool. [ Info: running task 60 on process 74 (Schedulers-primary-eA3Kzwpz); 5 workers total; 40 tasks left in task-pool. [ Info: running task 61 on process 70 (Schedulers-primary-eA3Kzwpz); 5 workers total; 39 tasks left in task-pool. [ Info: running task 62 on process 71 (Schedulers-primary-eA3Kzwpz); 5 workers total; 38 tasks left in task-pool. [ Info: running task 63 on process 72 (Schedulers-primary-eA3Kzwpz); 5 workers total; 37 tasks left in task-pool. [ Info: running task 64 on process 73 (Schedulers-primary-eA3Kzwpz); 5 workers total; 35 tasks left in task-pool. [ Info: running task 65 on process 74 (Schedulers-primary-eA3Kzwpz); 5 workers total; 35 tasks left in task-pool. [ Info: running task 66 on process 70 (Schedulers-primary-eA3Kzwpz); 5 workers total; 34 tasks left in task-pool. [ Info: running task 67 on process 71 (Schedulers-primary-eA3Kzwpz); 5 workers total; 33 tasks left in task-pool. [ Info: running task 68 on process 72 (Schedulers-primary-eA3Kzwpz); 5 workers total; 32 tasks left in task-pool. [ Info: running task 69 on process 73 (Schedulers-primary-eA3Kzwpz); 5 workers total; 30 tasks left in task-pool. [ Info: running task 70 on process 74 (Schedulers-primary-eA3Kzwpz); 5 workers total; 30 tasks left in task-pool. [ Info: running task 71 on process 70 (Schedulers-primary-eA3Kzwpz); 5 workers total; 29 tasks left in task-pool. [ Info: running task 72 on process 71 (Schedulers-primary-eA3Kzwpz); 5 workers total; 28 tasks left in task-pool. [ Info: running task 73 on process 72 (Schedulers-primary-eA3Kzwpz); 5 workers total; 27 tasks left in task-pool. [ Info: running task 74 on process 73 (Schedulers-primary-eA3Kzwpz); 5 workers total; 25 tasks left in task-pool. [ Info: running task 75 on process 74 (Schedulers-primary-eA3Kzwpz); 5 workers total; 25 tasks left in task-pool. [ Info: running task 76 on process 70 (Schedulers-primary-eA3Kzwpz); 5 workers total; 24 tasks left in task-pool. [ Info: running task 77 on process 71 (Schedulers-primary-eA3Kzwpz); 5 workers total; 23 tasks left in task-pool. [ Info: running task 78 on process 72 (Schedulers-primary-eA3Kzwpz); 5 workers total; 22 tasks left in task-pool. [ Info: running task 79 on process 73 (Schedulers-primary-eA3Kzwpz); 5 workers total; 20 tasks left in task-pool. [ Info: running task 80 on process 74 (Schedulers-primary-eA3Kzwpz); 5 workers total; 20 tasks left in task-pool. [ Info: running task 81 on process 70 (Schedulers-primary-eA3Kzwpz); 5 workers total; 19 tasks left in task-pool. [ Info: running task 82 on process 71 (Schedulers-primary-eA3Kzwpz); 5 workers total; 18 tasks left in task-pool. [ Info: running task 83 on process 72 (Schedulers-primary-eA3Kzwpz); 5 workers total; 17 tasks left in task-pool. [ Info: running task 84 on process 73 (Schedulers-primary-eA3Kzwpz); 5 workers total; 16 tasks left in task-pool. [ Info: running task 85 on process 74 (Schedulers-primary-eA3Kzwpz); 5 workers total; 15 tasks left in task-pool. [ Info: running task 86 on process 70 (Schedulers-primary-eA3Kzwpz); 5 workers total; 14 tasks left in task-pool. [ Info: running task 87 on process 71 (Schedulers-primary-eA3Kzwpz); 5 workers total; 13 tasks left in task-pool. [ Info: running task 88 on process 72 (Schedulers-primary-eA3Kzwpz); 5 workers total; 12 tasks left in task-pool. [ Info: running task 89 on process 73 (Schedulers-primary-eA3Kzwpz); 5 workers total; 10 tasks left in task-pool. [ Info: running task 90 on process 74 (Schedulers-primary-eA3Kzwpz); 5 workers total; 10 tasks left in task-pool. [ Info: running task 91 on process 70 (Schedulers-primary-eA3Kzwpz); 5 workers total; 9 tasks left in task-pool. [ Info: running task 92 on process 71 (Schedulers-primary-eA3Kzwpz); 5 workers total; 8 tasks left in task-pool. [ Info: running task 93 on process 72 (Schedulers-primary-eA3Kzwpz); 5 workers total; 7 tasks left in task-pool. [ Info: running task 94 on process 73 (Schedulers-primary-eA3Kzwpz); 5 workers total; 6 tasks left in task-pool. [ Info: running task 95 on process 74 (Schedulers-primary-eA3Kzwpz); 5 workers total; 5 tasks left in task-pool. [ Info: running task 96 on process 70 (Schedulers-primary-eA3Kzwpz); 5 workers total; 4 tasks left in task-pool. [ Info: running task 97 on process 71 (Schedulers-primary-eA3Kzwpz); 5 workers total; 3 tasks left in task-pool. [ Info: running task 98 on process 72 (Schedulers-primary-eA3Kzwpz); 5 workers total; 2 tasks left in task-pool. [ Info: running task 99 on process 73 (Schedulers-primary-eA3Kzwpz); 5 workers total; 0 tasks left in task-pool. [ Info: running task 100 on process 74 (Schedulers-primary-eA3Kzwpz); 5 workers total; 0 tasks left in task-pool. [ Info: reducing from 5 checkpoints using process 70 (5 workers, 1 reduce workers). [ Info: reducing from 4 checkpoints using process 71 (5 workers, 2 reduce workers). [ Info: reducing from 2 checkpoints using process 70 (5 workers, 2 reduce workers). [ Info: reducing from 2 checkpoints using process 70 (5 workers, 2 reduce workers). Test Summary: | Pass Total Time pmapreduce, stable cluster test | 2 2 2m57.9s [ Info: running task 1 on process 75 (Schedulers-primary-eA3Kzwpz); 5 workers total; 99 tasks left in task-pool. [ Info: running task 2 on process 76 (Schedulers-primary-eA3Kzwpz); 5 workers total; 98 tasks left in task-pool. [ Info: running task 3 on process 77 (Schedulers-primary-eA3Kzwpz); 5 workers total; 97 tasks left in task-pool. [ Info: running task 4 on process 79 (Schedulers-primary-eA3Kzwpz); 5 workers total; 96 tasks left in task-pool. [ Info: running task 5 on process 76 (Schedulers-primary-eA3Kzwpz); 5 workers total; 95 tasks left in task-pool. [ Info: running task 6 on process 75 (Schedulers-primary-eA3Kzwpz); 5 workers total; 94 tasks left in task-pool. [ Info: running task 7 on process 77 (Schedulers-primary-eA3Kzwpz); 5 workers total; 93 tasks left in task-pool. [ Info: running task 8 on process 79 (Schedulers-primary-eA3Kzwpz); 5 workers total; 92 tasks left in task-pool. [ Info: running task 9 on process 76 (Schedulers-primary-eA3Kzwpz); 5 workers total; 91 tasks left in task-pool. [ Info: running task 10 on process 75 (Schedulers-primary-eA3Kzwpz); 5 workers total; 90 tasks left in task-pool. [ Info: running task 11 on process 77 (Schedulers-primary-eA3Kzwpz); 5 workers total; 89 tasks left in task-pool. [ Info: running task 12 on process 79 (Schedulers-primary-eA3Kzwpz); 5 workers total; 88 tasks left in task-pool. [ Info: running task 13 on process 76 (Schedulers-primary-eA3Kzwpz); 5 workers total; 87 tasks left in task-pool. [ Info: running task 14 on process 75 (Schedulers-primary-eA3Kzwpz); 5 workers total; 86 tasks left in task-pool. [ Info: running task 15 on process 77 (Schedulers-primary-eA3Kzwpz); 5 workers total; 85 tasks left in task-pool. [ Info: running task 16 on process 79 (Schedulers-primary-eA3Kzwpz); 5 workers total; 84 tasks left in task-pool. [ Info: running task 17 on process 76 (Schedulers-primary-eA3Kzwpz); 5 workers total; 83 tasks left in task-pool. [ Info: running task 18 on process 80 (Schedulers-primary-eA3Kzwpz); 5 workers total; 82 tasks left in task-pool. [ Info: running task 19 on process 75 (Schedulers-primary-eA3Kzwpz); 5 workers total; 81 tasks left in task-pool. [ Info: running task 20 on process 77 (Schedulers-primary-eA3Kzwpz); 5 workers total; 80 tasks left in task-pool. [ Info: running task 21 on process 79 (Schedulers-primary-eA3Kzwpz); 5 workers total; 79 tasks left in task-pool. [ Info: running task 22 on process 76 (Schedulers-primary-eA3Kzwpz); 5 workers total; 78 tasks left in task-pool. [ Info: running task 23 on process 80 (Schedulers-primary-eA3Kzwpz); 5 workers total; 77 tasks left in task-pool. [ Info: running task 24 on process 75 (Schedulers-primary-eA3Kzwpz); 5 workers total; 76 tasks left in task-pool. [ Info: running task 25 on process 77 (Schedulers-primary-eA3Kzwpz); 5 workers total; 75 tasks left in task-pool. [ Info: running task 26 on process 79 (Schedulers-primary-eA3Kzwpz); 5 workers total; 74 tasks left in task-pool. [ Info: running task 27 on process 76 (Schedulers-primary-eA3Kzwpz); 5 workers total; 73 tasks left in task-pool. [ Info: running task 28 on process 80 (Schedulers-primary-eA3Kzwpz); 5 workers total; 72 tasks left in task-pool. [ Info: running task 29 on process 75 (Schedulers-primary-eA3Kzwpz); 5 workers total; 71 tasks left in task-pool. [ Info: running task 30 on process 77 (Schedulers-primary-eA3Kzwpz); 5 workers total; 70 tasks left in task-pool. [ Info: running task 31 on process 79 (Schedulers-primary-eA3Kzwpz); 5 workers total; 69 tasks left in task-pool. [ Info: running task 32 on process 76 (Schedulers-primary-eA3Kzwpz); 5 workers total; 68 tasks left in task-pool. [ Info: running task 33 on process 80 (Schedulers-primary-eA3Kzwpz); 5 workers total; 67 tasks left in task-pool. [ Info: running task 34 on process 75 (Schedulers-primary-eA3Kzwpz); 5 workers total; 66 tasks left in task-pool. [ Info: running task 35 on process 77 (Schedulers-primary-eA3Kzwpz); 5 workers total; 65 tasks left in task-pool. [ Info: running task 36 on process 79 (Schedulers-primary-eA3Kzwpz); 5 workers total; 64 tasks left in task-pool. [ Info: running task 37 on process 76 (Schedulers-primary-eA3Kzwpz); 5 workers total; 63 tasks left in task-pool. [ Info: running task 38 on process 80 (Schedulers-primary-eA3Kzwpz); 5 workers total; 62 tasks left in task-pool. [ Info: running task 39 on process 75 (Schedulers-primary-eA3Kzwpz); 5 workers total; 61 tasks left in task-pool. [ Info: running task 40 on process 77 (Schedulers-primary-eA3Kzwpz); 5 workers total; 60 tasks left in task-pool. [ Info: running task 41 on process 79 (Schedulers-primary-eA3Kzwpz); 5 workers total; 59 tasks left in task-pool. [ Info: running task 42 on process 76 (Schedulers-primary-eA3Kzwpz); 5 workers total; 58 tasks left in task-pool. [ Info: running task 43 on process 80 (Schedulers-primary-eA3Kzwpz); 5 workers total; 57 tasks left in task-pool. [ Info: running task 44 on process 75 (Schedulers-primary-eA3Kzwpz); 5 workers total; 56 tasks left in task-pool. [ Info: running task 45 on process 77 (Schedulers-primary-eA3Kzwpz); 5 workers total; 55 tasks left in task-pool. [ Info: running task 46 on process 79 (Schedulers-primary-eA3Kzwpz); 5 workers total; 54 tasks left in task-pool. [ Info: running task 47 on process 76 (Schedulers-primary-eA3Kzwpz); 5 workers total; 53 tasks left in task-pool. [ Info: running task 48 on process 80 (Schedulers-primary-eA3Kzwpz); 5 workers total; 52 tasks left in task-pool. [ Info: running task 49 on process 75 (Schedulers-primary-eA3Kzwpz); 5 workers total; 51 tasks left in task-pool. [ Info: running task 50 on process 77 (Schedulers-primary-eA3Kzwpz); 5 workers total; 50 tasks left in task-pool. [ Info: running task 51 on process 79 (Schedulers-primary-eA3Kzwpz); 5 workers total; 49 tasks left in task-pool. [ Info: running task 52 on process 76 (Schedulers-primary-eA3Kzwpz); 5 workers total; 48 tasks left in task-pool. [ Info: running task 53 on process 80 (Schedulers-primary-eA3Kzwpz); 5 workers total; 47 tasks left in task-pool. [ Info: running task 54 on process 75 (Schedulers-primary-eA3Kzwpz); 5 workers total; 46 tasks left in task-pool. [ Info: running task 55 on process 77 (Schedulers-primary-eA3Kzwpz); 5 workers total; 45 tasks left in task-pool. [ Info: running task 56 on process 79 (Schedulers-primary-eA3Kzwpz); 5 workers total; 44 tasks left in task-pool. [ Info: running task 57 on process 76 (Schedulers-primary-eA3Kzwpz); 5 workers total; 43 tasks left in task-pool. [ Info: running task 58 on process 80 (Schedulers-primary-eA3Kzwpz); 5 workers total; 42 tasks left in task-pool. [ Info: running task 59 on process 75 (Schedulers-primary-eA3Kzwpz); 5 workers total; 41 tasks left in task-pool. [ Info: running task 60 on process 77 (Schedulers-primary-eA3Kzwpz); 5 workers total; 40 tasks left in task-pool. [ Info: running task 61 on process 79 (Schedulers-primary-eA3Kzwpz); 5 workers total; 39 tasks left in task-pool. [ Info: running task 62 on process 76 (Schedulers-primary-eA3Kzwpz); 5 workers total; 38 tasks left in task-pool. [ Info: running task 63 on process 80 (Schedulers-primary-eA3Kzwpz); 5 workers total; 37 tasks left in task-pool. [ Info: running task 64 on process 75 (Schedulers-primary-eA3Kzwpz); 5 workers total; 36 tasks left in task-pool. [ Info: running task 65 on process 77 (Schedulers-primary-eA3Kzwpz); 5 workers total; 35 tasks left in task-pool. [ Info: running task 66 on process 79 (Schedulers-primary-eA3Kzwpz); 5 workers total; 34 tasks left in task-pool. [ Info: running task 67 on process 76 (Schedulers-primary-eA3Kzwpz); 5 workers total; 33 tasks left in task-pool. [ Info: running task 68 on process 80 (Schedulers-primary-eA3Kzwpz); 5 workers total; 32 tasks left in task-pool. [ Info: running task 69 on process 75 (Schedulers-primary-eA3Kzwpz); 5 workers total; 31 tasks left in task-pool. [ Info: running task 70 on process 77 (Schedulers-primary-eA3Kzwpz); 5 workers total; 30 tasks left in task-pool. [ Info: running task 71 on process 79 (Schedulers-primary-eA3Kzwpz); 5 workers total; 29 tasks left in task-pool. [ Info: running task 72 on process 76 (Schedulers-primary-eA3Kzwpz); 5 workers total; 28 tasks left in task-pool. [ Info: running task 73 on process 80 (Schedulers-primary-eA3Kzwpz); 5 workers total; 27 tasks left in task-pool. [ Info: running task 74 on process 75 (Schedulers-primary-eA3Kzwpz); 5 workers total; 26 tasks left in task-pool. [ Info: running task 75 on process 77 (Schedulers-primary-eA3Kzwpz); 5 workers total; 25 tasks left in task-pool. [ Info: running task 76 on process 79 (Schedulers-primary-eA3Kzwpz); 5 workers total; 24 tasks left in task-pool. [ Info: running task 77 on process 76 (Schedulers-primary-eA3Kzwpz); 5 workers total; 23 tasks left in task-pool. [ Info: running task 78 on process 80 (Schedulers-primary-eA3Kzwpz); 5 workers total; 22 tasks left in task-pool. [ Info: running task 79 on process 75 (Schedulers-primary-eA3Kzwpz); 5 workers total; 21 tasks left in task-pool. [ Info: running task 80 on process 77 (Schedulers-primary-eA3Kzwpz); 5 workers total; 20 tasks left in task-pool. [ Info: running task 81 on process 79 (Schedulers-primary-eA3Kzwpz); 5 workers total; 19 tasks left in task-pool. [ Info: running task 82 on process 76 (Schedulers-primary-eA3Kzwpz); 5 workers total; 18 tasks left in task-pool. [ Info: running task 83 on process 80 (Schedulers-primary-eA3Kzwpz); 5 workers total; 17 tasks left in task-pool. [ Info: running task 84 on process 75 (Schedulers-primary-eA3Kzwpz); 5 workers total; 16 tasks left in task-pool. [ Info: running task 85 on process 77 (Schedulers-primary-eA3Kzwpz); 5 workers total; 15 tasks left in task-pool. [ Info: running task 86 on process 79 (Schedulers-primary-eA3Kzwpz); 5 workers total; 14 tasks left in task-pool. [ Info: running task 87 on process 76 (Schedulers-primary-eA3Kzwpz); 5 workers total; 13 tasks left in task-pool. [ Info: running task 88 on process 80 (Schedulers-primary-eA3Kzwpz); 5 workers total; 12 tasks left in task-pool. [ Info: running task 89 on process 75 (Schedulers-primary-eA3Kzwpz); 5 workers total; 11 tasks left in task-pool. [ Info: running task 90 on process 77 (Schedulers-primary-eA3Kzwpz); 5 workers total; 10 tasks left in task-pool. [ Info: running task 91 on process 79 (Schedulers-primary-eA3Kzwpz); 5 workers total; 9 tasks left in task-pool. [ Info: running task 92 on process 76 (Schedulers-primary-eA3Kzwpz); 5 workers total; 8 tasks left in task-pool. [ Info: running task 93 on process 80 (Schedulers-primary-eA3Kzwpz); 5 workers total; 7 tasks left in task-pool. [ Info: running task 94 on process 75 (Schedulers-primary-eA3Kzwpz); 5 workers total; 6 tasks left in task-pool. [ Info: running task 95 on process 77 (Schedulers-primary-eA3Kzwpz); 5 workers total; 5 tasks left in task-pool. [ Info: running task 96 on process 79 (Schedulers-primary-eA3Kzwpz); 5 workers total; 4 tasks left in task-pool. [ Info: running task 97 on process 76 (Schedulers-primary-eA3Kzwpz); 5 workers total; 3 tasks left in task-pool. [ Info: running task 98 on process 80 (Schedulers-primary-eA3Kzwpz); 5 workers total; 2 tasks left in task-pool. [ Info: running task 99 on process 75 (Schedulers-primary-eA3Kzwpz); 5 workers total; 1 tasks left in task-pool. [ Info: running task 100 on process 77 (Schedulers-primary-eA3Kzwpz); 5 workers total; 0 tasks left in task-pool. [ Info: reducing from 3 checkpoints using process 79 (5 workers, 1 reduce workers). [ Info: reducing from 4 checkpoints using process 75 (5 workers, 2 reduce workers). [ Info: reducing from 2 checkpoints using process 79 (5 workers, 2 reduce workers). [ Info: reducing from 2 checkpoints using process 79 (5 workers, 2 reduce workers). Test Summary: | Pass Total Time pmapreduce, cluster with ProcessExitedException during tasks | 4 4 3m29.6s [ Info: running task 1 on process 81 (Schedulers-primary-eA3Kzwpz); 5 workers total; 99 tasks left in task-pool. [ Info: running task 2 on process 82 (Schedulers-primary-eA3Kzwpz); 5 workers total; 98 tasks left in task-pool. [ Info: running task 3 on process 83 (Schedulers-primary-eA3Kzwpz); 5 workers total; 97 tasks left in task-pool. [ Info: running task 4 on process 84 (Schedulers-primary-eA3Kzwpz); 5 workers total; 95 tasks left in task-pool. [ Info: running task 5 on process 85 (Schedulers-primary-eA3Kzwpz); 5 workers total; 95 tasks left in task-pool. [ Info: running task 6 on process 81 (Schedulers-primary-eA3Kzwpz); 5 workers total; 94 tasks left in task-pool. [ Info: running task 7 on process 82 (Schedulers-primary-eA3Kzwpz); 5 workers total; 93 tasks left in task-pool. [ Info: running task 8 on process 83 (Schedulers-primary-eA3Kzwpz); 5 workers total; 92 tasks left in task-pool. [ Info: running task 9 on process 84 (Schedulers-primary-eA3Kzwpz); 5 workers total; 90 tasks left in task-pool. [ Info: running task 10 on process 85 (Schedulers-primary-eA3Kzwpz); 5 workers total; 90 tasks left in task-pool. [ Info: running task 11 on process 81 (Schedulers-primary-eA3Kzwpz); 5 workers total; 89 tasks left in task-pool. [ Info: running task 12 on process 82 (Schedulers-primary-eA3Kzwpz); 5 workers total; 88 tasks left in task-pool. [ Info: running task 13 on process 83 (Schedulers-primary-eA3Kzwpz); 5 workers total; 87 tasks left in task-pool. ┌ Warning: pid=81 (Schedulers-primary-eA3Kzwpz), checkpoint=/tmp/jl_7rsBvB/checkpoint-Iab0sT-322, task loop, caught exception during save_checkpoint └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1410 ┌ Warning: TaskFailedException │ │ nested task error: On worker 81: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] #default_threadpool_checkpoint_call#25 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:284 │ [5] default_threadpool_checkpoint_call │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:250 │ [6] #handle_msg##8 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:303 │ [7] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:70 │ [8] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:79 │ [9] #schedule_call##2 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:88 │ │ nested task error: foo,x=0.94477903715177 │ Stacktrace: │ [1] error │ @ ./error.jl:44 │ [2] test_save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:448 │ [3] save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1785 │ [4] #28 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:262 │ Stacktrace: │ [1] remotecall_wait(::Function, ::Distributed.Worker, ::Nothing, ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:507 │ [2] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:497 [inlined] │ [3] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:519 [inlined] │ [4] (::Schedulers.var"#34#35"{@Kwargs{}, Nothing, Schedulers.var"#140#141", Schedulers.var"#142#143", Int64, typeof(Schedulers.save_checkpoint), Int64, Tuple{typeof(test_save_checkpoint), typeof(fetch), String, Future, DataType}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:288 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] remotecall_wait_timeout(::Vector{Float64}, ::Int64, ::Float64, ::Nothing, ::Function, ::Function, ::Int64, ::typeof(Schedulers.save_checkpoint), ::Int64, ::typeof(test_save_checkpoint), ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:298 │ [5] remotecall_wait_timeout(::Vector{Float64}, ::Int64, ::Float64, ::Nothing, ::Function, ::Function, ::Int64, ::Function, ::Int64, ::typeof(test_save_checkpoint), ::Vararg{Any}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:287 │ [6] (::Schedulers.var"#134#135"{Vector{Float32}, @Kwargs{}, typeof(foo7b), Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", typeof(nworkers), Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, String}, Dict{String, Any}, SchedulerOptions{String}, Tuple{Int64, Int64}, Vector{Float64}, Vector{Float64}, Vector{Float64}, Vector{Any}, Dict{Int64, Future}, Int64})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1405 │ │ nested task error: On worker 81: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] #default_threadpool_checkpoint_call#25 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:284 │ [5] default_threadpool_checkpoint_call │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:250 │ [6] #handle_msg##8 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:303 │ [7] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:70 │ [8] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:79 │ [9] #schedule_call##2 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:88 │ │ nested task error: foo,x=0.94477903715177 │ Stacktrace: │ [1] error │ @ ./error.jl:44 │ [2] test_save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:448 │ [3] save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1785 │ [4] #28 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:262 │ Stacktrace: │ [1] remotecall_wait(::Function, ::Distributed.Worker, ::Nothing, ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:507 │ [2] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:497 [inlined] │ [3] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:519 [inlined] │ [4] (::Schedulers.var"#34#35"{@Kwargs{}, Nothing, Schedulers.var"#140#141", Schedulers.var"#142#143", Int64, typeof(Schedulers.save_checkpoint), Int64, Tuple{typeof(test_save_checkpoint), typeof(fetch), String, Future, DataType}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:288 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 ┌ Warning: too many failures on process with id=81 (Schedulers-primary-eA3Kzwpz), removing from process list └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:393 ┌ Warning: pid=84 (Schedulers-primary-eA3Kzwpz), checkpoint=/tmp/jl_7rsBvB/checkpoint-Iab0sT-320, task loop, caught exception during save_checkpoint └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1410 ┌ Warning: TaskFailedException │ │ nested task error: On worker 84: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] #default_threadpool_checkpoint_call#25 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:284 │ [5] default_threadpool_checkpoint_call │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:250 │ [6] #handle_msg##8 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:303 │ [7] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:70 │ [8] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:79 │ [9] #schedule_call##2 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:88 │ │ nested task error: foo,x=0.8328938480943109 │ Stacktrace: │ [1] error │ @ ./error.jl:44 │ [2] test_save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:448 │ [3] save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1785 │ [4] #28 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:262 │ Stacktrace: │ [1] remotecall_wait(::Function, ::Distributed.Worker, ::Nothing, ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:507 │ [2] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:497 [inlined] │ [3] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:519 [inlined] │ [4] (::Schedulers.var"#34#35"{@Kwargs{}, Nothing, Schedulers.var"#140#141", Schedulers.var"#142#143", Int64, typeof(Schedulers.save_checkpoint), Int64, Tuple{typeof(test_save_checkpoint), typeof(fetch), String, Future, DataType}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:288 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] remotecall_wait_timeout(::Vector{Float64}, ::Int64, ::Float64, ::Nothing, ::Function, ::Function, ::Int64, ::typeof(Schedulers.save_checkpoint), ::Int64, ::typeof(test_save_checkpoint), ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:298 │ [5] remotecall_wait_timeout(::Vector{Float64}, ::Int64, ::Float64, ::Nothing, ::Function, ::Function, ::Int64, ::Function, ::Int64, ::typeof(test_save_checkpoint), ::Vararg{Any}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:287 │ [6] (::Schedulers.var"#134#135"{Vector{Float32}, @Kwargs{}, typeof(foo7b), Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", typeof(nworkers), Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, String}, Dict{String, Any}, SchedulerOptions{String}, Tuple{Int64, Int64}, Vector{Float64}, Vector{Float64}, Vector{Float64}, Vector{Any}, Dict{Int64, Future}, Int64})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1405 │ │ nested task error: On worker 84: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] #default_threadpool_checkpoint_call#25 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:284 │ [5] default_threadpool_checkpoint_call │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:250 │ [6] #handle_msg##8 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:303 │ [7] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:70 │ [8] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:79 │ [9] #schedule_call##2 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:88 │ │ nested task error: foo,x=0.8328938480943109 │ Stacktrace: │ [1] error │ @ ./error.jl:44 │ [2] test_save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:448 │ [3] save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1785 │ [4] #28 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:262 │ Stacktrace: │ [1] remotecall_wait(::Function, ::Distributed.Worker, ::Nothing, ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:507 │ [2] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:497 [inlined] │ [3] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:519 [inlined] │ [4] (::Schedulers.var"#34#35"{@Kwargs{}, Nothing, Schedulers.var"#140#141", Schedulers.var"#142#143", Int64, typeof(Schedulers.save_checkpoint), Int64, Tuple{typeof(test_save_checkpoint), typeof(fetch), String, Future, DataType}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:288 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 ┌ Warning: too many failures on process with id=84 (Schedulers-primary-eA3Kzwpz), removing from process list └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:393 [ Info: running task 14 on process 82 (Schedulers-primary-eA3Kzwpz); 5 workers total; 88 tasks left in task-pool. ┌ Warning: pid=85 (Schedulers-primary-eA3Kzwpz), checkpoint=/tmp/jl_7rsBvB/checkpoint-Iab0sT-321, task loop, caught exception during save_checkpoint └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1410 ┌ Warning: TaskFailedException │ │ nested task error: On worker 85: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] #default_threadpool_checkpoint_call#25 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:284 │ [5] default_threadpool_checkpoint_call │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:250 │ [6] #handle_msg##8 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:303 │ [7] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:70 │ [8] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:79 │ [9] #schedule_call##2 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:88 │ │ nested task error: foo,x=0.8025441934444892 │ Stacktrace: │ [1] error │ @ ./error.jl:44 │ [2] test_save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:448 │ [3] save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1785 │ [4] #28 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:262 │ Stacktrace: │ [1] remotecall_wait(::Function, ::Distributed.Worker, ::Nothing, ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:507 │ [2] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:497 [inlined] │ [3] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:519 [inlined] │ [4] (::Schedulers.var"#34#35"{@Kwargs{}, Nothing, Schedulers.var"#140#141", Schedulers.var"#142#143", Int64, typeof(Schedulers.save_checkpoint), Int64, Tuple{typeof(test_save_checkpoint), typeof(fetch), String, Future, DataType}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:288 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] remotecall_wait_timeout(::Vector{Float64}, ::Int64, ::Float64, ::Nothing, ::Function, ::Function, ::Int64, ::typeof(Schedulers.save_checkpoint), ::Int64, ::typeof(test_save_checkpoint), ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:298 │ [5] remotecall_wait_timeout(::Vector{Float64}, ::Int64, ::Float64, ::Nothing, ::Function, ::Function, ::Int64, ::Function, ::Int64, ::typeof(test_save_checkpoint), ::Vararg{Any}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:287 │ [6] (::Schedulers.var"#134#135"{Vector{Float32}, @Kwargs{}, typeof(foo7b), Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", typeof(nworkers), Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, String}, Dict{String, Any}, SchedulerOptions{String}, Tuple{Int64, Int64}, Vector{Float64}, Vector{Float64}, Vector{Float64}, Vector{Any}, Dict{Int64, Future}, Int64})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1405 │ │ nested task error: On worker 85: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] #default_threadpool_checkpoint_call#25 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:284 │ [5] default_threadpool_checkpoint_call │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:250 │ [6] #handle_msg##8 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:303 │ [7] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:70 │ [8] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:79 │ [9] #schedule_call##2 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:88 │ │ nested task error: foo,x=0.8025441934444892 │ Stacktrace: │ [1] error │ @ ./error.jl:44 │ [2] test_save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:448 │ [3] save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1785 │ [4] #28 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:262 │ Stacktrace: │ [1] remotecall_wait(::Function, ::Distributed.Worker, ::Nothing, ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:507 │ [2] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:497 [inlined] │ [3] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:519 [inlined] │ [4] (::Schedulers.var"#34#35"{@Kwargs{}, Nothing, Schedulers.var"#140#141", Schedulers.var"#142#143", Int64, typeof(Schedulers.save_checkpoint), Int64, Tuple{typeof(test_save_checkpoint), typeof(fetch), String, Future, DataType}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:288 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 ┌ Warning: too many failures on process with id=85 (Schedulers-primary-eA3Kzwpz), removing from process list └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:393 [ Info: running task 15 on process 83 (Schedulers-primary-eA3Kzwpz); 2 workers total; 88 tasks left in task-pool. [ Info: running task 16 on process 82 (Schedulers-primary-eA3Kzwpz); 2 workers total; 87 tasks left in task-pool. [ Info: running task 17 on process 83 (Schedulers-primary-eA3Kzwpz); 2 workers total; 86 tasks left in task-pool. [ Info: running task 18 on process 82 (Schedulers-primary-eA3Kzwpz); 2 workers total; 85 tasks left in task-pool. ┌ Warning: pid=83 (Schedulers-primary-eA3Kzwpz), checkpoint=/tmp/jl_7rsBvB/checkpoint-Iab0sT-331, task loop, caught exception during save_checkpoint └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1410 ┌ Warning: TaskFailedException │ │ nested task error: On worker 83: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] #default_threadpool_checkpoint_call#25 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:284 │ [5] default_threadpool_checkpoint_call │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:250 │ [6] #handle_msg##8 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:303 │ [7] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:70 │ [8] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:79 │ [9] #schedule_call##2 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:88 │ │ nested task error: foo,x=0.9805083698224824 │ Stacktrace: │ [1] error │ @ ./error.jl:44 │ [2] test_save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:448 │ [3] save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1785 │ [4] #28 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:262 │ Stacktrace: │ [1] remotecall_wait(::Function, ::Distributed.Worker, ::Nothing, ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:507 │ [2] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:497 [inlined] │ [3] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:519 [inlined] │ [4] (::Schedulers.var"#34#35"{@Kwargs{}, Nothing, Schedulers.var"#140#141", Schedulers.var"#142#143", Int64, typeof(Schedulers.save_checkpoint), Int64, Tuple{typeof(test_save_checkpoint), typeof(fetch), String, Future, DataType}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:288 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] remotecall_wait_timeout(::Vector{Float64}, ::Int64, ::Float64, ::Nothing, ::Function, ::Function, ::Int64, ::typeof(Schedulers.save_checkpoint), ::Int64, ::typeof(test_save_checkpoint), ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:298 │ [5] remotecall_wait_timeout(::Vector{Float64}, ::Int64, ::Float64, ::Nothing, ::Function, ::Function, ::Int64, ::Function, ::Int64, ::typeof(test_save_checkpoint), ::Vararg{Any}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:287 │ [6] (::Schedulers.var"#134#135"{Vector{Float32}, @Kwargs{}, typeof(foo7b), Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", typeof(nworkers), Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, String}, Dict{String, Any}, SchedulerOptions{String}, Tuple{Int64, Int64}, Vector{Float64}, Vector{Float64}, Vector{Float64}, Vector{Any}, Dict{Int64, Future}, Int64})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1405 │ │ nested task error: On worker 83: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] #default_threadpool_checkpoint_call#25 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:284 │ [5] default_threadpool_checkpoint_call │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:250 │ [6] #handle_msg##8 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:303 │ [7] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:70 │ [8] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:79 │ [9] #schedule_call##2 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:88 │ │ nested task error: foo,x=0.9805083698224824 │ Stacktrace: │ [1] error │ @ ./error.jl:44 │ [2] test_save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:448 │ [3] save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1785 │ [4] #28 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:262 │ Stacktrace: │ [1] remotecall_wait(::Function, ::Distributed.Worker, ::Nothing, ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:507 │ [2] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:497 [inlined] │ [3] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:519 [inlined] │ [4] (::Schedulers.var"#34#35"{@Kwargs{}, Nothing, Schedulers.var"#140#141", Schedulers.var"#142#143", Int64, typeof(Schedulers.save_checkpoint), Int64, Tuple{typeof(test_save_checkpoint), typeof(fetch), String, Future, DataType}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:288 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 ┌ Warning: too many failures on process with id=83 (Schedulers-primary-eA3Kzwpz), removing from process list └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:393 ┌ Warning: pid=82 (Schedulers-primary-eA3Kzwpz), checkpoint=/tmp/jl_7rsBvB/checkpoint-Iab0sT-332, task loop, caught exception during save_checkpoint └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1410 ┌ Warning: TaskFailedException │ │ nested task error: On worker 82: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] #default_threadpool_checkpoint_call#25 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:284 │ [5] default_threadpool_checkpoint_call │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:250 │ [6] #handle_msg##8 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:303 │ [7] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:70 │ [8] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:79 │ [9] #schedule_call##2 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:88 │ │ nested task error: foo,x=0.8956201391971675 │ Stacktrace: │ [1] error │ @ ./error.jl:44 │ [2] test_save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:448 │ [3] save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1785 │ [4] #28 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:262 │ Stacktrace: │ [1] remotecall_wait(::Function, ::Distributed.Worker, ::Nothing, ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:507 │ [2] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:497 [inlined] │ [3] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:519 [inlined] │ [4] (::Schedulers.var"#34#35"{@Kwargs{}, Nothing, Schedulers.var"#140#141", Schedulers.var"#142#143", Int64, typeof(Schedulers.save_checkpoint), Int64, Tuple{typeof(test_save_checkpoint), typeof(fetch), String, Future, DataType}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:288 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] remotecall_wait_timeout(::Vector{Float64}, ::Int64, ::Float64, ::Nothing, ::Function, ::Function, ::Int64, ::typeof(Schedulers.save_checkpoint), ::Int64, ::typeof(test_save_checkpoint), ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:298 │ [5] remotecall_wait_timeout(::Vector{Float64}, ::Int64, ::Float64, ::Nothing, ::Function, ::Function, ::Int64, ::Function, ::Int64, ::typeof(test_save_checkpoint), ::Vararg{Any}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:287 │ [6] (::Schedulers.var"#134#135"{Vector{Float32}, @Kwargs{}, typeof(foo7b), Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", typeof(nworkers), Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, String}, Dict{String, Any}, SchedulerOptions{String}, Tuple{Int64, Int64}, Vector{Float64}, Vector{Float64}, Vector{Float64}, Vector{Any}, Dict{Int64, Future}, Int64})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1405 │ │ nested task error: On worker 82: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] #default_threadpool_checkpoint_call#25 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:284 │ [5] default_threadpool_checkpoint_call │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:250 │ [6] #handle_msg##8 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:303 │ [7] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:70 │ [8] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:79 │ [9] #schedule_call##2 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:88 │ │ nested task error: foo,x=0.8956201391971675 │ Stacktrace: │ [1] error │ @ ./error.jl:44 │ [2] test_save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:448 │ [3] save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1785 │ [4] #28 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:262 │ Stacktrace: │ [1] remotecall_wait(::Function, ::Distributed.Worker, ::Nothing, ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:507 │ [2] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:497 [inlined] │ [3] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:519 [inlined] │ [4] (::Schedulers.var"#34#35"{@Kwargs{}, Nothing, Schedulers.var"#140#141", Schedulers.var"#142#143", Int64, typeof(Schedulers.save_checkpoint), Int64, Tuple{typeof(test_save_checkpoint), typeof(fetch), String, Future, DataType}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:288 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 ┌ Warning: too many failures on process with id=82 (Schedulers-primary-eA3Kzwpz), removing from process list └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:393 [ Info: running task 19 on process 86 (Schedulers-primary-eA3Kzwpz); 5 workers total; 86 tasks left in task-pool. [ Info: running task 20 on process 88 (Schedulers-primary-eA3Kzwpz); 5 workers total; 85 tasks left in task-pool. [ Info: running task 21 on process 87 (Schedulers-primary-eA3Kzwpz); 5 workers total; 84 tasks left in task-pool. [ Info: running task 22 on process 86 (Schedulers-primary-eA3Kzwpz); 5 workers total; 83 tasks left in task-pool. [ Info: running task 23 on process 88 (Schedulers-primary-eA3Kzwpz); 5 workers total; 82 tasks left in task-pool. [ Info: running task 24 on process 87 (Schedulers-primary-eA3Kzwpz); 5 workers total; 81 tasks left in task-pool. [ Info: running task 25 on process 86 (Schedulers-primary-eA3Kzwpz); 5 workers total; 80 tasks left in task-pool. [ Info: running task 26 on process 88 (Schedulers-primary-eA3Kzwpz); 5 workers total; 79 tasks left in task-pool. [ Info: running task 27 on process 87 (Schedulers-primary-eA3Kzwpz); 5 workers total; 78 tasks left in task-pool. [ Info: running task 28 on process 86 (Schedulers-primary-eA3Kzwpz); 5 workers total; 77 tasks left in task-pool. [ Info: running task 29 on process 88 (Schedulers-primary-eA3Kzwpz); 5 workers total; 76 tasks left in task-pool. [ Info: running task 30 on process 87 (Schedulers-primary-eA3Kzwpz); 5 workers total; 75 tasks left in task-pool. [ Info: running task 31 on process 86 (Schedulers-primary-eA3Kzwpz); 5 workers total; 74 tasks left in task-pool. [ Info: running task 32 on process 88 (Schedulers-primary-eA3Kzwpz); 5 workers total; 73 tasks left in task-pool. [ Info: running task 33 on process 87 (Schedulers-primary-eA3Kzwpz); 5 workers total; 72 tasks left in task-pool. [ Info: running task 34 on process 86 (Schedulers-primary-eA3Kzwpz); 5 workers total; 71 tasks left in task-pool. ┌ Warning: pid=88 (Schedulers-primary-eA3Kzwpz), checkpoint=/tmp/jl_7rsBvB/checkpoint-Iab0sT-345, task loop, caught exception during save_checkpoint └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1410 ┌ Warning: TaskFailedException │ │ nested task error: On worker 88: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] #default_threadpool_checkpoint_call#25 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:284 │ [5] default_threadpool_checkpoint_call │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:250 │ [6] #handle_msg##8 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:303 │ [7] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:70 │ [8] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:79 │ [9] #schedule_call##2 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:88 │ │ nested task error: foo,x=0.9459786525562597 │ Stacktrace: │ [1] error │ @ ./error.jl:44 │ [2] test_save_checkpoint │ @ ./none:-1 [inlined] │ [3] test_save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:0 │ [4] save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1785 │ [5] #28 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:262 │ Stacktrace: │ [1] remotecall_wait(::Function, ::Distributed.Worker, ::Nothing, ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:507 │ [2] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:497 [inlined] │ [3] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:519 [inlined] │ [4] (::Schedulers.var"#34#35"{@Kwargs{}, Nothing, Schedulers.var"#140#141", Schedulers.var"#142#143", Int64, typeof(Schedulers.save_checkpoint), Int64, Tuple{typeof(test_save_checkpoint), typeof(fetch), String, Future, DataType}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:288 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] remotecall_wait_timeout(::Vector{Float64}, ::Int64, ::Float64, ::Nothing, ::Function, ::Function, ::Int64, ::typeof(Schedulers.save_checkpoint), ::Int64, ::typeof(test_save_checkpoint), ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:298 │ [5] remotecall_wait_timeout(::Vector{Float64}, ::Int64, ::Float64, ::Nothing, ::Function, ::Function, ::Int64, ::Function, ::Int64, ::typeof(test_save_checkpoint), ::Vararg{Any}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:287 │ [6] (::Schedulers.var"#134#135"{Vector{Float32}, @Kwargs{}, typeof(foo7b), Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", typeof(nworkers), Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, String}, Dict{String, Any}, SchedulerOptions{String}, Tuple{Int64, Int64}, Vector{Float64}, Vector{Float64}, Vector{Float64}, Vector{Any}, Dict{Int64, Future}, Int64})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1405 │ │ nested task error: On worker 88: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] #default_threadpool_checkpoint_call#25 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:284 │ [5] default_threadpool_checkpoint_call │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:250 │ [6] #handle_msg##8 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:303 │ [7] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:70 │ [8] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:79 │ [9] #schedule_call##2 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:88 │ │ nested task error: foo,x=0.9459786525562597 │ Stacktrace: │ [1] error │ @ ./error.jl:44 │ [2] test_save_checkpoint │ @ ./none:-1 [inlined] │ [3] test_save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:0 │ [4] save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1785 │ [5] #28 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:262 │ Stacktrace: │ [1] remotecall_wait(::Function, ::Distributed.Worker, ::Nothing, ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:507 │ [2] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:497 [inlined] │ [3] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:519 [inlined] │ [4] (::Schedulers.var"#34#35"{@Kwargs{}, Nothing, Schedulers.var"#140#141", Schedulers.var"#142#143", Int64, typeof(Schedulers.save_checkpoint), Int64, Tuple{typeof(test_save_checkpoint), typeof(fetch), String, Future, DataType}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:288 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 ┌ Warning: too many failures on process with id=88 (Schedulers-primary-eA3Kzwpz), removing from process list └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:393 [ Info: running task 35 on process 87 (Schedulers-primary-eA3Kzwpz); 5 workers total; 71 tasks left in task-pool. [ Info: running task 36 on process 86 (Schedulers-primary-eA3Kzwpz); 4 workers total; 70 tasks left in task-pool. [ Info: running task 37 on process 89 (Schedulers-primary-eA3Kzwpz); 4 workers total; 69 tasks left in task-pool. [ Info: running task 38 on process 90 (Schedulers-primary-eA3Kzwpz); 4 workers total; 68 tasks left in task-pool. [ Info: running task 39 on process 87 (Schedulers-primary-eA3Kzwpz); 4 workers total; 67 tasks left in task-pool. ┌ Warning: pid=86 (Schedulers-primary-eA3Kzwpz), checkpoint=/tmp/jl_7rsBvB/checkpoint-Iab0sT-350, task loop, caught exception during save_checkpoint └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1410 ┌ Warning: TaskFailedException │ │ nested task error: On worker 86: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] #default_threadpool_checkpoint_call#25 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:284 │ [5] default_threadpool_checkpoint_call │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:250 │ [6] #handle_msg##8 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:303 │ [7] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:70 │ [8] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:79 │ [9] #schedule_call##2 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:88 │ │ nested task error: foo,x=0.9787601185702006 │ Stacktrace: │ [1] error │ @ ./error.jl:44 │ [2] test_save_checkpoint │ @ ./none:-1 [inlined] │ [3] test_save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:0 │ [4] save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1785 │ [5] #28 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:262 │ Stacktrace: │ [1] remotecall_wait(::Function, ::Distributed.Worker, ::Nothing, ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:507 │ [2] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:497 [inlined] │ [3] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:519 [inlined] │ [4] (::Schedulers.var"#34#35"{@Kwargs{}, Nothing, Schedulers.var"#140#141", Schedulers.var"#142#143", Int64, typeof(Schedulers.save_checkpoint), Int64, Tuple{typeof(test_save_checkpoint), typeof(fetch), String, Future, DataType}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:288 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] remotecall_wait_timeout(::Vector{Float64}, ::Int64, ::Float64, ::Nothing, ::Function, ::Function, ::Int64, ::typeof(Schedulers.save_checkpoint), ::Int64, ::typeof(test_save_checkpoint), ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:298 │ [5] remotecall_wait_timeout(::Vector{Float64}, ::Int64, ::Float64, ::Nothing, ::Function, ::Function, ::Int64, ::Function, ::Int64, ::typeof(test_save_checkpoint), ::Vararg{Any}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:287 │ [6] (::Schedulers.var"#134#135"{Vector{Float32}, @Kwargs{}, typeof(foo7b), Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", typeof(nworkers), Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, String}, Dict{String, Any}, SchedulerOptions{String}, Tuple{Int64, Int64}, Vector{Float64}, Vector{Float64}, Vector{Float64}, Vector{Any}, Dict{Int64, Future}, Int64})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1405 │ │ nested task error: On worker 86: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] #default_threadpool_checkpoint_call#25 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:284 │ [5] default_threadpool_checkpoint_call │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:250 │ [6] #handle_msg##8 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:303 │ [7] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:70 │ [8] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:79 │ [9] #schedule_call##2 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:88 │ │ nested task error: foo,x=0.9787601185702006 │ Stacktrace: │ [1] error │ @ ./error.jl:44 │ [2] test_save_checkpoint │ @ ./none:-1 [inlined] │ [3] test_save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:0 │ [4] save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1785 │ [5] #28 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:262 │ Stacktrace: │ [1] remotecall_wait(::Function, ::Distributed.Worker, ::Nothing, ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:507 │ [2] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:497 [inlined] │ [3] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:519 [inlined] │ [4] (::Schedulers.var"#34#35"{@Kwargs{}, Nothing, Schedulers.var"#140#141", Schedulers.var"#142#143", Int64, typeof(Schedulers.save_checkpoint), Int64, Tuple{typeof(test_save_checkpoint), typeof(fetch), String, Future, DataType}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:288 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 ┌ Warning: too many failures on process with id=86 (Schedulers-primary-eA3Kzwpz), removing from process list └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:393 ┌ Warning: pid=87 (Schedulers-primary-eA3Kzwpz), checkpoint=/tmp/jl_7rsBvB/checkpoint-Iab0sT-351, task loop, caught exception during save_checkpoint └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1410 ┌ Warning: TaskFailedException │ │ nested task error: On worker 87: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] #default_threadpool_checkpoint_call#25 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:284 │ [5] default_threadpool_checkpoint_call │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:250 │ [6] #handle_msg##8 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:303 │ [7] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:70 │ [8] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:79 │ [9] #schedule_call##2 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:88 │ │ nested task error: foo,x=0.8023363059020172 │ Stacktrace: │ [1] error │ @ ./error.jl:44 │ [2] test_save_checkpoint │ @ ./none:-1 [inlined] │ [3] test_save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:0 │ [4] save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1785 │ [5] #28 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:262 │ Stacktrace: │ [1] remotecall_wait(::Function, ::Distributed.Worker, ::Nothing, ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:507 │ [2] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:497 [inlined] │ [3] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:519 [inlined] │ [4] (::Schedulers.var"#34#35"{@Kwargs{}, Nothing, Schedulers.var"#140#141", Schedulers.var"#142#143", Int64, typeof(Schedulers.save_checkpoint), Int64, Tuple{typeof(test_save_checkpoint), typeof(fetch), String, Future, DataType}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:288 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] remotecall_wait_timeout(::Vector{Float64}, ::Int64, ::Float64, ::Nothing, ::Function, ::Function, ::Int64, ::typeof(Schedulers.save_checkpoint), ::Int64, ::typeof(test_save_checkpoint), ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:298 │ [5] remotecall_wait_timeout(::Vector{Float64}, ::Int64, ::Float64, ::Nothing, ::Function, ::Function, ::Int64, ::Function, ::Int64, ::typeof(test_save_checkpoint), ::Vararg{Any}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:287 │ [6] (::Schedulers.var"#134#135"{Vector{Float32}, @Kwargs{}, typeof(foo7b), Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", typeof(nworkers), Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, String}, Dict{String, Any}, SchedulerOptions{String}, Tuple{Int64, Int64}, Vector{Float64}, Vector{Float64}, Vector{Float64}, Vector{Any}, Dict{Int64, Future}, Int64})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1405 │ │ nested task error: On worker 87: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] #default_threadpool_checkpoint_call#25 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:284 │ [5] default_threadpool_checkpoint_call │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:250 │ [6] #handle_msg##8 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:303 │ [7] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:70 │ [8] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:79 │ [9] #schedule_call##2 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:88 │ │ nested task error: foo,x=0.8023363059020172 │ Stacktrace: │ [1] error │ @ ./error.jl:44 │ [2] test_save_checkpoint │ @ ./none:-1 [inlined] │ [3] test_save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:0 │ [4] save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1785 │ [5] #28 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:262 │ Stacktrace: │ [1] remotecall_wait(::Function, ::Distributed.Worker, ::Nothing, ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:507 │ [2] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:497 [inlined] │ [3] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:519 [inlined] │ [4] (::Schedulers.var"#34#35"{@Kwargs{}, Nothing, Schedulers.var"#140#141", Schedulers.var"#142#143", Int64, typeof(Schedulers.save_checkpoint), Int64, Tuple{typeof(test_save_checkpoint), typeof(fetch), String, Future, DataType}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:288 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 ┌ Warning: too many failures on process with id=87 (Schedulers-primary-eA3Kzwpz), removing from process list └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:393 [ Info: running task 40 on process 89 (Schedulers-primary-eA3Kzwpz); 4 workers total; 68 tasks left in task-pool. ┌ Warning: pid=90 (Schedulers-primary-eA3Kzwpz), checkpoint=nothing, task loop, caught exception during save_checkpoint └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1410 ┌ Warning: TaskFailedException │ │ nested task error: On worker 90: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] #default_threadpool_checkpoint_call#25 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:284 │ [5] default_threadpool_checkpoint_call │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:250 │ [6] #handle_msg##8 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:303 │ [7] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:70 │ [8] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:79 │ [9] #schedule_call##2 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:88 │ │ nested task error: foo,x=0.8164347909907699 │ Stacktrace: │ [1] error │ @ ./error.jl:44 │ [2] test_save_checkpoint │ @ ./none:-1 [inlined] │ [3] test_save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:0 │ [4] save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1785 │ [5] #28 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:262 │ Stacktrace: │ [1] remotecall_wait(::Function, ::Distributed.Worker, ::Nothing, ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:507 │ [2] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:497 [inlined] │ [3] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:519 [inlined] │ [4] (::Schedulers.var"#34#35"{@Kwargs{}, Nothing, Schedulers.var"#140#141", Schedulers.var"#142#143", Int64, typeof(Schedulers.save_checkpoint), Int64, Tuple{typeof(test_save_checkpoint), typeof(fetch), String, Future, DataType}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:288 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] remotecall_wait_timeout(::Vector{Float64}, ::Int64, ::Float64, ::Nothing, ::Function, ::Function, ::Int64, ::typeof(Schedulers.save_checkpoint), ::Int64, ::typeof(test_save_checkpoint), ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:298 │ [5] remotecall_wait_timeout(::Vector{Float64}, ::Int64, ::Float64, ::Nothing, ::Function, ::Function, ::Int64, ::Function, ::Int64, ::typeof(test_save_checkpoint), ::Vararg{Any}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:287 │ [6] (::Schedulers.var"#134#135"{Vector{Float32}, @Kwargs{}, typeof(foo7b), Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", typeof(nworkers), Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, String}, Dict{String, Any}, SchedulerOptions{String}, Tuple{Int64, Int64}, Vector{Float64}, Vector{Float64}, Vector{Float64}, Vector{Any}, Dict{Int64, Future}, Int64})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1405 │ │ nested task error: On worker 90: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] #default_threadpool_checkpoint_call#25 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:284 │ [5] default_threadpool_checkpoint_call │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:250 │ [6] #handle_msg##8 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:303 │ [7] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:70 │ [8] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:79 │ [9] #schedule_call##2 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:88 │ │ nested task error: foo,x=0.8164347909907699 │ Stacktrace: │ [1] error │ @ ./error.jl:44 │ [2] test_save_checkpoint │ @ ./none:-1 [inlined] │ [3] test_save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:0 │ [4] save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1785 │ [5] #28 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:262 │ Stacktrace: │ [1] remotecall_wait(::Function, ::Distributed.Worker, ::Nothing, ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:507 │ [2] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:497 [inlined] │ [3] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:519 [inlined] │ [4] (::Schedulers.var"#34#35"{@Kwargs{}, Nothing, Schedulers.var"#140#141", Schedulers.var"#142#143", Int64, typeof(Schedulers.save_checkpoint), Int64, Tuple{typeof(test_save_checkpoint), typeof(fetch), String, Future, DataType}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:288 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 ┌ Warning: too many failures on process with id=90 (Schedulers-primary-eA3Kzwpz), removing from process list └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:393 ┌ Warning: pid=89 (Schedulers-primary-eA3Kzwpz), checkpoint=/tmp/jl_7rsBvB/checkpoint-Iab0sT-354, task loop, caught exception during save_checkpoint └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1410 ┌ Warning: TaskFailedException │ │ nested task error: On worker 89: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] #default_threadpool_checkpoint_call#25 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:284 │ [5] default_threadpool_checkpoint_call │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:250 │ [6] #handle_msg##8 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:303 │ [7] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:70 │ [8] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:79 │ [9] #schedule_call##2 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:88 │ │ nested task error: foo,x=0.8765120161421198 │ Stacktrace: │ [1] error │ @ ./error.jl:44 │ [2] test_save_checkpoint │ @ ./none:-1 [inlined] │ [3] test_save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:0 │ [4] save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1785 │ [5] #28 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:262 │ Stacktrace: │ [1] remotecall_wait(::Function, ::Distributed.Worker, ::Nothing, ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:507 │ [2] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:497 [inlined] │ [3] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:519 [inlined] │ [4] (::Schedulers.var"#34#35"{@Kwargs{}, Nothing, Schedulers.var"#140#141", Schedulers.var"#142#143", Int64, typeof(Schedulers.save_checkpoint), Int64, Tuple{typeof(test_save_checkpoint), typeof(fetch), String, Future, DataType}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:288 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] remotecall_wait_timeout(::Vector{Float64}, ::Int64, ::Float64, ::Nothing, ::Function, ::Function, ::Int64, ::typeof(Schedulers.save_checkpoint), ::Int64, ::typeof(test_save_checkpoint), ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:298 │ [5] remotecall_wait_timeout(::Vector{Float64}, ::Int64, ::Float64, ::Nothing, ::Function, ::Function, ::Int64, ::Function, ::Int64, ::typeof(test_save_checkpoint), ::Vararg{Any}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:287 │ [6] (::Schedulers.var"#134#135"{Vector{Float32}, @Kwargs{}, typeof(foo7b), Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", typeof(nworkers), Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, String}, Dict{String, Any}, SchedulerOptions{String}, Tuple{Int64, Int64}, Vector{Float64}, Vector{Float64}, Vector{Float64}, Vector{Any}, Dict{Int64, Future}, Int64})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1405 │ │ nested task error: On worker 89: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] #default_threadpool_checkpoint_call#25 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:284 │ [5] default_threadpool_checkpoint_call │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:250 │ [6] #handle_msg##8 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:303 │ [7] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:70 │ [8] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:79 │ [9] #schedule_call##2 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:88 │ │ nested task error: foo,x=0.8765120161421198 │ Stacktrace: │ [1] error │ @ ./error.jl:44 │ [2] test_save_checkpoint │ @ ./none:-1 [inlined] │ [3] test_save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:0 │ [4] save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1785 │ [5] #28 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:262 │ Stacktrace: │ [1] remotecall_wait(::Function, ::Distributed.Worker, ::Nothing, ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:507 │ [2] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:497 [inlined] │ [3] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:519 [inlined] │ [4] (::Schedulers.var"#34#35"{@Kwargs{}, Nothing, Schedulers.var"#140#141", Schedulers.var"#142#143", Int64, typeof(Schedulers.save_checkpoint), Int64, Tuple{typeof(test_save_checkpoint), typeof(fetch), String, Future, DataType}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:288 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 ┌ Warning: too many failures on process with id=89 (Schedulers-primary-eA3Kzwpz), removing from process list └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:393 [ Info: running task 41 on process 91 (Schedulers-primary-eA3Kzwpz); 5 workers total; 69 tasks left in task-pool. [ Info: running task 42 on process 91 (Schedulers-primary-eA3Kzwpz); 5 workers total; 68 tasks left in task-pool. [ Info: running task 43 on process 91 (Schedulers-primary-eA3Kzwpz); 5 workers total; 67 tasks left in task-pool. [ Info: running task 44 on process 91 (Schedulers-primary-eA3Kzwpz); 5 workers total; 66 tasks left in task-pool. [ Info: running task 45 on process 91 (Schedulers-primary-eA3Kzwpz); 5 workers total; 65 tasks left in task-pool. [ Info: running task 46 on process 91 (Schedulers-primary-eA3Kzwpz); 5 workers total; 64 tasks left in task-pool. From worker 95: From worker 95: ====================================================================================== From worker 95: Information request received. A stacktrace will print followed by a 1.0 second profile From worker 95: ====================================================================================== From worker 95: From worker 95: cmd: /opt/julia/bin/julia 384 running 1 of 1 From worker 95: From worker 95: signal (10): User defined signal 1 From worker 95: invalidate_mt_cache at /source/src/gf.c:2414 From worker 95: jl_method_table_activate at /source/src/gf.c:2790 From worker 95: jl_activate_methods at /source/src/staticdata_utils.c:729 [inlined] From worker 95: jl_restore_package_image_from_stream at /source/src/staticdata.c:4439 From worker 95: jl_restore_incremental_from_buf at /source/src/staticdata.c:4481 From worker 95: ijl_restore_package_image_from_file at /source/src/staticdata.c:4548 From worker 95: #_include_from_serialized#869 at ./loading.jl:1270 From worker 95: _include_from_serialized at ./loading.jl:1246 [inlined] From worker 95: #_require_search_from_serialized#885 at ./loading.jl:2075 From worker 95: _require_search_from_serialized at ./loading.jl:1982 From worker 95: jfptr__require_search_from_serialized_57975.1 at /opt/julia/lib/julia/sys.so (unknown line) From worker 95: _jl_invoke at /source/src/gf.c:3789 [inlined] From worker 95: ijl_apply_generic at /source/src/gf.c:3989 From worker 95: __require_prelocked at ./loading.jl:2600 ┌ Warning: pid=91 (Schedulers-primary-eA3Kzwpz), checkpoint=/tmp/jl_7rsBvB/checkpoint-Iab0sT-361, task loop, caught exception during save_checkpoint └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1410 ┌ Warning: TaskFailedException │ │ nested task error: On worker 91: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] #default_threadpool_checkpoint_call#25 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:284 │ [5] default_threadpool_checkpoint_call │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:250 │ [6] #handle_msg##8 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:303 │ [7] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:70 │ [8] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:79 │ [9] #schedule_call##2 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:88 │ │ nested task error: foo,x=0.9309603465806658 │ Stacktrace: │ [1] error │ @ ./error.jl:44 │ [2] test_save_checkpoint │ @ ./none:-1 [inlined] │ [3] test_save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:0 │ [4] save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1785 │ [5] #28 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:262 │ Stacktrace: │ [1] remotecall_wait(::Function, ::Distributed.Worker, ::Nothing, ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:507 │ [2] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:497 [inlined] │ [3] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:519 [inlined] │ [4] (::Schedulers.var"#34#35"{@Kwargs{}, Nothing, Schedulers.var"#140#141", Schedulers.var"#142#143", Int64, typeof(Schedulers.save_checkpoint), Int64, Tuple{typeof(test_save_checkpoint), typeof(fetch), String, Future, DataType}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:288 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] remotecall_wait_timeout(::Vector{Float64}, ::Int64, ::Float64, ::Nothing, ::Function, ::Function, ::Int64, ::typeof(Schedulers.save_checkpoint), ::Int64, ::typeof(test_save_checkpoint), ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:298 │ [5] remotecall_wait_timeout(::Vector{Float64}, ::Int64, ::Float64, ::Nothing, ::Function, ::Function, ::Int64, ::Function, ::Int64, ::typeof(test_save_checkpoint), ::Vararg{Any}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:287 │ [6] (::Schedulers.var"#134#135"{Vector{Float32}, @Kwargs{}, typeof(foo7b), Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", typeof(nworkers), Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, String}, Dict{String, Any}, SchedulerOptions{String}, Tuple{Int64, Int64}, Vector{Float64}, Vector{Float64}, Vector{Float64}, Vector{Any}, Dict{Int64, Future}, Int64})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1405 │ │ nested task error: On worker 91: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] #default_threadpool_checkpoint_call#25 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:284 │ [5] default_threadpool_checkpoint_call │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:250 │ [6] #handle_msg##8 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:303 │ [7] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:70 │ [8] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:79 │ [9] #schedule_call##2 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:88 │ │ nested task error: foo,x=0.9309603465806658 │ Stacktrace: │ [1] error │ @ ./error.jl:44 │ [2] test_save_checkpoint │ @ ./none:-1 [inlined] │ [3] test_save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:0 │ [4] save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1785 │ [5] #28 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:262 │ Stacktrace: │ [1] remotecall_wait(::Function, ::Distributed.Worker, ::Nothing, ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:507 │ [2] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:497 [inlined] │ [3] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:519 [inlined] │ [4] (::Schedulers.var"#34#35"{@Kwargs{}, Nothing, Schedulers.var"#140#141", Schedulers.var"#142#143", Int64, typeof(Schedulers.save_checkpoint), Int64, Tuple{typeof(test_save_checkpoint), typeof(fetch), String, Future, DataType}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:288 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 ┌ Warning: too many failures on process with id=91 (Schedulers-primary-eA3Kzwpz), removing from process list └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:393 From worker 95: _require_prelocked at ./loading.jl:2466 From worker 95: macro expansion at ./loading.jl:2394 [inlined] From worker 95: macro expansion at ./lock.jl:376 [inlined] From worker 95: __require at ./loading.jl:2359 From worker 95: jfptr___require_71428.1 at /opt/julia/lib/julia/sys.so (unknown line) From worker 95: _jl_invoke at /source/src/gf.c:3789 [inlined] From worker 95: ijl_apply_generic at /source/src/gf.c:3989 From worker 95: jl_apply at /source/src/julia.h:2345 [inlined] From worker 95: jl_f_invoke_in_world at /source/src/builtins.c:896 From worker 95: require at ./loading.jl:2335 [inlined] From worker 95: eval_import_path at ./module.jl:36 [inlined] From worker 95: eval_import_path_all at ./module.jl:60 From worker 95: _eval_using at ./module.jl:137 From worker 95: jfptr__eval_using_68366.1 at /opt/julia/lib/julia/sys.so (unknown line) From worker 95: _jl_invoke at /source/src/gf.c:3789 [inlined] From worker 95: ijl_apply_generic at /source/src/gf.c:3989 From worker 95: jl_apply at /source/src/julia.h:2345 [inlined] From worker 95: do_call at /source/src/interpreter.c:123 From worker 95: eval_value at /source/src/interpreter.c:243 From worker 95: eval_stmt_value at /source/src/interpreter.c:194 [inlined] From worker 95: eval_body at /source/src/interpreter.c:708 From worker 95: jl_interpret_toplevel_thunk at /source/src/interpreter.c:899 From worker 95: jl_toplevel_eval_flex at /source/src/toplevel.c:773 From worker 95: ijl_toplevel_eval at /source/src/toplevel.c:785 From worker 95: ijl_toplevel_eval_in at /source/src/toplevel.c:830 From worker 95: eval at ./boot.jl:489 From worker 95: eval at ./Base_compiler.jl:149 From worker 95: unknown function (ip: 0x7c4a74f3bf46) at (unknown file) From worker 95: _jl_invoke at /source/src/gf.c:3789 [inlined] From worker 95: ijl_apply_generic at /source/src/gf.c:3989 From worker 95: jl_apply at /source/src/julia.h:2345 [inlined] From worker 95: jl_f_invokelatest at /source/src/builtins.c:877 From worker 95: _jl_invoke at /source/src/gf.c:3789 [inlined] From worker 95: ijl_apply_generic at /source/src/gf.c:3989 From worker 95: jl_apply at /source/src/julia.h:2345 [inlined] From worker 95: jl_f__apply_iterate at /source/src/builtins.c:864 From worker 95: #handle_msg##4 at /source/usr/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:287 From worker 95: run_work_thunk at /source/usr/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:70 From worker 95: #handle_msg##2 at /source/usr/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:287 From worker 95: unknown function (ip: 0x7c4a74f3bce7) at (unknown file) From worker 95: _jl_invoke at /source/src/gf.c:3789 [inlined] From worker 95: ijl_apply_generic at /source/src/gf.c:3989 From worker 95: jl_apply at /source/src/julia.h:2345 [inlined] From worker 95: start_task at /source/src/task.c:1249 From worker 95: unknown function (ip: (nil)) at (unknown file) From worker 95: From worker 95: ============================================================== From worker 95: Profile collected. A report will print at the next yield point From worker 95: ============================================================== From worker 95: From worker 94: From worker 94: ====================================================================================== From worker 94: Information request received. A stacktrace will print followed by a 1.0 second profile From worker 94: ====================================================================================== From worker 94: From worker 94: cmd: /opt/julia/bin/julia 383 running 1 of 1 From worker 94: From worker 94: signal (10): User defined signal 1 From worker 94: _ZN4llvm13isNoAliasCallEPKNS_5ValueE at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) From worker 94: _ZN4llvm18isIdentifiedObjectEPKNS_5ValueE at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) From worker 94: _ZN4llvm13BasicAAResult10aliasCheckEPKNS_5ValueENS_12LocationSizeES3_S4_RNS_11AAQueryInfoEPKNS_11InstructionE at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) From worker 94: _ZN4llvm13BasicAAResult5aliasERKNS_14MemoryLocationES3_RNS_11AAQueryInfoEPKNS_11InstructionE at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) From worker 94: _ZN4llvm9AAResults5aliasERKNS_14MemoryLocationES3_RNS_11AAQueryInfoEPKNS_11InstructionE at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) From worker 94: _ZN4llvm23MemoryDependenceResults30getSimplePointerDependencyFromERKNS_14MemoryLocationEbNS_21ilist_iterator_w_bitsINS_12ilist_detail12node_optionsINS_11InstructionELb1ELb0EvLb1ENS_10BasicBlockEEELb0ELb0EEEPS8_PS7_PjRNS_14BatchAAResultsE at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) From worker 94: _ZN4llvm23MemoryDependenceResults24getPointerDependencyFromERKNS_14MemoryLocationEbNS_21ilist_iterator_w_bitsINS_12ilist_detail12node_optionsINS_11InstructionELb1ELb0EvLb1ENS_10BasicBlockEEELb0ELb0EEEPS8_PS7_PjRNS_14BatchAAResultsE at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) From worker 94: _ZN4llvm23MemoryDependenceResults23getNonLocalInfoForBlockEPNS_11InstructionERKNS_14MemoryLocationEbPNS_10BasicBlockEPSt6vectorINS_16NonLocalDepEntryESaIS9_EEjRNS_14BatchAAResultsE at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) From worker 94: _ZN4llvm23MemoryDependenceResults27getNonLocalPointerDepFromBBEPNS_11InstructionERKNS_12PHITransAddrERKNS_14MemoryLocationEbPNS_10BasicBlockERNS_15SmallVectorImplINS_17NonLocalDepResultEEERNS_13SmallDenseMapISA_PNS_5ValueELj16ENS_12DenseMapInfoISA_vEENS_6detail12DenseMapPairISA_SH_EEEEbb at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) From worker 94: _ZN4llvm23MemoryDependenceResults28getNonLocalPointerDependencyEPNS_11InstructionERNS_15SmallVectorImplINS_17NonLocalDepResultEEE at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) From worker 94: _ZN4llvm7GVNPass19processNonLocalLoadEPNS_8LoadInstE.part.0 at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) From worker 94: _ZN4llvm7GVNPass18processInstructionEPNS_11InstructionE.part.0 at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) From worker 94: _ZN4llvm7GVNPass12processBlockEPNS_10BasicBlockE.part.0 at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) From worker 94: _ZN4llvm7GVNPass17iterateOnFunctionERNS_8FunctionE at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) From worker 94: _ZN4llvm7GVNPass7runImplERNS_8FunctionERNS_15AssumptionCacheERNS_13DominatorTreeERKNS_17TargetLibraryInfoERNS_9AAResultsEPNS_23MemoryDependenceResultsERNS_8LoopInfoEPNS_25OptimizationRemarkEmitterEPNS_9MemorySSAE at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) From worker 94: _ZN4llvm7GVNPass3runERNS_8FunctionERNS_15AnalysisManagerIS1_JEEE at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) From worker 94: run at /source/usr/include/llvm/IR/PassManagerInternal.h:91 From worker 94: _ZN4llvm11PassManagerINS_8FunctionENS_15AnalysisManagerIS1_JEEEJEE3runERS1_RS3_ at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) From worker 94: run at /source/usr/include/llvm/IR/PassManagerInternal.h:91 From worker 94: _ZN4llvm27ModuleToFunctionPassAdaptor3runERNS_6ModuleERNS_15AnalysisManagerIS1_JEEE at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) From worker 94: run at /source/usr/include/llvm/IR/PassManagerInternal.h:91 From worker 94: _ZN4llvm11PassManagerINS_6ModuleENS_15AnalysisManagerIS1_JEEEJEE3runERS1_RS3_ at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) From worker 94: run at /source/src/pipeline.cpp:791 From worker 94: operator() at /source/src/jitlayers.cpp:1518 From worker 94: withModuleDo<(anonymous namespace)::sizedOptimizerT::operator()(llvm::orc::ThreadSafeModule) [with long unsigned int N = 4]:: > at /source/usr/include/llvm/ExecutionEngine/Orc/ThreadSafeModule.h:136 [inlined] From worker 94: operator() at /source/src/jitlayers.cpp:1479 [inlined] From worker 94: operator() at /source/src/jitlayers.cpp:1631 [inlined] From worker 94: addModule at /source/src/jitlayers.cpp:2088 From worker 94: jl_compile_codeinst_now at /source/src/jitlayers.cpp:683 From worker 94: jl_compile_codeinst_impl at /source/src/jitlayers.cpp:883 From worker 94: jl_compile_method_internal at /source/src/gf.c:3310 From worker 94: _jl_invoke at /source/src/gf.c:3781 [inlined] From worker 94: ijl_apply_generic at /source/src/gf.c:3989 From worker 94: handle_deserialize at /source/usr/share/julia/stdlib/v1.13/Serialization/src/Serialization.jl:994 From worker 94: deserialize at /source/usr/share/julia/stdlib/v1.13/Serialization/src/Serialization.jl:848 [inlined] From worker 94: deserialize at /source/usr/share/julia/stdlib/v1.13/Serialization/src/Serialization.jl:1081 From worker 94: unknown function (ip: 0x791e8c6e6392) at (unknown file) From worker 94: _jl_invoke at /source/src/gf.c:3789 [inlined] From worker 94: ijl_apply_generic at /source/src/gf.c:3989 From worker 94: handle_deserialize at /source/usr/share/julia/stdlib/v1.13/Serialization/src/Serialization.jl:994 From worker 94: deserialize at /source/usr/share/julia/stdlib/v1.13/Serialization/src/Serialization.jl:848 From worker 94: #deserialize_tuple##0 at /source/usr/share/julia/stdlib/v1.13/Serialization/src/Serialization.jl:1007 From worker 94: ntupleany at ./ntuple.jl:45 From worker 94: deserialize_tuple at /source/usr/share/julia/stdlib/v1.13/Serialization/src/Serialization.jl:1007 From worker 94: handle_deserialize at /source/usr/share/julia/stdlib/v1.13/Serialization/src/Serialization.jl:891 From worker 94: deserialize at /source/usr/share/julia/stdlib/v1.13/Serialization/src/Serialization.jl:848 [inlined] From worker 94: deserialize_msg at /source/usr/share/julia/stdlib/v1.13/Distributed/src/messages.jl:87 From worker 94: unknown function (ip: 0x791e8c6bbc13) at (unknown file) From worker 94: _jl_invoke at /source/src/gf.c:3789 [inlined] From worker 94: ijl_apply_generic at /source/src/gf.c:3989 From worker 94: jl_apply at /source/src/julia.h:2345 [inlined] From worker 94: jl_f_invokelatest at /source/src/builtins.c:877 From worker 94: message_handler_loop at /source/usr/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:176 From worker 94: process_tcp_streams at /source/usr/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:133 From worker 94: #process_messages##0 at /source/usr/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:121 From worker 94: unknown function (ip: 0x791e8c6b5f95) at (unknown file) From worker 94: _jl_invoke at /source/src/gf.c:3789 [inlined] From worker 94: ijl_apply_generic at /source/src/gf.c:3989 From worker 94: jl_apply at /source/src/julia.h:2345 [inlined] From worker 94: start_task at /source/src/task.c:1249 From worker 94: unknown function (ip: (nil)) at (unknown file) From worker 94: From worker 94: ============================================================== From worker 94: Profile collected. A report will print at the next yield point From worker 94: ============================================================== From worker 94: From worker 95: ┌ Warning: There were no samples collected in one or more groups. From worker 95: │ This may be due to idle threads, or you may need to run your From worker 95: │ program longer (perhaps by running it multiple times), From worker 95: │ or adjust the delay between samples with `Profile.init()`. From worker 95: └ @ Profile /opt/julia/share/julia/stdlib/v1.13/Profile/src/Profile.jl:1362 From worker 95: Overhead ╎ [+additional indent] Count File:Line Function From worker 95: ========================================================= From worker 95: Thread 1 (default) Task 0x00007c4a551a5000 Total snapshots: 1. Utilization: 100% From worker 95: ╎1 @Distributed/…messages.jl:287 (::Distributed.var"#handle_msg##2#handle_msg… From worker 95: ╎ 1 @Distributed/…messages.jl:70 run_work_thunk(thunk::Distributed.var"#handl… From worker 95: ╎ 1 @Distributed/…essages.jl:287 (::Distributed.var"#handle_msg##4#handle_ms… From worker 95: ╎ 1 @Base/Base_compiler.jl:149 eval(m::Module, x::Expr) From worker 95: ╎ 1 @Base/boot.jl:489 eval(m::Module, e::Any) From worker 95: ╎ 1 @Base/module.jl:137 _eval_using(to::Module, path::Expr) From worker 95: ╎ ╎ 1 @Base/module.jl:60 eval_import_path_all(at::Module, path::Expr, keyw… From worker 95: ╎ ╎ 1 @Base/module.jl:36 eval_import_path From worker 95: ╎ ╎ 1 @Base/loading.jl:2335 require From worker 95: ╎ ╎ 1 @Base/loading.jl:2359 __require(into::Module, mod::Symbol) From worker 95: ╎ ╎ 1 @Base/lock.jl:376 macro expansion From worker 95: ╎ ╎ ╎ 1 @Base/loading.jl:2394 macro expansion From worker 95: ╎ ╎ ╎ 1 @Base/loading.jl:2466 _require_prelocked(uuidkey::Base.PkgId, … From worker 95: ╎ ╎ ╎ 1 @Base/loading.jl:2600 __require_prelocked(pkg::Base.PkgId, en… From worker 95: ╎ ╎ ╎ 1 @Base/loading.jl:1982 kwcall(::@NamedTuple{reasons::Dict{Str… From worker 95: ╎ ╎ ╎ 1 @Base/loading.jl:2075 _require_search_from_serialized(pkg::… From worker 95: ╎ ╎ ╎ ╎ 1 @Base/loading.jl:1246 _include_from_serialized From worker 95: ╎ ╎ ╎ ╎ 1 @Base/loading.jl:1290 _include_from_serialized(pkg::Base.… From worker 95: ╎ ╎ ╎ ╎ 1 @Base/…aticdata.jl:23 insert_backedges From worker 95: ╎ ╎ ╎ ╎ 1 @Base/…aticdata.jl:30 _insert_backedges From worker 95: ╎ ╎ ╎ ╎ 1 @Base/…ticdata.jl:33 _insert_backedges(edges::Vector{A… From worker 95: ╎ ╎ ╎ ╎ ╎ 1 @Base/…ticdata.jl:59 verify_method_graph(codeinst::Co… From worker 95: ╎ ╎ ╎ ╎ ╎ 1 @Base/…icdata.jl:171 verify_method(codeinst::Core.Co… From worker 95: ╎ ╎ ╎ ╎ ╎ 1 @Base/…icdata.jl:320 verify_call From worker 95: ╎ ╎ ╎ ╎ ╎ 1 @Base/…rnals.jl:1562 _methods_by_ftype From worker 95: From worker 93: From worker 93: ====================================================================================== From worker 93: Information request received. A stacktrace will print followed by a 1.0 second profile From worker 93: ====================================================================================== From worker 93: From worker 93: cmd: /opt/julia/bin/julia 380 running 0 of 1 From worker 93: From worker 93: signal (10): User defined signal 1 From worker 93: epoll_pwait at /lib/x86_64-linux-gnu/libc.so.6 (unknown line) From worker 93: uv__io_poll at /workspace/srcdir/libuv/src/unix/linux.c:1404 From worker 93: uv_run at /workspace/srcdir/libuv/src/unix/core.c:430 From worker 93: ijl_task_get_next at /source/src/scheduler.c:457 [ Info: running task 47 on process 92 (Schedulers-primary-eA3Kzwpz); 4 workers total; 64 tasks left in task-pool. [ Info: running task 48 on process 93 (Schedulers-primary-eA3Kzwpz); 4 workers total; 63 tasks left in task-pool. From worker 93: wait at ./task.jl:1213 From worker 93: wait_forever at ./task.jl:1150 From worker 93: jfptr_wait_forever_41430.1 at /opt/julia/lib/julia/sys.so (unknown line) From worker 93: _jl_invoke at /source/src/gf.c:3789 [inlined] From worker 93: ijl_apply_generic at /source/src/gf.c:3989 From worker 93: jl_apply at /source/src/julia.h:2345 [inlined] From worker 93: start_task at /source/src/task.c:1249 From worker 93: unknown function (ip: (nil)) at (unknown file) From worker 93: From worker 93: ============================================================== From worker 93: Profile collected. A report will print at the next yield point From worker 93: ============================================================== From worker 93: From worker 94: ┌ Warning: There were no samples collected in one or more groups. From worker 94: │ This may be due to idle threads, or you may need to run your From worker 94: │ program longer (perhaps by running it multiple times), From worker 94: │ or adjust the delay between samples with `Profile.init()`. From worker 94: └ @ Profile /opt/julia/share/julia/stdlib/v1.13/Profile/src/Profile.jl:1362 From worker 94: Overhead ╎ [+additional indent] Count File:Line Function From worker 94: ========================================================= From worker 94: Thread 1 (default) Task 0x0000791e77c70e20 Total snapshots: 1. Utilization: 100% From worker 94: ╎1 @Distributed/…messages.jl:121 (::Distributed.var"#process_messages##0#proc… From worker 94: ╎ 1 @Distributed/…essages.jl:133 process_tcp_streams(r_stream::Sockets.TCPSoc… From worker 94: ╎ 1 @Distributed/…essages.jl:176 message_handler_loop(r_stream::Sockets.TCPS… From worker 94: ╎ 1 @Distributed/…essages.jl:87 deserialize_msg(s::Distributed.ClusterSeria… From worker 94: ╎ 1 @Serialization/…tion.jl:848 deserialize From worker 94: ╎ 1 @Serialization/…tion.jl:891 handle_deserialize(s::Distributed.Cluster… From worker 94: ╎ ╎ 1 @Serialization/…ion.jl:1007 deserialize_tuple(s::Distributed.Cluster… From worker 94: ╎ ╎ 1 @Base/ntuple.jl:45 ntupleany(f::Serialization.var"#deserialize_tupl… From worker 94: ╎ ╎ 1 @Serialization/…on.jl:1007 (::Serialization.var"#deserialize_tuple… From worker 94: ╎ ╎ 1 @Serialization/…on.jl:848 deserialize(s::Distributed.ClusterSeria… From worker 94: ╎ ╎ 1 @Serialization/…on.jl:994 handle_deserialize(s::Distributed.Clus… From worker 94: ╎ ╎ ╎ 1 @Serialization/…n.jl:1081 deserialize(s::Distributed.ClusterSer… From worker 94: ╎ ╎ ╎ 1 @Serialization/…n.jl:848 deserialize From worker 94: ╎ ╎ ╎ 1 @Serialization/…n.jl:994 handle_deserialize(s::Distributed.Cl… From worker 94: ╎ ╎ ╎ 1 @Serialization/….jl:1245 deserialize(s::Distributed.ClusterS… From worker 94: ╎ ╎ ╎ 1 @Serialization/….jl:848 deserialize From worker 94: ╎ ╎ ╎ ╎ 1 @Serialization/…jl:899 handle_deserialize(s::Distributed.C… From worker 94: ╎ ╎ ╎ ╎ 1 @Serialization/…jl:1363 deserialize_array(s::Distributed.… From worker 94: [ Info: running task 49 on process 94 (Schedulers-primary-eA3Kzwpz); 5 workers total; 62 tasks left in task-pool. From worker 92: From worker 92: ====================================================================================== From worker 92: Information request received. A stacktrace will print followed by a 1.0 second profile From worker 92: ====================================================================================== From worker 92: From worker 92: cmd: /opt/julia/bin/julia 379 running 0 of 1 From worker 92: From worker 92: signal (10): User defined signal 1 From worker 92: epoll_pwait at /lib/x86_64-linux-gnu/libc.so.6 (unknown line) From worker 92: uv__io_poll at /workspace/srcdir/libuv/src/unix/linux.c:1404 From worker 92: uv_run at /workspace/srcdir/libuv/src/unix/core.c:430 From worker 92: ijl_task_get_next at /source/src/scheduler.c:457 [ Info: running task 50 on process 95 (Schedulers-primary-eA3Kzwpz); 5 workers total; 61 tasks left in task-pool. From worker 92: wait at ./task.jl:1213 From worker 92: wait_forever at ./task.jl:1150 From worker 92: jfptr_wait_forever_41430.1 at /opt/julia/lib/julia/sys.so (unknown line) From worker 92: _jl_invoke at /source/src/gf.c:3789 [inlined] From worker 92: ijl_apply_generic at /source/src/gf.c:3989 From worker 92: jl_apply at /source/src/julia.h:2345 [inlined] From worker 92: start_task at /source/src/task.c:1249 From worker 92: unknown function (ip: (nil)) at (unknown file) From worker 92: From worker 92: ============================================================== From worker 92: Profile collected. A report will print at the next yield point From worker 92: ============================================================== From worker 92: From worker 93: ┌ Warning: There were no samples collected in one or more groups. From worker 93: │ This may be due to idle threads, or you may need to run your From worker 93: │ program longer (perhaps by running it multiple times), From worker 93: │ or adjust the delay between samples with `Profile.init()`. From worker 93: └ @ Profile /opt/julia/share/julia/stdlib/v1.13/Profile/src/Profile.jl:1362 From worker 93: Overhead ╎ [+additional indent] Count File:Line Function From worker 93: ========================================================= From worker 93: Thread 1 (default) Task 0x0000794a94a014b0 Total snapshots: 1. Utilization: 100% From worker 93: ╎1 @Distributed/…messages.jl:121 (::Distributed.var"#process_messages##0#proc… From worker 93: ╎ 1 @Distributed/…essages.jl:133 process_tcp_streams(r_stream::Sockets.TCPSoc… From worker 93: ╎ 1 @Distributed/…essages.jl:176 message_handler_loop(r_stream::Sockets.TCPS… From worker 93: ╎ 1 @Distributed/…essages.jl:87 deserialize_msg(s::Distributed.ClusterSeria… From worker 93: ╎ 1 @Serialization/…tion.jl:848 deserialize From worker 93: ╎ 1 @Serialization/…tion.jl:891 handle_deserialize(s::Distributed.Cluster… From worker 93: ╎ ╎ 1 @Serialization/…ion.jl:1007 deserialize_tuple(s::Distributed.Cluster… From worker 93: ╎ ╎ 1 @Base/ntuple.jl:45 ntupleany(f::Serialization.var"#deserialize_tupl… From worker 93: ╎ ╎ 1 @Serialization/…on.jl:1007 (::Serialization.var"#deserialize_tuple… From worker 93: ╎ ╎ 1 @Serialization/…on.jl:848 deserialize(s::Distributed.ClusterSeria… From worker 93: ╎ ╎ 1 @Serialization/…on.jl:917 handle_deserialize(s::Distributed.Clus… From worker 93: ╎ ╎ ╎ 1 @Distributed/…all.jl:383 deserialize(s::Distributed.ClusterSeri… From worker 93: ╎ ╎ ╎ 1 @Serialization/…n.jl:1616 deserialize(s::Distributed.ClusterSe… From worker 93: ╎ ╎ ╎ 1 @Serialization/…n.jl:917 handle_deserialize(s::Distributed.Cl… From worker 93: ====================================================================================== Information request received. A stacktrace will print followed by a 1.0 second profile ====================================================================================== cmd: /opt/julia/bin/julia 15 running 0 of 1 signal (10): User defined signal 1 epoll_pwait at /lib/x86_64-linux-gnu/libc.so.6 (unknown line) uv__io_poll at /workspace/srcdir/libuv/src/unix/linux.c:1404 uv_run at /workspace/srcdir/libuv/src/unix/core.c:430 ijl_task_get_next at /source/src/scheduler.c:457 ┌ Warning: pid=92 (Schedulers-primary-eA3Kzwpz), checkpoint=nothing, task loop, caught exception during save_checkpoint └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1410 wait at ./task.jl:1213 ┌ Warning: TaskFailedException │ │ nested task error: On worker 92: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] #default_threadpool_checkpoint_call#25 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:284 │ [5] default_threadpool_checkpoint_call │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:250 │ [6] #handle_msg##8 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:303 │ [7] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:70 │ [8] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:79 │ [9] #schedule_call##2 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:88 │ │ nested task error: foo,x=0.9215242000458715 │ Stacktrace: │ [1] error │ @ ./error.jl:44 │ [2] test_save_checkpoint │ @ ./none:-1 [inlined] │ [3] test_save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:0 │ [4] save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1785 │ [5] #28 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:262 │ Stacktrace: │ [1] remotecall_wait(::Function, ::Distributed.Worker, ::Nothing, ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:507 │ [2] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:497 [inlined] │ [3] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:519 [inlined] │ [4] (::Schedulers.var"#34#35"{@Kwargs{}, Nothing, Schedulers.var"#140#141", Schedulers.var"#142#143", Int64, typeof(Schedulers.save_checkpoint), Int64, Tuple{typeof(test_save_checkpoint), typeof(fetch), String, Future, DataType}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:288 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] remotecall_wait_timeout(::Vector{Float64}, ::Int64, ::Float64, ::Nothing, ::Function, ::Function, ::Int64, ::typeof(Schedulers.save_checkpoint), ::Int64, ::typeof(test_save_checkpoint), ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:298 │ [5] remotecall_wait_timeout(::Vector{Float64}, ::Int64, ::Float64, ::Nothing, ::Function, ::Function, ::Int64, ::Function, ::Int64, ::typeof(test_save_checkpoint), ::Vararg{Any}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:287 │ [6] (::Schedulers.var"#134#135"{Vector{Float32}, @Kwargs{}, typeof(foo7b), Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", typeof(nworkers), Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, String}, Dict{String, Any}, SchedulerOptions{String}, Tuple{Int64, Int64}, Vector{Float64}, Vector{Float64}, Vector{Float64}, Vector{Any}, Dict{Int64, Future}, Int64})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1405 │ │ nested task error: On worker 92: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] #default_threadpool_checkpoint_call#25 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:284 │ [5] default_threadpool_checkpoint_call │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:250 │ [6] #handle_msg##8 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:303 │ [7] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:70 │ [8] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:79 │ [9] #schedule_call##2 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:88 │ │ nested task error: foo,x=0.9215242000458715 │ Stacktrace: │ [1] error │ @ ./error.jl:44 │ [2] test_save_checkpoint │ @ ./none:-1 [inlined] │ [3] test_save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:0 │ [4] save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1785 │ [5] #28 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:262 │ Stacktrace: │ [1] remotecall_wait(::Function, ::Distributed.Worker, ::Nothing, ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:507 │ [2] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:497 [inlined] │ [3] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:519 [inlined] │ [4] (::Schedulers.var"#34#35"{@Kwargs{}, Nothing, Schedulers.var"#140#141", Schedulers.var"#142#143", Int64, typeof(Schedulers.save_checkpoint), Int64, Tuple{typeof(test_save_checkpoint), typeof(fetch), String, Future, DataType}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:288 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 ┌ Warning: too many failures on process with id=92 (Schedulers-primary-eA3Kzwpz), removing from process list └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:393 wait_forever at ./task.jl:1150 jfptr_wait_forever_41430.1 at /opt/julia/lib/julia/sys.so (unknown line) _jl_invoke at /source/src/gf.c:3789 [inlined] ijl_apply_generic at /source/src/gf.c:3989 jl_apply at /source/src/julia.h:2345 [inlined] start_task at /source/src/task.c:1249 unknown function (ip: (nil)) at (unknown file) ============================================================== Profile collected. A report will print at the next yield point ============================================================== [ Info: running task 51 on process 93 (Schedulers-primary-eA3Kzwpz); 5 workers total; 61 tasks left in task-pool. [ Info: running task 52 on process 94 (Schedulers-primary-eA3Kzwpz); 5 workers total; 60 tasks left in task-pool. [ Info: running task 53 on process 95 (Schedulers-primary-eA3Kzwpz); 5 workers total; 59 tasks left in task-pool. From worker 92: ┌ Warning: There were no samples collected in one or more groups. From worker 92: │ This may be due to idle threads, or you may need to run your From worker 92: │ program longer (perhaps by running it multiple times), From worker 92: │ or adjust the delay between samples with `Profile.init()`. From worker 92: └ @ Profile /opt/julia/share/julia/stdlib/v1.13/Profile/src/Profile.jl:1362 From worker 92: Overhead ╎ [+additional indent] Count File:Line Function From worker 92: ========================================================= From worker 92: Thread 1 (default) Task 0x000079a22b2c9870 Total snapshots: 1. Utilization: 100% From worker 92: ╎1 @Schedulers/…chedulers.jl:262 (::Schedulers.var"#28#29"{Base.Pairs{Symbol,… From worker 92: ╎ 1 @Compiler/…/typeinfer.jl:1570 typeinf_ext_toplevel(mi::Core.MethodInstanc… From worker 92: ╎ 1 @Compiler/…/typeinfer.jl:1561 typeinf_ext_toplevel(interp::Compiler.Nati… From worker 92: ╎ 1 @Compiler/…typeinfer.jl:1378 typeinf_ext(interp::Compiler.NativeInterpr… From worker 92: ╎ 1 @Compiler/…pretation.jl:4469 typeinf(interp::Compiler.NativeInterprete… From worker 92: ╎ 1 @Compiler/…retation.jl:4311 typeinf_local(interp::Compiler.NativeInte… From worker 92: ╎ ╎ 1 @Compiler/…retation.jl:3762 abstract_eval_basic_statement From worker 92: ╎ ╎ 1 @Compiler/…retation.jl:3805 abstract_eval_basic_statement From worker 92: ╎ ╎ 1 @Compiler/…etation.jl:3430 abstract_eval_statement_expr(interp::Co… From worker 92: ╎ ╎ 1 @Compiler/…etation.jl:3074 abstract_eval_call(interp::Compiler.Na… From worker 92: ╎ ╎ 1 @Compiler/…tation.jl:3056 abstract_call(interp::Compiler.NativeI… From worker 92: ╎ ╎ ╎ 1 @Compiler/…tation.jl:2896 abstract_call(interp::Compiler.Native… From worker 92: ╎ ╎ ╎ 1 @Compiler/…tation.jl:2903 abstract_call(interp::Compiler.Nativ… From worker 92: ╎ ╎ ╎ 1 @Compiler/…ation.jl:2796 abstract_call_known(interp::Compiler… From worker 92: ╎ ╎ ╎ 1 @Compiler/…ation.jl:120 abstract_call_gf_by_type(interp::Com… From worker 92: ╎ ╎ ╎ 1 @Compiler/…ation.jl:342 kwcall(::@NamedTuple{max_methods::I… From worker 92: ╎ ╎ ╎ ╎ 1 @Compiler/…tion.jl:348 find_method_matches(interp::Compile… From worker 92: ╎ ╎ ╎ ╎ 1 @Compiler/…tion.jl:386 find_simple_method_matches(interp:… From worker 92: ╎ ╎ ╎ ╎ 1 @Compiler/…able.jl:102 findall From worker 92: ╎ ╎ ╎ ╎ 1 @Compiler/…ble.jl:111 findall(sig::Type, table::Compile… From worker 92: ╎ ╎ ╎ ╎ 1 @Base/iddict.jl:86 setindex! From worker 92: [ Info: running task 54 on process 93 (Schedulers-primary-eA3Kzwpz); 5 workers total; 58 tasks left in task-pool. [ Info: running task 55 on process 94 (Schedulers-primary-eA3Kzwpz); 5 workers total; 57 tasks left in task-pool. [ Info: running task 56 on process 95 (Schedulers-primary-eA3Kzwpz); 5 workers total; 56 tasks left in task-pool. ====================================================================================== Information request received. A stacktrace will print followed by a 1.0 second profile ====================================================================================== cmd: /opt/julia/bin/julia 1 running 0 of 1 signal (10): User defined signal 1 epoll_pwait at /lib/x86_64-linux-gnu/libc.so.6 (unknown line) uv__io_poll at /workspace/srcdir/libuv/src/unix/linux.c:1404 uv_run at /workspace/srcdir/libuv/src/unix/core.c:430 ijl_task_get_next at /source/src/scheduler.c:457 wait at ./task.jl:1213 wait_forever at ./task.jl:1150 [ Info: running task 57 on process 93 (Schedulers-primary-eA3Kzwpz); 4 workers total; 55 tasks left in task-pool. jfptr_wait_forever_41430.1 at /opt/julia/lib/julia/sys.so (unknown line) _jl_invoke at /source/src/gf.c:3789 [inlined] ijl_apply_generic at /source/src/gf.c:3989 jl_apply at /source/src/julia.h:2345 [inlined] start_task at /source/src/task.c:1249 unknown function (ip: (nil)) at (unknown file) ============================================================== Profile collected. A report will print at the next yield point ============================================================== [ Info: running task 58 on process 95 (Schedulers-primary-eA3Kzwpz); 4 workers total; 54 tasks left in task-pool. ┌ Warning: pid=94 (Schedulers-primary-eA3Kzwpz), checkpoint=/tmp/jl_7rsBvB/checkpoint-Iab0sT-368, task loop, caught exception during save_checkpoint └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1410 ┌ Warning: TaskFailedException │ │ nested task error: On worker 94: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] #default_threadpool_checkpoint_call#25 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:284 │ [5] default_threadpool_checkpoint_call │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:250 │ [6] #handle_msg##8 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:303 │ [7] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:70 │ [8] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:79 │ [9] #schedule_call##2 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:88 │ │ nested task error: foo,x=0.8882861164771655 │ Stacktrace: │ [1] error │ @ ./error.jl:44 │ [2] test_save_checkpoint │ @ ./none:-1 [inlined] │ [3] test_save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:0 │ [4] save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1785 │ [5] #28 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:262 │ Stacktrace: │ [1] remotecall_wait(::Function, ::Distributed.Worker, ::Nothing, ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:507 │ [2] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:497 [inlined] │ [3] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:519 [inlined] │ [4] (::Schedulers.var"#34#35"{@Kwargs{}, Nothing, Schedulers.var"#140#141", Schedulers.var"#142#143", Int64, typeof(Schedulers.save_checkpoint), Int64, Tuple{typeof(test_save_checkpoint), typeof(fetch), String, Future, DataType}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:288 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] remotecall_wait_timeout(::Vector{Float64}, ::Int64, ::Float64, ::Nothing, ::Function, ::Function, ::Int64, ::typeof(Schedulers.save_checkpoint), ::Int64, ::typeof(test_save_checkpoint), ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:298 │ [5] remotecall_wait_timeout(::Vector{Float64}, ::Int64, ::Float64, ::Nothing, ::Function, ::Function, ::Int64, ::Function, ::Int64, ::typeof(test_save_checkpoint), ::Vararg{Any}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:287 │ [6] (::Schedulers.var"#134#135"{Vector{Float32}, @Kwargs{}, typeof(foo7b), Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", typeof(nworkers), Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, String}, Dict{String, Any}, SchedulerOptions{String}, Tuple{Int64, Int64}, Vector{Float64}, Vector{Float64}, Vector{Float64}, Vector{Any}, Dict{Int64, Future}, Int64})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1405 │ │ nested task error: On worker 94: │ TaskFailedException │ Stacktrace: │ [1] #wait#583 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] #default_threadpool_checkpoint_call#25 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:284 │ [5] default_threadpool_checkpoint_call │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:250 │ [6] #handle_msg##8 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:303 │ [7] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:70 │ [8] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:79 │ [9] #schedule_call##2 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:88 │ │ nested task error: foo,x=0.8882861164771655 │ Stacktrace: │ [1] error │ @ ./error.jl:44 │ [2] test_save_checkpoint │ @ ./none:-1 [inlined] │ [3] test_save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:0 │ [4] save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1785 │ [5] #28 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:262 │ Stacktrace: │ [1] remotecall_wait(::Function, ::Distributed.Worker, ::Nothing, ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:507 │ [2] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:497 [inlined] │ [3] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:519 [inlined] │ [4] (::Schedulers.var"#34#35"{@Kwargs{}, Nothing, Schedulers.var"#140#141", Schedulers.var"#142#143", Int64, typeof(Schedulers.save_checkpoint), Int64, Tuple{typeof(test_save_checkpoint), typeof(fetch), String, Future, DataType}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:288 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 ┌ Warning: too many failures on process with id=94 (Schedulers-primary-eA3Kzwpz), removing from process list └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:393 [ Info: running task 59 on process 93 (Schedulers-primary-eA3Kzwpz); 4 workers total; 54 tasks left in task-pool. [ Info: running task 60 on process 95 (Schedulers-primary-eA3Kzwpz); 4 workers total; 53 tasks left in task-pool. ┌ Warning: There were no samples collected in one or more groups. │ This may be due to idle threads, or you may need to run your │ program longer (perhaps by running it multiple times), │ or adjust the delay between samples with `Profile.init()`. └ @ Profile /opt/julia/share/julia/stdlib/v1.13/Profile/src/Profile.jl:1362 Overhead ╎ [+additional indent] Count File:Line Function ========================================================= Thread 1 (default) Task 0x0000763afa3dba30 Total snapshots: 1. Utilization: 0% ╎1 @Base/task.jl:1150 wait_forever() ╎ 1 @Base/task.jl:1213 wait() [ Info: running task 61 on process 93 (Schedulers-primary-eA3Kzwpz); 4 workers total; 52 tasks left in task-pool. [1] signal 15: Terminated in expression starting at /PkgEval.jl/scripts/evaluate.jl:210 epoll_pwait at /lib/x86_64-linux-gnu/libc.so.6 (unknown line) uv__io_poll at /workspace/srcdir/libuv/src/unix/linux.c:1404 uv_run at /workspace/srcdir/libuv/src/unix/core.c:430 ijl_task_get_next at /source/src/scheduler.c:457 [399] signal 15: Terminated in expression starting at none:0 _ZN4llvm15ScalarEvolution19getMinTrailingZerosEPKNS_4SCEVE at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) _ZN4llvm15ScalarEvolution11getRangeRefEPKNS_4SCEVENS0_13RangeSignHintEj at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) [15] signal 15: Terminated in expression starting at /home/pkgeval/.julia/packages/Schedulers/JBB7J/test/runtests.jl:433 _ZN4llvm13AnalysisUsage15setPreservesCFGEv at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) unknown function (ip: 0x130136f) at (unknown file) unknown function (ip: 0x7ffffffff) at (unknown file) unknown function (ip: 0x73ce0dfa67c3) at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl _ZN4llvm21MachineModuleAnalysis3KeyE at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) unknown function (ip: (nil)) at (unknown file) Allocations: 54765693 (Pool: 54764639; Big: 1054); GC: 63 wait at ./task.jl:1213 _ZL21StrengthenNoWrapFlagsPN4llvm15ScalarEvolutionENS_9SCEVTypesENS_8ArrayRefIPKNS_4SCEVEEENS4_11NoWrapFlagsE at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) _ZN4llvm15ScalarEvolution10getAddExprERNS_15SmallVectorImplIPKNS_4SCEVEEENS2_11NoWrapFlagsEj at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) wait_forever at ./task.jl:1150 _ZN4llvm15ScalarEvolution10getAddExprEPKNS_4SCEVES3_NS1_11NoWrapFlagsEj at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) _ZN4llvm12SCEVExpander25getAddRecExprPHILiterallyEPKNS_14SCEVAddRecExprEPKNS_4LoopERPNS_4TypeERb at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) _ZN4llvm12SCEVExpander25expandAddRecExprLiterallyEPKNS_14SCEVAddRecExprE at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) _ZN4llvm12SCEVExpander6expandEPKNS_4SCEVE at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) _ZN4llvm12SCEVExpander13expandCodeForEPKNS_4SCEVEPNS_4TypeE at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) jfptr_wait_forever_41430.1 at /opt/julia/lib/julia/sys.so (unknown line) _jl_invoke at /source/src/gf.c:3789 [inlined] ijl_apply_generic at /source/src/gf.c:3989 jl_apply at /source/src/julia.h:2345 [inlined] start_task at /source/src/task.c:1249 unknown function (ip: (nil)) at (unknown file) Allocations: 10613177 (Pool: 10613044; Big: 133); GC: 9 _ZNK12_GLOBAL__N_111LSRInstance6ExpandERKNS_6LSRUseERKNS_8LSRFixupERKNS_7FormulaEN4llvm21ilist_iterator_w_bitsINSA_12ilist_detail12node_optionsINSA_11InstructionELb1ELb0EvLb1ENSA_10BasicBlockEEELb0ELb0EEERNSA_15SmallVectorImplINSA_14WeakTrackingVHEEE.constprop.0 at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) _ZN12_GLOBAL__N_111LSRInstanceC1EPN4llvm4LoopERNS1_7IVUsersERNS1_15ScalarEvolutionERNS1_13DominatorTreeERNS1_8LoopInfoERKNS1_19TargetTransformInfoERNS1_15AssumptionCacheERNS1_17TargetLibraryInfoEPNS1_16MemorySSAUpdaterE at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) _ZL18ReduceLoopStrengthPN4llvm4LoopERNS_7IVUsersERNS_15ScalarEvolutionERNS_13DominatorTreeERNS_8LoopInfoERKNS_19TargetTransformInfoERNS_15AssumptionCacheERNS_17TargetLibraryInfoEPNS_9MemorySSAE at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) _ZN12_GLOBAL__N_118LoopStrengthReduce9runOnLoopEPN4llvm4LoopERNS1_13LPPassManagerE at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) _ZN4llvm13LPPassManager13runOnFunctionERNS_8FunctionE at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) _ZN4llvm13FPPassManager13runOnFunctionERNS_8FunctionE at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) _ZN4llvm13FPPassManager11runOnModuleERNS_6ModuleE at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) _ZN4llvm6legacy15PassManagerImpl3runERNS_6ModuleE at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) _ZN4llvm3orc14SimpleCompilerclERNS_6ModuleE at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) operator() at /source/src/jitlayers.cpp:1620 addModule at /source/src/jitlayers.cpp:2101 jl_compile_codeinst_now at /source/src/jitlayers.cpp:683 jl_compile_codeinst_impl at /source/src/jitlayers.cpp:883 jl_compile_method_internal at /source/src/gf.c:3310 _jl_invoke at /source/src/gf.c:3781 [inlined] ijl_apply_generic at /source/src/gf.c:3989 jl_apply at /source/src/julia.h:2345 [inlined] jl_f_invokelatest at /source/src/builtins.c:877 exec_options at ./client.jl:266 _start at ./client.jl:557 jfptr__start_53432.1 at /opt/julia/lib/julia/sys.so (unknown line) _jl_invoke at /source/src/gf.c:3789 [inlined] ijl_apply_generic at /source/src/gf.c:3989 jl_apply at /source/src/julia.h:2345 [inlined] true_main at /source/src/jlapi.c:964 jl_repl_entrypoint at /source/src/jlapi.c:1132 main at /source/cli/loader_exe.c:58 unknown function (ip: 0x7140cff25249) at /lib/x86_64-linux-gnu/libc.so.6 __libc_start_main at /lib/x86_64-linux-gnu/libc.so.6 (unknown line) unknown function (ip: 0x4010b8) at /workspace/srcdir/glibc-2.17/csu/../sysdeps/x86_64/start.S unknown function (ip: (nil)) at (unknown file) Allocations: 1746795 (Pool: 1746793; Big: 2); GC: 2 PkgEval terminated after 2765.77s: test duration exceeded the time limit