Package evaluation of Schedulers on Julia 1.13.0-DEV.626 (157b8bc303*) started at 2025-05-22T23:48:02.100 ################################################################################ # Set-up # Installing PkgEval dependencies (TestEnv)... Set-up completed after 8.16s ################################################################################ # Installation # Installing Schedulers... Resolving package versions... Updating `~/.julia/environments/v1.13/Project.toml` [fcb5363a] + Schedulers v0.11.3 Updating `~/.julia/environments/v1.13/Manifest.toml` [682c06a0] + JSON v0.21.4 [69de0a69] + Parsers v2.8.3 [aea7be01] + PrecompileTools v1.3.2 [21216c6a] + Preferences v1.4.3 [fcb5363a] + Schedulers v0.11.3 [10745b16] + Statistics v1.11.1 [56f22d72] + Artifacts v1.11.0 [ade2ca70] + Dates v1.11.0 [8ba89e20] + Distributed v1.11.0 [8f399da3] + Libdl v1.11.0 [37e2e46d] + LinearAlgebra v1.12.0 [56ddb016] + Logging v1.11.0 [a63ad114] + Mmap v1.11.0 [de0858da] + Printf v1.11.0 [9a3f8284] + Random v1.11.0 [ea8e919c] + SHA v0.7.0 [9e88b42a] + Serialization v1.11.0 [6462fe0b] + Sockets v1.11.0 [fa267f1f] + TOML v1.0.3 [cf7118a7] + UUIDs v1.11.0 [4ec0a83e] + Unicode v1.11.0 [e66e0078] + CompilerSupportLibraries_jll v1.3.0+1 [4536629a] + OpenBLAS_jll v0.3.29+0 [8e850b90] + libblastrampoline_jll v5.12.0+0 Installation completed after 1.13s ################################################################################ # Precompilation # Precompiling PkgEval dependencies... Precompiling package dependencies... Precompilation completed after 15.65s ################################################################################ # Testing # Testing Schedulers Status `/tmp/jl_xvxFtR/Project.toml` [fcb5363a] Schedulers v0.11.3 [8ba89e20] Distributed v1.11.0 [56ddb016] Logging v1.11.0 [9a3f8284] Random v1.11.0 [9e88b42a] Serialization v1.11.0 [8dfed614] Test v1.11.0 Status `/tmp/jl_xvxFtR/Manifest.toml` [682c06a0] JSON v0.21.4 [69de0a69] Parsers v2.8.3 [aea7be01] PrecompileTools v1.3.2 [21216c6a] Preferences v1.4.3 [fcb5363a] Schedulers v0.11.3 [10745b16] Statistics v1.11.1 [56f22d72] Artifacts v1.11.0 [2a0f44e3] Base64 v1.11.0 [ade2ca70] Dates v1.11.0 [8ba89e20] Distributed v1.11.0 [b77e0a4c] InteractiveUtils v1.11.0 [ac6e5ff7] JuliaSyntaxHighlighting v1.12.0 [8f399da3] Libdl v1.11.0 [37e2e46d] LinearAlgebra v1.12.0 [56ddb016] Logging v1.11.0 [d6f4376e] Markdown v1.11.0 [a63ad114] Mmap v1.11.0 [de0858da] Printf v1.11.0 [9a3f8284] Random v1.11.0 [ea8e919c] SHA v0.7.0 [9e88b42a] Serialization v1.11.0 [6462fe0b] Sockets v1.11.0 [f489334b] StyledStrings v1.11.0 [fa267f1f] TOML v1.0.3 [8dfed614] Test v1.11.0 [cf7118a7] UUIDs v1.11.0 [4ec0a83e] Unicode v1.11.0 [e66e0078] CompilerSupportLibraries_jll v1.3.0+1 [4536629a] OpenBLAS_jll v0.3.29+0 [8e850b90] libblastrampoline_jll v5.12.0+0 Testing Running tests... [ Info: running task 1 on process 2 (Schedulers-primary-qO2tilUH); 5 workers total; 9 tasks left in task-pool. [ Info: running task 2 on process 3 (Schedulers-primary-qO2tilUH); 5 workers total; 5 tasks left in task-pool. [ Info: running task 3 on process 4 (Schedulers-primary-qO2tilUH); 5 workers total; 5 tasks left in task-pool. [ Info: running task 4 on process 5 (Schedulers-primary-qO2tilUH); 5 workers total; 5 tasks left in task-pool. [ Info: running task 5 on process 6 (Schedulers-primary-qO2tilUH); 5 workers total; 5 tasks left in task-pool. [ Info: running task 6 on process 2 (Schedulers-primary-qO2tilUH); 5 workers total; 4 tasks left in task-pool. [ Info: running task 7 on process 3 (Schedulers-primary-qO2tilUH); 5 workers total; 3 tasks left in task-pool. [ Info: running task 8 on process 4 (Schedulers-primary-qO2tilUH); 5 workers total; 0 tasks left in task-pool. [ Info: running task 9 on process 5 (Schedulers-primary-qO2tilUH); 5 workers total; 0 tasks left in task-pool. [ Info: running task 10 on process 6 (Schedulers-primary-qO2tilUH); 5 workers total; 0 tasks left in task-pool. Test Summary: | Pass Total Time pmap, stable cluster test | 16 16 2m34.0s [ Info: running task 1 on process 7 (Schedulers-primary-qO2tilUH); 10 workers total; 99 tasks left in task-pool. [ Info: running task 2 on process 8 (Schedulers-primary-qO2tilUH); 10 workers total; 98 tasks left in task-pool. [ Info: running task 3 on process 8 (Schedulers-primary-qO2tilUH); 10 workers total; 97 tasks left in task-pool. [ Info: running task 4 on process 7 (Schedulers-primary-qO2tilUH); 10 workers total; 96 tasks left in task-pool. ┌ Warning: addprocs/rmprocs taking longer than expected, cancelling. └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:776 ┌ Warning: problem adding or removing processes └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:737 ┌ Warning: TaskFailedException │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1213 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#193 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#582 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch(t::Task) │ @ Base ./task.jl:525 │ [4] loop(eloop::Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, journal::Dict{String, Any}, journal_task_callback::Function, tsk_map::Task, tsk_reduce::Task) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:735 │ [5] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1046 │ [6] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1041 │ [7] top-level scope │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:46 │ [8] macro expansion │ @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1833 [inlined] │ [9] macro expansion │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:55 [inlined] │ [10] include(mapexpr::Function, mod::Module, _path::String) │ @ Base ./Base.jl:310 │ [11] top-level scope │ @ none:6 │ [12] eval(m::Module, e::Any) │ @ Core ./boot.jl:489 │ [13] exec_options(opts::Base.JLOptions) │ @ Base ./client.jl:297 │ [14] _start() │ @ Base ./client.jl:564 │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1213 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#193 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 [ Info: running task 5 on process 8 (Schedulers-primary-qO2tilUH); 10 workers total; 95 tasks left in task-pool. [ Info: running task 6 on process 7 (Schedulers-primary-qO2tilUH); 10 workers total; 94 tasks left in task-pool. ┌ Warning: problem adding or removing processes └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:737 ┌ Warning: TaskFailedException │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1213 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#193 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#582 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch(t::Task) │ @ Base ./task.jl:525 │ [4] loop(eloop::Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, journal::Dict{String, Any}, journal_task_callback::Function, tsk_map::Task, tsk_reduce::Task) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:735 │ [5] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1046 │ [6] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1041 │ [7] top-level scope │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:46 │ [8] macro expansion │ @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1833 [inlined] │ [9] macro expansion │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:55 [inlined] │ [10] include(mapexpr::Function, mod::Module, _path::String) │ @ Base ./Base.jl:310 │ [11] top-level scope │ @ none:6 │ [12] eval(m::Module, e::Any) │ @ Core ./boot.jl:489 │ [13] exec_options(opts::Base.JLOptions) │ @ Base ./client.jl:297 │ [14] _start() │ @ Base ./client.jl:564 │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1213 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#193 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 ┌ Warning: problem adding or removing processes └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:737 ┌ Warning: TaskFailedException │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1213 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#193 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#582 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch(t::Task) │ @ Base ./task.jl:525 │ [4] loop(eloop::Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, journal::Dict{String, Any}, journal_task_callback::Function, tsk_map::Task, tsk_reduce::Task) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:735 │ [5] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1046 │ [6] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1041 │ [7] top-level scope │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:46 │ [8] macro expansion │ @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1833 [inlined] │ [9] macro expansion │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:55 [inlined] │ [10] include(mapexpr::Function, mod::Module, _path::String) │ @ Base ./Base.jl:310 │ [11] top-level scope │ @ none:6 │ [12] eval(m::Module, e::Any) │ @ Core ./boot.jl:489 │ [13] exec_options(opts::Base.JLOptions) │ @ Base ./client.jl:297 │ [14] _start() │ @ Base ./client.jl:564 │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1213 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#193 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 [ Info: running task 7 on process 8 (Schedulers-primary-qO2tilUH); 10 workers total; 93 tasks left in task-pool. [ Info: running task 8 on process 7 (Schedulers-primary-qO2tilUH); 10 workers total; 92 tasks left in task-pool. ┌ Warning: problem adding or removing processes └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:737 ┌ Warning: TaskFailedException │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1213 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#193 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#582 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch(t::Task) │ @ Base ./task.jl:525 │ [4] loop(eloop::Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, journal::Dict{String, Any}, journal_task_callback::Function, tsk_map::Task, tsk_reduce::Task) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:735 │ [5] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1046 │ [6] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1041 │ [7] top-level scope │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:46 │ [8] macro expansion │ @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1833 [inlined] │ [9] macro expansion │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:55 [inlined] │ [10] include(mapexpr::Function, mod::Module, _path::String) │ @ Base ./Base.jl:310 │ [11] top-level scope │ @ none:6 │ [12] eval(m::Module, e::Any) │ @ Core ./boot.jl:489 │ [13] exec_options(opts::Base.JLOptions) │ @ Base ./client.jl:297 │ [14] _start() │ @ Base ./client.jl:564 │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1213 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#193 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 ┌ Warning: problem adding or removing processes └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:737 ┌ Warning: TaskFailedException │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1213 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#193 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#582 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch(t::Task) │ @ Base ./task.jl:525 │ [4] loop(eloop::Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, journal::Dict{String, Any}, journal_task_callback::Function, tsk_map::Task, tsk_reduce::Task) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:735 │ [5] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1046 │ [6] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1041 │ [7] top-level scope │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:46 │ [8] macro expansion │ @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1833 [inlined] │ [9] macro expansion │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:55 [inlined] │ [10] include(mapexpr::Function, mod::Module, _path::String) │ @ Base ./Base.jl:310 │ [11] top-level scope │ @ none:6 │ [12] eval(m::Module, e::Any) │ @ Core ./boot.jl:489 │ [13] exec_options(opts::Base.JLOptions) │ @ Base ./client.jl:297 │ [14] _start() │ @ Base ./client.jl:564 │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1213 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#193 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 ┌ Warning: problem adding or removing processes └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:737 ┌ Warning: TaskFailedException │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1213 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#193 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#582 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch(t::Task) │ @ Base ./task.jl:525 │ [4] loop(eloop::Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, journal::Dict{String, Any}, journal_task_callback::Function, tsk_map::Task, tsk_reduce::Task) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:735 │ [5] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1046 │ [6] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1041 │ [7] top-level scope │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:46 │ [8] macro expansion │ @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1833 [inlined] │ [9] macro expansion │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:55 [inlined] │ [10] include(mapexpr::Function, mod::Module, _path::String) │ @ Base ./Base.jl:310 │ [11] top-level scope │ @ none:6 │ [12] eval(m::Module, e::Any) │ @ Core ./boot.jl:489 │ [13] exec_options(opts::Base.JLOptions) │ @ Base ./client.jl:297 │ [14] _start() │ @ Base ./client.jl:564 │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1213 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#193 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 [ Info: running task 9 on process 8 (Schedulers-primary-qO2tilUH); 10 workers total; 91 tasks left in task-pool. [ Info: running task 10 on process 7 (Schedulers-primary-qO2tilUH); 10 workers total; 90 tasks left in task-pool. ┌ Warning: problem adding or removing processes └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:737 ┌ Warning: TaskFailedException │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1213 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#193 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#582 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch(t::Task) │ @ Base ./task.jl:525 │ [4] loop(eloop::Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, journal::Dict{String, Any}, journal_task_callback::Function, tsk_map::Task, tsk_reduce::Task) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:735 │ [5] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1046 │ [6] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1041 │ [7] top-level scope │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:46 │ [8] macro expansion │ @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1833 [inlined] │ [9] macro expansion │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:55 [inlined] │ [10] include(mapexpr::Function, mod::Module, _path::String) │ @ Base ./Base.jl:310 │ [11] top-level scope │ @ none:6 │ [12] eval(m::Module, e::Any) │ @ Core ./boot.jl:489 │ [13] exec_options(opts::Base.JLOptions) │ @ Base ./client.jl:297 │ [14] _start() │ @ Base ./client.jl:564 │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1213 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#193 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 From worker 15: ErrorException("Process(1) - Invalid connection credentials sent by remote.")CapturedException(ErrorException("Process(1) - Invalid connection credentials sent by remote."), Any[(error(s::String) at error.jl:44, 1), (process_hdr(s::Sockets.TCPSocket, validate_cookie::Bool) at process_messages.jl:266, 1), (message_handler_loop(r_stream::Sockets.TCPSocket, w_stream::Sockets.TCPSocket, incoming::Bool) at process_messages.jl:158, 1), (process_tcp_streams(r_stream::Sockets.TCPSocket, w_stream::Sockets.TCPSocket, incoming::Bool) at process_messages.jl:133, 1), ((::Distributed.var"#process_messages##0#process_messages##1"{Sockets.TCPSocket, Sockets.TCPSocket, Bool})() at process_messages.jl:121, 1)]) From worker 15: Process(1) - Unknown remote, closing connection. Worker 15 terminated. UNHANDLED TASK ERROR: Version read failed. Connection closed by peer. Stacktrace: [1] error(s::String) @ Base ./error.jl:44 [2] process_hdr(s::Sockets.TCPSocket, validate_cookie::Bool) @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:276 [3] message_handler_loop(r_stream::Sockets.TCPSocket, w_stream::Sockets.TCPSocket, incoming::Bool) @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:158 [4] process_tcp_streams(r_stream::Sockets.TCPSocket, w_stream::Sockets.TCPSocket, incoming::Bool) @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:133 [5] (::Distributed.var"#process_messages##0#process_messages##1"{Sockets.TCPSocket, Sockets.TCPSocket, Bool})() @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:121 [ Info: running task 11 on process 8 (Schedulers-primary-qO2tilUH); 9 workers total; 89 tasks left in task-pool. [ Info: running task 12 on process 7 (Schedulers-primary-qO2tilUH); 9 workers total; 88 tasks left in task-pool. ┌ Warning: problem adding or removing processes └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:737 ┌ Warning: TaskFailedException │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1213 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#193 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#582 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch(t::Task) │ @ Base ./task.jl:525 │ [4] loop(eloop::Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, journal::Dict{String, Any}, journal_task_callback::Function, tsk_map::Task, tsk_reduce::Task) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:735 │ [5] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1046 │ [6] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1041 │ [7] top-level scope │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:46 │ [8] macro expansion │ @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1833 [inlined] │ [9] macro expansion │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:55 [inlined] │ [10] include(mapexpr::Function, mod::Module, _path::String) │ @ Base ./Base.jl:310 │ [11] top-level scope │ @ none:6 │ [12] eval(m::Module, e::Any) │ @ Core ./boot.jl:489 │ [13] exec_options(opts::Base.JLOptions) │ @ Base ./client.jl:297 │ [14] _start() │ @ Base ./client.jl:564 │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1213 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#193 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 From worker 15: Master process (id 1) could not connect within 120.0 seconds. From worker 15: exiting. [ Info: running task 13 on process 9 (Schedulers-primary-qO2tilUH); 10 workers total; 87 tasks left in task-pool. [ Info: running task 14 on process 10 (Schedulers-primary-qO2tilUH); 10 workers total; 81 tasks left in task-pool. [ Info: running task 15 on process 11 (Schedulers-primary-qO2tilUH); 10 workers total; 81 tasks left in task-pool. [ Info: running task 16 on process 12 (Schedulers-primary-qO2tilUH); 10 workers total; 81 tasks left in task-pool. [ Info: running task 17 on process 13 (Schedulers-primary-qO2tilUH); 10 workers total; 81 tasks left in task-pool. [ Info: running task 18 on process 14 (Schedulers-primary-qO2tilUH); 10 workers total; 81 tasks left in task-pool. [ Info: running task 19 on process 16 (Schedulers-primary-qO2tilUH); 10 workers total; 81 tasks left in task-pool. [ Info: running task 20 on process 8 (Schedulers-primary-qO2tilUH); 10 workers total; 80 tasks left in task-pool. [ Info: running task 21 on process 7 (Schedulers-primary-qO2tilUH); 10 workers total; 79 tasks left in task-pool. [ Info: running task 22 on process 8 (Schedulers-primary-qO2tilUH); 10 workers total; 78 tasks left in task-pool. [ Info: running task 23 on process 9 (Schedulers-primary-qO2tilUH); 10 workers total; 77 tasks left in task-pool. [ Info: running task 24 on process 10 (Schedulers-primary-qO2tilUH); 10 workers total; 71 tasks left in task-pool. [ Info: running task 25 on process 11 (Schedulers-primary-qO2tilUH); 10 workers total; 71 tasks left in task-pool. [ Info: running task 26 on process 12 (Schedulers-primary-qO2tilUH); 10 workers total; 71 tasks left in task-pool. [ Info: running task 27 on process 13 (Schedulers-primary-qO2tilUH); 10 workers total; 71 tasks left in task-pool. [ Info: running task 28 on process 14 (Schedulers-primary-qO2tilUH); 10 workers total; 71 tasks left in task-pool. [ Info: running task 29 on process 16 (Schedulers-primary-qO2tilUH); 10 workers total; 71 tasks left in task-pool. [ Info: running task 30 on process 7 (Schedulers-primary-qO2tilUH); 10 workers total; 70 tasks left in task-pool. [ Info: running task 31 on process 17 (Schedulers-primary-qO2tilUH); 10 workers total; 69 tasks left in task-pool. [ Info: running task 32 on process 8 (Schedulers-primary-qO2tilUH); 10 workers total; 68 tasks left in task-pool. [ Info: running task 33 on process 9 (Schedulers-primary-qO2tilUH); 10 workers total; 67 tasks left in task-pool. [ Info: running task 34 on process 10 (Schedulers-primary-qO2tilUH); 10 workers total; 61 tasks left in task-pool. [ Info: running task 35 on process 11 (Schedulers-primary-qO2tilUH); 10 workers total; 61 tasks left in task-pool. [ Info: running task 36 on process 12 (Schedulers-primary-qO2tilUH); 10 workers total; 61 tasks left in task-pool. [ Info: running task 37 on process 13 (Schedulers-primary-qO2tilUH); 10 workers total; 61 tasks left in task-pool. [ Info: running task 38 on process 14 (Schedulers-primary-qO2tilUH); 10 workers total; 61 tasks left in task-pool. [ Info: running task 39 on process 16 (Schedulers-primary-qO2tilUH); 10 workers total; 61 tasks left in task-pool. [ Info: running task 40 on process 7 (Schedulers-primary-qO2tilUH); 10 workers total; 60 tasks left in task-pool. [ Info: running task 41 on process 17 (Schedulers-primary-qO2tilUH); 10 workers total; 59 tasks left in task-pool. [ Info: running task 42 on process 8 (Schedulers-primary-qO2tilUH); 10 workers total; 58 tasks left in task-pool. [ Info: running task 43 on process 9 (Schedulers-primary-qO2tilUH); 10 workers total; 57 tasks left in task-pool. [ Info: running task 44 on process 10 (Schedulers-primary-qO2tilUH); 10 workers total; 51 tasks left in task-pool. [ Info: running task 45 on process 11 (Schedulers-primary-qO2tilUH); 10 workers total; 51 tasks left in task-pool. [ Info: running task 46 on process 12 (Schedulers-primary-qO2tilUH); 10 workers total; 51 tasks left in task-pool. [ Info: running task 47 on process 13 (Schedulers-primary-qO2tilUH); 10 workers total; 51 tasks left in task-pool. [ Info: running task 48 on process 14 (Schedulers-primary-qO2tilUH); 10 workers total; 51 tasks left in task-pool. [ Info: running task 49 on process 16 (Schedulers-primary-qO2tilUH); 10 workers total; 51 tasks left in task-pool. [ Info: running task 50 on process 7 (Schedulers-primary-qO2tilUH); 10 workers total; 50 tasks left in task-pool. [ Info: running task 51 on process 17 (Schedulers-primary-qO2tilUH); 10 workers total; 49 tasks left in task-pool. [ Info: running task 52 on process 8 (Schedulers-primary-qO2tilUH); 10 workers total; 48 tasks left in task-pool. [ Info: running task 53 on process 9 (Schedulers-primary-qO2tilUH); 10 workers total; 47 tasks left in task-pool. [ Info: running task 54 on process 10 (Schedulers-primary-qO2tilUH); 10 workers total; 41 tasks left in task-pool. [ Info: running task 55 on process 11 (Schedulers-primary-qO2tilUH); 10 workers total; 41 tasks left in task-pool. [ Info: running task 56 on process 12 (Schedulers-primary-qO2tilUH); 10 workers total; 41 tasks left in task-pool. [ Info: running task 57 on process 13 (Schedulers-primary-qO2tilUH); 10 workers total; 41 tasks left in task-pool. [ Info: running task 58 on process 14 (Schedulers-primary-qO2tilUH); 10 workers total; 41 tasks left in task-pool. [ Info: running task 59 on process 16 (Schedulers-primary-qO2tilUH); 10 workers total; 41 tasks left in task-pool. [ Info: running task 60 on process 7 (Schedulers-primary-qO2tilUH); 10 workers total; 40 tasks left in task-pool. [ Info: running task 61 on process 17 (Schedulers-primary-qO2tilUH); 10 workers total; 39 tasks left in task-pool. [ Info: running task 62 on process 8 (Schedulers-primary-qO2tilUH); 10 workers total; 38 tasks left in task-pool. [ Info: running task 63 on process 9 (Schedulers-primary-qO2tilUH); 10 workers total; 37 tasks left in task-pool. [ Info: running task 64 on process 10 (Schedulers-primary-qO2tilUH); 10 workers total; 34 tasks left in task-pool. [ Info: running task 65 on process 11 (Schedulers-primary-qO2tilUH); 10 workers total; 34 tasks left in task-pool. [ Info: running task 66 on process 12 (Schedulers-primary-qO2tilUH); 10 workers total; 34 tasks left in task-pool. [ Info: running task 67 on process 13 (Schedulers-primary-qO2tilUH); 10 workers total; 33 tasks left in task-pool. [ Info: running task 68 on process 14 (Schedulers-primary-qO2tilUH); 10 workers total; 31 tasks left in task-pool. [ Info: running task 69 on process 16 (Schedulers-primary-qO2tilUH); 10 workers total; 31 tasks left in task-pool. [ Info: running task 70 on process 7 (Schedulers-primary-qO2tilUH); 10 workers total; 30 tasks left in task-pool. [ Info: running task 71 on process 17 (Schedulers-primary-qO2tilUH); 10 workers total; 29 tasks left in task-pool. [ Info: running task 72 on process 8 (Schedulers-primary-qO2tilUH); 10 workers total; 28 tasks left in task-pool. [ Info: running task 73 on process 9 (Schedulers-primary-qO2tilUH); 10 workers total; 27 tasks left in task-pool. [ Info: running task 74 on process 10 (Schedulers-primary-qO2tilUH); 10 workers total; 24 tasks left in task-pool. [ Info: running task 75 on process 11 (Schedulers-primary-qO2tilUH); 10 workers total; 24 tasks left in task-pool. [ Info: running task 76 on process 12 (Schedulers-primary-qO2tilUH); 10 workers total; 24 tasks left in task-pool. [ Info: running task 77 on process 13 (Schedulers-primary-qO2tilUH); 10 workers total; 23 tasks left in task-pool. [ Info: running task 78 on process 14 (Schedulers-primary-qO2tilUH); 10 workers total; 21 tasks left in task-pool. [ Info: running task 79 on process 16 (Schedulers-primary-qO2tilUH); 10 workers total; 21 tasks left in task-pool. [ Info: running task 80 on process 7 (Schedulers-primary-qO2tilUH); 10 workers total; 20 tasks left in task-pool. [ Info: running task 81 on process 17 (Schedulers-primary-qO2tilUH); 10 workers total; 19 tasks left in task-pool. [ Info: running task 82 on process 8 (Schedulers-primary-qO2tilUH); 10 workers total; 18 tasks left in task-pool. [ Info: running task 83 on process 9 (Schedulers-primary-qO2tilUH); 10 workers total; 17 tasks left in task-pool. [ Info: running task 84 on process 10 (Schedulers-primary-qO2tilUH); 10 workers total; 14 tasks left in task-pool. [ Info: running task 85 on process 11 (Schedulers-primary-qO2tilUH); 10 workers total; 14 tasks left in task-pool. [ Info: running task 86 on process 12 (Schedulers-primary-qO2tilUH); 10 workers total; 14 tasks left in task-pool. [ Info: running task 87 on process 13 (Schedulers-primary-qO2tilUH); 10 workers total; 13 tasks left in task-pool. [ Info: running task 88 on process 14 (Schedulers-primary-qO2tilUH); 10 workers total; 11 tasks left in task-pool. [ Info: running task 89 on process 16 (Schedulers-primary-qO2tilUH); 10 workers total; 11 tasks left in task-pool. [ Info: running task 90 on process 7 (Schedulers-primary-qO2tilUH); 10 workers total; 10 tasks left in task-pool. [ Info: running task 91 on process 17 (Schedulers-primary-qO2tilUH); 10 workers total; 9 tasks left in task-pool. [ Info: running task 92 on process 8 (Schedulers-primary-qO2tilUH); 10 workers total; 8 tasks left in task-pool. [ Info: running task 93 on process 9 (Schedulers-primary-qO2tilUH); 10 workers total; 7 tasks left in task-pool. [ Info: running task 94 on process 10 (Schedulers-primary-qO2tilUH); 10 workers total; 4 tasks left in task-pool. [ Info: running task 95 on process 11 (Schedulers-primary-qO2tilUH); 10 workers total; 4 tasks left in task-pool. [ Info: running task 96 on process 12 (Schedulers-primary-qO2tilUH); 10 workers total; 4 tasks left in task-pool. [ Info: running task 97 on process 13 (Schedulers-primary-qO2tilUH); 10 workers total; 3 tasks left in task-pool. [ Info: running task 98 on process 14 (Schedulers-primary-qO2tilUH); 10 workers total; 1 tasks left in task-pool. [ Info: running task 99 on process 16 (Schedulers-primary-qO2tilUH); 10 workers total; 1 tasks left in task-pool. [ Info: running task 100 on process 7 (Schedulers-primary-qO2tilUH); 10 workers total; 0 tasks left in task-pool. Test Summary: | Pass Total Time pmap, growing cluster test | 111 111 5m04.6s [ Info: running task 1 on process 18 (Schedulers-primary-qO2tilUH); 8 workers total; 99 tasks left in task-pool. [ Info: running task 2 on process 21 (Schedulers-primary-qO2tilUH); 8 workers total; 92 tasks left in task-pool. [ Info: running task 3 on process 22 (Schedulers-primary-qO2tilUH); 8 workers total; 92 tasks left in task-pool. [ Info: running task 4 on process 23 (Schedulers-primary-qO2tilUH); 8 workers total; 92 tasks left in task-pool. [ Info: running task 5 on process 24 (Schedulers-primary-qO2tilUH); 8 workers total; 92 tasks left in task-pool. [ Info: running task 6 on process 25 (Schedulers-primary-qO2tilUH); 8 workers total; 92 tasks left in task-pool. [ Info: running task 7 on process 26 (Schedulers-primary-qO2tilUH); 8 workers total; 92 tasks left in task-pool. [ Info: running task 8 on process 27 (Schedulers-primary-qO2tilUH); 8 workers total; 92 tasks left in task-pool. [ Info: running task 9 on process 18 (Schedulers-primary-qO2tilUH); 10 workers total; 91 tasks left in task-pool. [ Info: running task 10 on process 21 (Schedulers-primary-qO2tilUH); 10 workers total; 90 tasks left in task-pool. [ Info: running task 11 on process 22 (Schedulers-primary-qO2tilUH); 10 workers total; 84 tasks left in task-pool. [ Info: running task 12 on process 23 (Schedulers-primary-qO2tilUH); 10 workers total; 84 tasks left in task-pool. [ Info: running task 13 on process 24 (Schedulers-primary-qO2tilUH); 10 workers total; 84 tasks left in task-pool. [ Info: running task 14 on process 25 (Schedulers-primary-qO2tilUH); 10 workers total; 84 tasks left in task-pool. [ Info: running task 15 on process 26 (Schedulers-primary-qO2tilUH); 10 workers total; 84 tasks left in task-pool. [ Info: running task 16 on process 27 (Schedulers-primary-qO2tilUH); 10 workers total; 84 tasks left in task-pool. [ Info: running task 17 on process 18 (Schedulers-primary-qO2tilUH); 10 workers total; 83 tasks left in task-pool. [ Info: running task 18 on process 21 (Schedulers-primary-qO2tilUH); 10 workers total; 82 tasks left in task-pool. [ Info: running task 19 on process 22 (Schedulers-primary-qO2tilUH); 10 workers total; 81 tasks left in task-pool. [ Info: running task 20 on process 23 (Schedulers-primary-qO2tilUH); 10 workers total; 76 tasks left in task-pool. [ Info: running task 21 on process 24 (Schedulers-primary-qO2tilUH); 10 workers total; 76 tasks left in task-pool. [ Info: running task 22 on process 25 (Schedulers-primary-qO2tilUH); 10 workers total; 76 tasks left in task-pool. [ Info: running task 23 on process 26 (Schedulers-primary-qO2tilUH); 10 workers total; 76 tasks left in task-pool. [ Info: running task 24 on process 27 (Schedulers-primary-qO2tilUH); 10 workers total; 76 tasks left in task-pool. [ Info: running task 25 on process 28 (Schedulers-primary-qO2tilUH); 10 workers total; 75 tasks left in task-pool. [ Info: running task 26 on process 29 (Schedulers-primary-qO2tilUH); 10 workers total; 74 tasks left in task-pool. [ Info: running task 27 on process 18 (Schedulers-primary-qO2tilUH); 10 workers total; 73 tasks left in task-pool. [ Info: running task 28 on process 21 (Schedulers-primary-qO2tilUH); 10 workers total; 72 tasks left in task-pool. [ Info: running task 29 on process 22 (Schedulers-primary-qO2tilUH); 10 workers total; 71 tasks left in task-pool. [ Info: running task 30 on process 23 (Schedulers-primary-qO2tilUH); 10 workers total; 69 tasks left in task-pool. [ Info: running task 31 on process 24 (Schedulers-primary-qO2tilUH); 10 workers total; 69 tasks left in task-pool. [ Info: running task 32 on process 25 (Schedulers-primary-qO2tilUH); 10 workers total; 68 tasks left in task-pool. [ Info: running task 33 on process 26 (Schedulers-primary-qO2tilUH); 10 workers total; 66 tasks left in task-pool. [ Info: running task 34 on process 27 (Schedulers-primary-qO2tilUH); 10 workers total; 66 tasks left in task-pool. [ Info: running task 35 on process 28 (Schedulers-primary-qO2tilUH); 10 workers total; 65 tasks left in task-pool. [ Info: running task 36 on process 29 (Schedulers-primary-qO2tilUH); 10 workers total; 64 tasks left in task-pool. [ Info: running task 37 on process 18 (Schedulers-primary-qO2tilUH); 10 workers total; 63 tasks left in task-pool. [ Info: running task 38 on process 21 (Schedulers-primary-qO2tilUH); 10 workers total; 62 tasks left in task-pool. [ Info: running task 39 on process 22 (Schedulers-primary-qO2tilUH); 10 workers total; 61 tasks left in task-pool. [ Info: running task 40 on process 23 (Schedulers-primary-qO2tilUH); 10 workers total; 60 tasks left in task-pool. [ Info: running task 41 on process 24 (Schedulers-primary-qO2tilUH); 10 workers total; 59 tasks left in task-pool. [ Info: running task 42 on process 25 (Schedulers-primary-qO2tilUH); 10 workers total; 58 tasks left in task-pool. [ Info: running task 43 on process 26 (Schedulers-primary-qO2tilUH); 10 workers total; 56 tasks left in task-pool. [ Info: running task 44 on process 27 (Schedulers-primary-qO2tilUH); 10 workers total; 56 tasks left in task-pool. [ Info: running task 45 on process 28 (Schedulers-primary-qO2tilUH); 10 workers total; 55 tasks left in task-pool. [ Info: running task 46 on process 29 (Schedulers-primary-qO2tilUH); 10 workers total; 54 tasks left in task-pool. [ Info: running task 47 on process 18 (Schedulers-primary-qO2tilUH); 10 workers total; 53 tasks left in task-pool. [ Info: running task 48 on process 21 (Schedulers-primary-qO2tilUH); 10 workers total; 52 tasks left in task-pool. [ Info: running task 49 on process 22 (Schedulers-primary-qO2tilUH); 10 workers total; 51 tasks left in task-pool. [ Info: running task 50 on process 23 (Schedulers-primary-qO2tilUH); 10 workers total; 50 tasks left in task-pool. [ Info: running task 51 on process 24 (Schedulers-primary-qO2tilUH); 10 workers total; 49 tasks left in task-pool. [ Info: running task 52 on process 25 (Schedulers-primary-qO2tilUH); 10 workers total; 48 tasks left in task-pool. [ Info: running task 53 on process 26 (Schedulers-primary-qO2tilUH); 10 workers total; 47 tasks left in task-pool. [ Info: running task 54 on process 27 (Schedulers-primary-qO2tilUH); 10 workers total; 46 tasks left in task-pool. [ Info: running task 55 on process 28 (Schedulers-primary-qO2tilUH); 10 workers total; 45 tasks left in task-pool. [ Info: running task 56 on process 29 (Schedulers-primary-qO2tilUH); 10 workers total; 44 tasks left in task-pool. [ Info: running task 57 on process 18 (Schedulers-primary-qO2tilUH); 10 workers total; 43 tasks left in task-pool. [ Info: running task 58 on process 21 (Schedulers-primary-qO2tilUH); 10 workers total; 42 tasks left in task-pool. [ Info: running task 59 on process 22 (Schedulers-primary-qO2tilUH); 10 workers total; 41 tasks left in task-pool. [ Info: running task 60 on process 23 (Schedulers-primary-qO2tilUH); 10 workers total; 40 tasks left in task-pool. [ Info: running task 61 on process 24 (Schedulers-primary-qO2tilUH); 10 workers total; 39 tasks left in task-pool. [ Info: running task 62 on process 25 (Schedulers-primary-qO2tilUH); 10 workers total; 38 tasks left in task-pool. [ Info: running task 63 on process 26 (Schedulers-primary-qO2tilUH); 10 workers total; 37 tasks left in task-pool. [ Info: running task 64 on process 27 (Schedulers-primary-qO2tilUH); 10 workers total; 36 tasks left in task-pool. [ Info: running task 65 on process 28 (Schedulers-primary-qO2tilUH); 10 workers total; 35 tasks left in task-pool. [ Info: running task 66 on process 29 (Schedulers-primary-qO2tilUH); 10 workers total; 34 tasks left in task-pool. [ Info: running task 67 on process 18 (Schedulers-primary-qO2tilUH); 10 workers total; 33 tasks left in task-pool. [ Info: running task 68 on process 21 (Schedulers-primary-qO2tilUH); 10 workers total; 32 tasks left in task-pool. [ Info: running task 69 on process 22 (Schedulers-primary-qO2tilUH); 10 workers total; 31 tasks left in task-pool. [ Info: running task 70 on process 23 (Schedulers-primary-qO2tilUH); 10 workers total; 30 tasks left in task-pool. [ Info: running task 71 on process 24 (Schedulers-primary-qO2tilUH); 10 workers total; 29 tasks left in task-pool. [ Info: running task 72 on process 25 (Schedulers-primary-qO2tilUH); 10 workers total; 28 tasks left in task-pool. [ Info: running task 73 on process 26 (Schedulers-primary-qO2tilUH); 10 workers total; 27 tasks left in task-pool. [ Info: running task 74 on process 27 (Schedulers-primary-qO2tilUH); 10 workers total; 26 tasks left in task-pool. [ Info: running task 75 on process 28 (Schedulers-primary-qO2tilUH); 10 workers total; 25 tasks left in task-pool. [ Info: running task 76 on process 29 (Schedulers-primary-qO2tilUH); 10 workers total; 24 tasks left in task-pool. [ Info: running task 77 on process 18 (Schedulers-primary-qO2tilUH); 10 workers total; 23 tasks left in task-pool. [ Info: running task 78 on process 21 (Schedulers-primary-qO2tilUH); 10 workers total; 22 tasks left in task-pool. [ Info: running task 79 on process 22 (Schedulers-primary-qO2tilUH); 10 workers total; 21 tasks left in task-pool. [ Info: running task 80 on process 23 (Schedulers-primary-qO2tilUH); 10 workers total; 20 tasks left in task-pool. [ Info: running task 81 on process 24 (Schedulers-primary-qO2tilUH); 10 workers total; 19 tasks left in task-pool. [ Info: running task 82 on process 25 (Schedulers-primary-qO2tilUH); 10 workers total; 18 tasks left in task-pool. [ Info: running task 83 on process 26 (Schedulers-primary-qO2tilUH); 10 workers total; 17 tasks left in task-pool. [ Info: running task 84 on process 27 (Schedulers-primary-qO2tilUH); 10 workers total; 16 tasks left in task-pool. [ Info: running task 85 on process 28 (Schedulers-primary-qO2tilUH); 10 workers total; 15 tasks left in task-pool. [ Info: running task 86 on process 29 (Schedulers-primary-qO2tilUH); 10 workers total; 14 tasks left in task-pool. [ Info: running task 87 on process 18 (Schedulers-primary-qO2tilUH); 10 workers total; 13 tasks left in task-pool. [ Info: running task 88 on process 23 (Schedulers-primary-qO2tilUH); 10 workers total; 12 tasks left in task-pool. [ Info: running task 89 on process 21 (Schedulers-primary-qO2tilUH); 10 workers total; 11 tasks left in task-pool. [ Info: running task 90 on process 22 (Schedulers-primary-qO2tilUH); 10 workers total; 10 tasks left in task-pool. [ Info: running task 91 on process 24 (Schedulers-primary-qO2tilUH); 10 workers total; 9 tasks left in task-pool. [ Info: running task 92 on process 25 (Schedulers-primary-qO2tilUH); 10 workers total; 8 tasks left in task-pool. [ Info: running task 93 on process 26 (Schedulers-primary-qO2tilUH); 10 workers total; 7 tasks left in task-pool. [ Info: running task 94 on process 27 (Schedulers-primary-qO2tilUH); 10 workers total; 6 tasks left in task-pool. [ Info: running task 95 on process 28 (Schedulers-primary-qO2tilUH); 10 workers total; 5 tasks left in task-pool. [ Info: running task 96 on process 29 (Schedulers-primary-qO2tilUH); 10 workers total; 4 tasks left in task-pool. [ Info: running task 97 on process 18 (Schedulers-primary-qO2tilUH); 10 workers total; 3 tasks left in task-pool. [ Info: running task 98 on process 23 (Schedulers-primary-qO2tilUH); 10 workers total; 2 tasks left in task-pool. [ Info: running task 99 on process 21 (Schedulers-primary-qO2tilUH); 10 workers total; 1 tasks left in task-pool. [ Info: running task 100 on process 22 (Schedulers-primary-qO2tilUH); 10 workers total; 0 tasks left in task-pool. Test Summary: | Pass Total Time pmap, elastic cluster with faults | 111 111 5m19.8s [ Info: running task 1 on process 30 (Schedulers-primary-qO2tilUH); 10 workers total; 104 tasks left in task-pool. [ Info: running task 2 on process 31 (Schedulers-primary-qO2tilUH); 10 workers total; 95 tasks left in task-pool. [ Info: running task 3 on process 32 (Schedulers-primary-qO2tilUH); 10 workers total; 95 tasks left in task-pool. [ Info: running task 4 on process 33 (Schedulers-primary-qO2tilUH); 10 workers total; 95 tasks left in task-pool. [ Info: running task 5 on process 34 (Schedulers-primary-qO2tilUH); 10 workers total; 95 tasks left in task-pool. [ Info: running task 6 on process 35 (Schedulers-primary-qO2tilUH); 10 workers total; 95 tasks left in task-pool. [ Info: running task 7 on process 36 (Schedulers-primary-qO2tilUH); 10 workers total; 95 tasks left in task-pool. [ Info: running task 8 on process 37 (Schedulers-primary-qO2tilUH); 10 workers total; 95 tasks left in task-pool. [ Info: running task 9 on process 38 (Schedulers-primary-qO2tilUH); 10 workers total; 95 tasks left in task-pool. [ Info: running task 10 on process 39 (Schedulers-primary-qO2tilUH); 10 workers total; 95 tasks left in task-pool. [ Info: running task 11 on process 30 (Schedulers-primary-qO2tilUH); 10 workers total; 94 tasks left in task-pool. [ Info: running task 12 on process 31 (Schedulers-primary-qO2tilUH); 10 workers total; 93 tasks left in task-pool. [ Info: running task 13 on process 32 (Schedulers-primary-qO2tilUH); 10 workers total; 91 tasks left in task-pool. [ Info: running task 14 on process 33 (Schedulers-primary-qO2tilUH); 10 workers total; 91 tasks left in task-pool. [ Info: running task 15 on process 34 (Schedulers-primary-qO2tilUH); 10 workers total; 90 tasks left in task-pool. [ Info: running task 16 on process 35 (Schedulers-primary-qO2tilUH); 10 workers total; 85 tasks left in task-pool. [ Info: running task 17 on process 36 (Schedulers-primary-qO2tilUH); 10 workers total; 85 tasks left in task-pool. [ Info: running task 18 on process 37 (Schedulers-primary-qO2tilUH); 10 workers total; 85 tasks left in task-pool. [ Info: running task 19 on process 38 (Schedulers-primary-qO2tilUH); 10 workers total; 85 tasks left in task-pool. [ Info: running task 20 on process 39 (Schedulers-primary-qO2tilUH); 10 workers total; 85 tasks left in task-pool. [ Info: running task 21 on process 30 (Schedulers-primary-qO2tilUH); 10 workers total; 84 tasks left in task-pool. [ Info: running task 22 on process 31 (Schedulers-primary-qO2tilUH); 10 workers total; 83 tasks left in task-pool. [ Info: running task 23 on process 32 (Schedulers-primary-qO2tilUH); 10 workers total; 81 tasks left in task-pool. [ Info: running task 24 on process 33 (Schedulers-primary-qO2tilUH); 10 workers total; 81 tasks left in task-pool. [ Info: running task 25 on process 34 (Schedulers-primary-qO2tilUH); 10 workers total; 80 tasks left in task-pool. [ Info: running task 26 on process 35 (Schedulers-primary-qO2tilUH); 10 workers total; 75 tasks left in task-pool. [ Info: running task 27 on process 36 (Schedulers-primary-qO2tilUH); 10 workers total; 75 tasks left in task-pool. [ Info: running task 28 on process 37 (Schedulers-primary-qO2tilUH); 10 workers total; 75 tasks left in task-pool. [ Info: running task 29 on process 38 (Schedulers-primary-qO2tilUH); 10 workers total; 75 tasks left in task-pool. [ Info: running task 30 on process 39 (Schedulers-primary-qO2tilUH); 10 workers total; 75 tasks left in task-pool. [ Info: running task 31 on process 30 (Schedulers-primary-qO2tilUH); 10 workers total; 74 tasks left in task-pool. [ Info: running task 32 on process 31 (Schedulers-primary-qO2tilUH); 10 workers total; 73 tasks left in task-pool. [ Info: running task 33 on process 32 (Schedulers-primary-qO2tilUH); 10 workers total; 65 tasks left in task-pool. [ Info: running task 34 on process 33 (Schedulers-primary-qO2tilUH); 10 workers total; 65 tasks left in task-pool. [ Info: running task 35 on process 34 (Schedulers-primary-qO2tilUH); 10 workers total; 65 tasks left in task-pool. [ Info: running task 36 on process 35 (Schedulers-primary-qO2tilUH); 10 workers total; 65 tasks left in task-pool. [ Info: running task 37 on process 36 (Schedulers-primary-qO2tilUH); 10 workers total; 65 tasks left in task-pool. [ Info: running task 38 on process 37 (Schedulers-primary-qO2tilUH); 10 workers total; 65 tasks left in task-pool. [ Info: running task 39 on process 38 (Schedulers-primary-qO2tilUH); 10 workers total; 65 tasks left in task-pool. [ Info: running task 40 on process 39 (Schedulers-primary-qO2tilUH); 10 workers total; 65 tasks left in task-pool. [ Info: running task 41 on process 30 (Schedulers-primary-qO2tilUH); 10 workers total; 64 tasks left in task-pool. [ Info: running task 42 on process 31 (Schedulers-primary-qO2tilUH); 10 workers total; 63 tasks left in task-pool. [ Info: running task 43 on process 32 (Schedulers-primary-qO2tilUH); 10 workers total; 62 tasks left in task-pool. [ Info: running task 44 on process 33 (Schedulers-primary-qO2tilUH); 10 workers total; 55 tasks left in task-pool. [ Info: running task 45 on process 34 (Schedulers-primary-qO2tilUH); 10 workers total; 55 tasks left in task-pool. [ Info: running task 46 on process 35 (Schedulers-primary-qO2tilUH); 10 workers total; 55 tasks left in task-pool. [ Info: running task 47 on process 36 (Schedulers-primary-qO2tilUH); 10 workers total; 55 tasks left in task-pool. [ Info: running task 48 on process 37 (Schedulers-primary-qO2tilUH); 10 workers total; 55 tasks left in task-pool. [ Info: running task 49 on process 38 (Schedulers-primary-qO2tilUH); 10 workers total; 55 tasks left in task-pool. [ Info: running task 50 on process 39 (Schedulers-primary-qO2tilUH); 10 workers total; 55 tasks left in task-pool. [ Info: running task 51 on process 30 (Schedulers-primary-qO2tilUH); 10 workers total; 54 tasks left in task-pool. [ Info: running task 52 on process 31 (Schedulers-primary-qO2tilUH); 10 workers total; 53 tasks left in task-pool. [ Info: running task 53 on process 32 (Schedulers-primary-qO2tilUH); 10 workers total; 52 tasks left in task-pool. [ Info: running task 54 on process 33 (Schedulers-primary-qO2tilUH); 10 workers total; 47 tasks left in task-pool. [ Info: running task 55 on process 34 (Schedulers-primary-qO2tilUH); 10 workers total; 47 tasks left in task-pool. [ Info: running task 56 on process 35 (Schedulers-primary-qO2tilUH); 10 workers total; 47 tasks left in task-pool. [ Info: running task 57 on process 36 (Schedulers-primary-qO2tilUH); 10 workers total; 47 tasks left in task-pool. [ Info: running task 58 on process 37 (Schedulers-primary-qO2tilUH); 10 workers total; 47 tasks left in task-pool. [ Info: running task 59 on process 38 (Schedulers-primary-qO2tilUH); 10 workers total; 46 tasks left in task-pool. [ Info: running task 60 on process 39 (Schedulers-primary-qO2tilUH); 10 workers total; 45 tasks left in task-pool. [ Info: running task 61 on process 30 (Schedulers-primary-qO2tilUH); 10 workers total; 44 tasks left in task-pool. [ Info: running task 62 on process 31 (Schedulers-primary-qO2tilUH); 10 workers total; 43 tasks left in task-pool. [ Info: running task 63 on process 32 (Schedulers-primary-qO2tilUH); 10 workers total; 42 tasks left in task-pool. [ Info: running task 64 on process 33 (Schedulers-primary-qO2tilUH); 10 workers total; 41 tasks left in task-pool. [ Info: running task 65 on process 34 (Schedulers-primary-qO2tilUH); 10 workers total; 40 tasks left in task-pool. [ Info: running task 66 on process 35 (Schedulers-primary-qO2tilUH); 10 workers total; 39 tasks left in task-pool. [ Info: running task 67 on process 36 (Schedulers-primary-qO2tilUH); 10 workers total; 38 tasks left in task-pool. [ Info: running task 68 on process 37 (Schedulers-primary-qO2tilUH); 10 workers total; 35 tasks left in task-pool. [ Info: running task 69 on process 38 (Schedulers-primary-qO2tilUH); 10 workers total; 35 tasks left in task-pool. [ Info: running task 70 on process 39 (Schedulers-primary-qO2tilUH); 10 workers total; 35 tasks left in task-pool. [ Info: running task 71 on process 30 (Schedulers-primary-qO2tilUH); 10 workers total; 34 tasks left in task-pool. [ Info: running task 72 on process 31 (Schedulers-primary-qO2tilUH); 10 workers total; 33 tasks left in task-pool. [ Info: running task 73 on process 32 (Schedulers-primary-qO2tilUH); 10 workers total; 32 tasks left in task-pool. [ Info: running task 74 on process 33 (Schedulers-primary-qO2tilUH); 10 workers total; 29 tasks left in task-pool. [ Info: running task 75 on process 34 (Schedulers-primary-qO2tilUH); 10 workers total; 29 tasks left in task-pool. [ Info: running task 76 on process 35 (Schedulers-primary-qO2tilUH); 10 workers total; 29 tasks left in task-pool. [ Info: running task 77 on process 36 (Schedulers-primary-qO2tilUH); 10 workers total; 28 tasks left in task-pool. [ Info: running task 78 on process 37 (Schedulers-primary-qO2tilUH); 10 workers total; 25 tasks left in task-pool. [ Info: running task 79 on process 38 (Schedulers-primary-qO2tilUH); 10 workers total; 25 tasks left in task-pool. [ Info: running task 80 on process 39 (Schedulers-primary-qO2tilUH); 10 workers total; 25 tasks left in task-pool. [ Info: running task 81 on process 30 (Schedulers-primary-qO2tilUH); 10 workers total; 24 tasks left in task-pool. [ Info: running task 82 on process 31 (Schedulers-primary-qO2tilUH); 10 workers total; 23 tasks left in task-pool. [ Info: running task 83 on process 32 (Schedulers-primary-qO2tilUH); 10 workers total; 22 tasks left in task-pool. [ Info: running task 84 on process 33 (Schedulers-primary-qO2tilUH); 10 workers total; 19 tasks left in task-pool. [ Info: running task 85 on process 34 (Schedulers-primary-qO2tilUH); 10 workers total; 19 tasks left in task-pool. [ Info: running task 86 on process 35 (Schedulers-primary-qO2tilUH); 10 workers total; 19 tasks left in task-pool. [ Info: running task 87 on process 36 (Schedulers-primary-qO2tilUH); 10 workers total; 18 tasks left in task-pool. [ Info: running task 88 on process 37 (Schedulers-primary-qO2tilUH); 10 workers total; 15 tasks left in task-pool. [ Info: running task 89 on process 38 (Schedulers-primary-qO2tilUH); 10 workers total; 15 tasks left in task-pool. [ Info: running task 90 on process 39 (Schedulers-primary-qO2tilUH); 10 workers total; 15 tasks left in task-pool. [ Info: running task 91 on process 30 (Schedulers-primary-qO2tilUH); 10 workers total; 14 tasks left in task-pool. [ Info: running task 92 on process 31 (Schedulers-primary-qO2tilUH); 10 workers total; 13 tasks left in task-pool. [ Info: running task 93 on process 32 (Schedulers-primary-qO2tilUH); 10 workers total; 12 tasks left in task-pool. [ Info: running task 94 on process 33 (Schedulers-primary-qO2tilUH); 10 workers total; 5 tasks left in task-pool. [ Info: running task 95 on process 34 (Schedulers-primary-qO2tilUH); 10 workers total; 5 tasks left in task-pool. [ Info: running task 96 on process 35 (Schedulers-primary-qO2tilUH); 10 workers total; 5 tasks left in task-pool. [ Info: running task 97 on process 36 (Schedulers-primary-qO2tilUH); 10 workers total; 5 tasks left in task-pool. [ Info: running task 98 on process 37 (Schedulers-primary-qO2tilUH); 10 workers total; 5 tasks left in task-pool. [ Info: running task 99 on process 38 (Schedulers-primary-qO2tilUH); 10 workers total; 5 tasks left in task-pool. [ Info: running task 100 on process 39 (Schedulers-primary-qO2tilUH); 10 workers total; 5 tasks left in task-pool. [ Info: running task 101 on process 30 (Schedulers-primary-qO2tilUH); 10 workers total; 4 tasks left in task-pool. [ Info: running task 102 on process 31 (Schedulers-primary-qO2tilUH); 10 workers total; 3 tasks left in task-pool. [ Info: running task 103 on process 32 (Schedulers-primary-qO2tilUH); 10 workers total; 2 tasks left in task-pool. [ Info: running task 104 on process 33 (Schedulers-primary-qO2tilUH); 10 workers total; 1 tasks left in task-pool. [ Info: running task 105 on process 34 (Schedulers-primary-qO2tilUH); 10 workers total; 0 tasks left in task-pool. Test Summary: | Pass Total Time pmap with shrinking cluster | 116 116 5m12.9s [ Info: running task 1 on process 40 (Schedulers-primary-qO2tilUH); 5 workers total; 99 tasks left in task-pool. [ Info: running task 2 on process 41 (Schedulers-primary-qO2tilUH); 5 workers total; 98 tasks left in task-pool. [ Info: running task 3 on process 41 (Schedulers-primary-qO2tilUH); 5 workers total; 97 tasks left in task-pool. [ Info: running task 4 on process 40 (Schedulers-primary-qO2tilUH); 5 workers total; 96 tasks left in task-pool. [ Info: running task 5 on process 41 (Schedulers-primary-qO2tilUH); 5 workers total; 95 tasks left in task-pool. [ Info: running task 6 on process 40 (Schedulers-primary-qO2tilUH); 5 workers total; 94 tasks left in task-pool. [ Info: running task 7 on process 41 (Schedulers-primary-qO2tilUH); 5 workers total; 93 tasks left in task-pool. [ Info: running task 8 on process 40 (Schedulers-primary-qO2tilUH); 5 workers total; 92 tasks left in task-pool. [ Info: running task 9 on process 41 (Schedulers-primary-qO2tilUH); 5 workers total; 91 tasks left in task-pool. [ Info: running task 10 on process 40 (Schedulers-primary-qO2tilUH); 5 workers total; 90 tasks left in task-pool. [ Info: running task 11 on process 41 (Schedulers-primary-qO2tilUH); 10 workers total; 89 tasks left in task-pool. [ Info: running task 12 on process 40 (Schedulers-primary-qO2tilUH); 10 workers total; 88 tasks left in task-pool. [ Info: running task 13 on process 41 (Schedulers-primary-qO2tilUH); 10 workers total; 87 tasks left in task-pool. [ Info: running task 14 on process 40 (Schedulers-primary-qO2tilUH); 10 workers total; 86 tasks left in task-pool. [ Info: running task 15 on process 42 (Schedulers-primary-qO2tilUH); 10 workers total; 85 tasks left in task-pool. [ Info: running task 16 on process 43 (Schedulers-primary-qO2tilUH); 10 workers total; 83 tasks left in task-pool. [ Info: running task 17 on process 44 (Schedulers-primary-qO2tilUH); 10 workers total; 83 tasks left in task-pool. [ Info: running task 18 on process 41 (Schedulers-primary-qO2tilUH); 10 workers total; 82 tasks left in task-pool. [ Info: running task 19 on process 40 (Schedulers-primary-qO2tilUH); 10 workers total; 81 tasks left in task-pool. [ Info: running task 20 on process 42 (Schedulers-primary-qO2tilUH); 10 workers total; 80 tasks left in task-pool. [ Info: running task 21 on process 43 (Schedulers-primary-qO2tilUH); 10 workers total; 78 tasks left in task-pool. [ Info: running task 22 on process 44 (Schedulers-primary-qO2tilUH); 10 workers total; 78 tasks left in task-pool. [ Info: running task 23 on process 41 (Schedulers-primary-qO2tilUH); 10 workers total; 77 tasks left in task-pool. ┌ Warning: addprocs/rmprocs taking longer than expected, cancelling. └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:776 [ Info: running task 24 on process 40 (Schedulers-primary-qO2tilUH); 10 workers total; 76 tasks left in task-pool. ┌ Warning: problem adding or removing processes └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:737 ┌ Warning: TaskFailedException │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1213 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#193 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#582 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch(t::Task) │ @ Base ./task.jl:525 │ [4] loop(eloop::Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, journal::Dict{String, Any}, journal_task_callback::Function, tsk_map::Task, tsk_reduce::Task) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:735 │ [5] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1046 │ [6] epmap │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1041 [inlined] │ [7] (::var"#11#12"{SchedulerOptions{String}, String})() │ @ Main ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:166 │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1213 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#193 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 ┌ Warning: problem adding or removing processes └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:737 ┌ Warning: TaskFailedException │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1213 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#193 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#582 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch(t::Task) │ @ Base ./task.jl:525 │ [4] loop(eloop::Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, journal::Dict{String, Any}, journal_task_callback::Function, tsk_map::Task, tsk_reduce::Task) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:735 │ [5] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1046 │ [6] epmap │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1041 [inlined] │ [7] (::var"#11#12"{SchedulerOptions{String}, String})() │ @ Main ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:166 │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1213 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#193 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 [ Info: running task 25 on process 41 (Schedulers-primary-qO2tilUH); 10 workers total; 75 tasks left in task-pool. [ Info: running task 26 on process 42 (Schedulers-primary-qO2tilUH); 10 workers total; 72 tasks left in task-pool. [ Info: running task 27 on process 43 (Schedulers-primary-qO2tilUH); 10 workers total; 72 tasks left in task-pool. [ Info: running task 28 on process 44 (Schedulers-primary-qO2tilUH); 10 workers total; 72 tasks left in task-pool. ┌ Warning: problem adding or removing processes └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:737 ┌ Warning: TaskFailedException │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1213 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#193 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#582 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch(t::Task) │ @ Base ./task.jl:525 │ [4] loop(eloop::Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, journal::Dict{String, Any}, journal_task_callback::Function, tsk_map::Task, tsk_reduce::Task) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:735 │ [5] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1046 │ [6] epmap │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1041 [inlined] │ [7] (::var"#11#12"{SchedulerOptions{String}, String})() │ @ Main ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:166 │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1213 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#193 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 [ Info: running task 29 on process 40 (Schedulers-primary-qO2tilUH); 10 workers total; 71 tasks left in task-pool. ┌ Warning: problem adding or removing processes └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:737 ┌ Warning: TaskFailedException │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1213 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#193 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#582 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch(t::Task) │ @ Base ./task.jl:525 │ [4] loop(eloop::Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, journal::Dict{String, Any}, journal_task_callback::Function, tsk_map::Task, tsk_reduce::Task) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:735 │ [5] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1046 │ [6] epmap │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1041 [inlined] │ [7] (::var"#11#12"{SchedulerOptions{String}, String})() │ @ Main ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:166 │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1213 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#193 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 ┌ Warning: problem adding or removing processes └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:737 ┌ Warning: TaskFailedException │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1213 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#193 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#582 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch(t::Task) │ @ Base ./task.jl:525 │ [4] loop(eloop::Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, journal::Dict{String, Any}, journal_task_callback::Function, tsk_map::Task, tsk_reduce::Task) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:735 │ [5] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1046 │ [6] epmap │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1041 [inlined] │ [7] (::var"#11#12"{SchedulerOptions{String}, String})() │ @ Main ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:166 │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1213 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#193 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 [ Info: running task 30 on process 41 (Schedulers-primary-qO2tilUH); 10 workers total; 70 tasks left in task-pool. [ Info: running task 31 on process 42 (Schedulers-primary-qO2tilUH); 10 workers total; 67 tasks left in task-pool. [ Info: running task 32 on process 43 (Schedulers-primary-qO2tilUH); 10 workers total; 67 tasks left in task-pool. [ Info: running task 33 on process 44 (Schedulers-primary-qO2tilUH); 10 workers total; 67 tasks left in task-pool. ┌ Warning: problem adding or removing processes └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:737 ┌ Warning: TaskFailedException │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1213 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#193 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#582 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch(t::Task) │ @ Base ./task.jl:525 │ [4] loop(eloop::Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, journal::Dict{String, Any}, journal_task_callback::Function, tsk_map::Task, tsk_reduce::Task) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:735 │ [5] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1046 │ [6] epmap │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1041 [inlined] │ [7] (::var"#11#12"{SchedulerOptions{String}, String})() │ @ Main ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:166 │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1213 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#193 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 [ Info: running task 34 on process 40 (Schedulers-primary-qO2tilUH); 10 workers total; 66 tasks left in task-pool. ┌ Warning: problem adding or removing processes └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:737 ┌ Warning: TaskFailedException │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1213 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#193 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#582 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch(t::Task) │ @ Base ./task.jl:525 │ [4] loop(eloop::Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, journal::Dict{String, Any}, journal_task_callback::Function, tsk_map::Task, tsk_reduce::Task) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:735 │ [5] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1046 │ [6] epmap │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1041 [inlined] │ [7] (::var"#11#12"{SchedulerOptions{String}, String})() │ @ Main ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:166 │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1213 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#193 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 ┌ Warning: problem adding or removing processes └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:737 ┌ Warning: TaskFailedException │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1213 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#193 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#582 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch(t::Task) │ @ Base ./task.jl:525 │ [4] loop(eloop::Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, journal::Dict{String, Any}, journal_task_callback::Function, tsk_map::Task, tsk_reduce::Task) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:735 │ [5] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1046 │ [6] epmap │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1041 [inlined] │ [7] (::var"#11#12"{SchedulerOptions{String}, String})() │ @ Main ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:166 │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1213 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#193 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 ┌ Warning: problem adding or removing processes └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:737 ┌ Warning: TaskFailedException │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1213 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#193 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#582 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch(t::Task) │ @ Base ./task.jl:525 │ [4] loop(eloop::Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, journal::Dict{String, Any}, journal_task_callback::Function, tsk_map::Task, tsk_reduce::Task) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:735 │ [5] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1046 │ [6] epmap │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1041 [inlined] │ [7] (::var"#11#12"{SchedulerOptions{String}, String})() │ @ Main ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:166 │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1213 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#193 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 [ Info: running task 35 on process 41 (Schedulers-primary-qO2tilUH); 10 workers total; 65 tasks left in task-pool. [ Info: running task 36 on process 42 (Schedulers-primary-qO2tilUH); 10 workers total; 61 tasks left in task-pool. [ Info: running task 37 on process 43 (Schedulers-primary-qO2tilUH); 10 workers total; 61 tasks left in task-pool. [ Info: running task 38 on process 44 (Schedulers-primary-qO2tilUH); 10 workers total; 61 tasks left in task-pool. [ Info: running task 39 on process 40 (Schedulers-primary-qO2tilUH); 10 workers total; 61 tasks left in task-pool. ┌ Warning: problem adding or removing processes └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:737 ┌ Warning: TaskFailedException │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1213 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#193 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#582 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch(t::Task) │ @ Base ./task.jl:525 │ [4] loop(eloop::Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, journal::Dict{String, Any}, journal_task_callback::Function, tsk_map::Task, tsk_reduce::Task) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:735 │ [5] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1046 │ [6] epmap │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1041 [inlined] │ [7] (::var"#11#12"{SchedulerOptions{String}, String})() │ @ Main ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:166 │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1213 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#193 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 ┌ Warning: problem adding or removing processes └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:737 ┌ Warning: TaskFailedException │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1213 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#193 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#582 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch(t::Task) │ @ Base ./task.jl:525 │ [4] loop(eloop::Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, journal::Dict{String, Any}, journal_task_callback::Function, tsk_map::Task, tsk_reduce::Task) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:735 │ [5] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1046 │ [6] epmap │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1041 [inlined] │ [7] (::var"#11#12"{SchedulerOptions{String}, String})() │ @ Main ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:166 │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1213 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#193 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 From worker 46: ErrorException("Process(1) - Invalid connection credentials sent by remote.")CapturedException(ErrorException("Process(1) - Invalid connection credentials sent by remote."), Any[(error(s::String) at error.jl:44, 1), (process_hdr(s::Sockets.TCPSocket, validate_cookie::Bool) at process_messages.jl:266, 1), (message_handler_loop(r_stream::Sockets.TCPSocket, w_stream::Sockets.TCPSocket, incoming::Bool) at process_messages.jl:158, 1), (process_tcp_streams(r_stream::Sockets.TCPSocket, w_stream::Sockets.TCPSocket, incoming::Bool) at process_messages.jl:133, 1), ((::Distributed.var"#process_messages##0#process_messages##1"{Sockets.TCPSocket, Sockets.TCPSocket, Bool})() at process_messages.jl:121, 1)]) From worker 46: Process(1) - Unknown remote, closing connection. Worker 46 terminated. UNHANDLED TASK ERROR: Version read failed. Connection closed by peer. Stacktrace: [1] error(s::String) @ Base ./error.jl:44 [2] process_hdr(s::Sockets.TCPSocket, validate_cookie::Bool) @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:276 [3] message_handler_loop(r_stream::Sockets.TCPSocket, w_stream::Sockets.TCPSocket, incoming::Bool) @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:158 [4] process_tcp_streams(r_stream::Sockets.TCPSocket, w_stream::Sockets.TCPSocket, incoming::Bool) @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:133 [5] (::Distributed.var"#process_messages##0#process_messages##1"{Sockets.TCPSocket, Sockets.TCPSocket, Bool})() @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:121 ┌ Warning: problem adding or removing processes └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:737 ┌ Warning: TaskFailedException │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1213 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#193 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#582 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch(t::Task) │ @ Base ./task.jl:525 │ [4] loop(eloop::Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, journal::Dict{String, Any}, journal_task_callback::Function, tsk_map::Task, tsk_reduce::Task) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:735 │ [5] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1046 │ [6] epmap │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1041 [inlined] │ [7] (::var"#11#12"{SchedulerOptions{String}, String})() │ @ Main ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:166 │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1213 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#193 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", var"#9#10", var"#7#8", Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 [ Info: running task 40 on process 41 (Schedulers-primary-qO2tilUH); 9 workers total; 60 tasks left in task-pool. [ Info: running task 41 on process 42 (Schedulers-primary-qO2tilUH); 9 workers total; 56 tasks left in task-pool. [ Info: running task 42 on process 43 (Schedulers-primary-qO2tilUH); 9 workers total; 56 tasks left in task-pool. [ Info: running task 43 on process 44 (Schedulers-primary-qO2tilUH); 9 workers total; 56 tasks left in task-pool. [ Info: running task 44 on process 40 (Schedulers-primary-qO2tilUH); 9 workers total; 56 tasks left in task-pool. [ Info: running task 45 on process 45 (Schedulers-primary-qO2tilUH); 9 workers total; 55 tasks left in task-pool. [ Info: running task 46 on process 47 (Schedulers-primary-qO2tilUH); 9 workers total; 52 tasks left in task-pool. [ Info: running task 47 on process 48 (Schedulers-primary-qO2tilUH); 9 workers total; 52 tasks left in task-pool. [ Info: running task 48 on process 49 (Schedulers-primary-qO2tilUH); 9 workers total; 52 tasks left in task-pool. [ Info: running task 49 on process 41 (Schedulers-primary-qO2tilUH); 10 workers total; 51 tasks left in task-pool. [ Info: running task 50 on process 42 (Schedulers-primary-qO2tilUH); 10 workers total; 49 tasks left in task-pool. [ Info: running task 51 on process 43 (Schedulers-primary-qO2tilUH); 10 workers total; 49 tasks left in task-pool. [ Info: running task 52 on process 44 (Schedulers-primary-qO2tilUH); 10 workers total; 48 tasks left in task-pool. [ Info: running task 53 on process 40 (Schedulers-primary-qO2tilUH); 10 workers total; 47 tasks left in task-pool. [ Info: running task 54 on process 45 (Schedulers-primary-qO2tilUH); 10 workers total; 46 tasks left in task-pool. [ Info: running task 55 on process 47 (Schedulers-primary-qO2tilUH); 10 workers total; 43 tasks left in task-pool. [ Info: running task 56 on process 48 (Schedulers-primary-qO2tilUH); 10 workers total; 43 tasks left in task-pool. [ Info: running task 57 on process 49 (Schedulers-primary-qO2tilUH); 10 workers total; 43 tasks left in task-pool. [ Info: running task 58 on process 50 (Schedulers-primary-qO2tilUH); 10 workers total; 42 tasks left in task-pool. [ Info: running task 59 on process 41 (Schedulers-primary-qO2tilUH); 10 workers total; 41 tasks left in task-pool. [ Info: running task 60 on process 42 (Schedulers-primary-qO2tilUH); 10 workers total; 37 tasks left in task-pool. [ Info: running task 61 on process 43 (Schedulers-primary-qO2tilUH); 10 workers total; 37 tasks left in task-pool. [ Info: running task 62 on process 44 (Schedulers-primary-qO2tilUH); 10 workers total; 37 tasks left in task-pool. [ Info: running task 63 on process 40 (Schedulers-primary-qO2tilUH); 10 workers total; 37 tasks left in task-pool. [ Info: running task 64 on process 45 (Schedulers-primary-qO2tilUH); 10 workers total; 36 tasks left in task-pool. [ Info: running task 65 on process 47 (Schedulers-primary-qO2tilUH); 10 workers total; 33 tasks left in task-pool. [ Info: running task 66 on process 48 (Schedulers-primary-qO2tilUH); 10 workers total; 33 tasks left in task-pool. [ Info: running task 67 on process 49 (Schedulers-primary-qO2tilUH); 10 workers total; 33 tasks left in task-pool. [ Info: running task 68 on process 50 (Schedulers-primary-qO2tilUH); 10 workers total; 32 tasks left in task-pool. [ Info: running task 69 on process 41 (Schedulers-primary-qO2tilUH); 10 workers total; 31 tasks left in task-pool. [ Info: running task 70 on process 42 (Schedulers-primary-qO2tilUH); 10 workers total; 27 tasks left in task-pool. [ Info: running task 71 on process 43 (Schedulers-primary-qO2tilUH); 10 workers total; 27 tasks left in task-pool. [ Info: running task 72 on process 44 (Schedulers-primary-qO2tilUH); 10 workers total; 27 tasks left in task-pool. [ Info: running task 73 on process 40 (Schedulers-primary-qO2tilUH); 10 workers total; 27 tasks left in task-pool. From worker 46: Master process (id 1) could not connect within 120.0 seconds. From worker 46: exiting. [ Info: running task 74 on process 45 (Schedulers-primary-qO2tilUH); 10 workers total; 26 tasks left in task-pool. [ Info: running task 75 on process 47 (Schedulers-primary-qO2tilUH); 10 workers total; 24 tasks left in task-pool. [ Info: running task 76 on process 48 (Schedulers-primary-qO2tilUH); 10 workers total; 24 tasks left in task-pool. [ Info: running task 77 on process 49 (Schedulers-primary-qO2tilUH); 10 workers total; 23 tasks left in task-pool. [ Info: running task 78 on process 50 (Schedulers-primary-qO2tilUH); 10 workers total; 22 tasks left in task-pool. [ Info: running task 79 on process 41 (Schedulers-primary-qO2tilUH); 10 workers total; 21 tasks left in task-pool. [ Info: running task 80 on process 42 (Schedulers-primary-qO2tilUH); 10 workers total; 20 tasks left in task-pool. [ Info: running task 81 on process 43 (Schedulers-primary-qO2tilUH); 10 workers total; 19 tasks left in task-pool. [ Info: running task 82 on process 44 (Schedulers-primary-qO2tilUH); 10 workers total; 18 tasks left in task-pool. [ Info: running task 83 on process 40 (Schedulers-primary-qO2tilUH); 10 workers total; 17 tasks left in task-pool. [ Info: running task 84 on process 45 (Schedulers-primary-qO2tilUH); 10 workers total; 16 tasks left in task-pool. [ Info: running task 85 on process 47 (Schedulers-primary-qO2tilUH); 10 workers total; 14 tasks left in task-pool. [ Info: running task 86 on process 48 (Schedulers-primary-qO2tilUH); 10 workers total; 14 tasks left in task-pool. [ Info: running task 87 on process 49 (Schedulers-primary-qO2tilUH); 10 workers total; 13 tasks left in task-pool. [ Info: running task 88 on process 50 (Schedulers-primary-qO2tilUH); 10 workers total; 12 tasks left in task-pool. [ Info: running task 89 on process 41 (Schedulers-primary-qO2tilUH); 10 workers total; 11 tasks left in task-pool. [ Info: running task 90 on process 42 (Schedulers-primary-qO2tilUH); 10 workers total; 10 tasks left in task-pool. [ Info: running task 91 on process 43 (Schedulers-primary-qO2tilUH); 10 workers total; 7 tasks left in task-pool. [ Info: running task 92 on process 44 (Schedulers-primary-qO2tilUH); 10 workers total; 7 tasks left in task-pool. [ Info: running task 93 on process 40 (Schedulers-primary-qO2tilUH); 10 workers total; 7 tasks left in task-pool. [ Info: running task 94 on process 45 (Schedulers-primary-qO2tilUH); 10 workers total; 6 tasks left in task-pool. [ Info: running task 95 on process 47 (Schedulers-primary-qO2tilUH); 10 workers total; 3 tasks left in task-pool. [ Info: running task 96 on process 48 (Schedulers-primary-qO2tilUH); 10 workers total; 3 tasks left in task-pool. [ Info: running task 97 on process 49 (Schedulers-primary-qO2tilUH); 10 workers total; 3 tasks left in task-pool. [ Info: running task 98 on process 50 (Schedulers-primary-qO2tilUH); 10 workers total; 2 tasks left in task-pool. [ Info: running task 99 on process 41 (Schedulers-primary-qO2tilUH); 10 workers total; 1 tasks left in task-pool. [ Info: running task 100 on process 42 (Schedulers-primary-qO2tilUH); 10 workers total; 0 tasks left in task-pool. Test Summary: | Pass Total Time pmap with interactive growing cluster | 111 111 4m44.4s [ Info: nworkers()=1, i=1 ┌ Warning: problem adding or removing processes └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:737 ┌ Warning: TaskFailedException │ │ nested task error: MethodError: no method matching addprocs(::Int64, ::String) │ The function `addprocs` exists, but no method is defined for this combination of argument types. │ │ Closest candidates are: │ addprocs(::Integer; restrict, kwargs...) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ addprocs(; ...) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ addprocs(!Matched::AbstractVector; kwargs...) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:158 │ ... │ │ Stacktrace: │ [1] (::var"#myaddprocs#myaddprocs##0")(n::Int64) │ @ Main ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:208 │ [2] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{var"#myaddprocs#myaddprocs##0", Schedulers.var"#14#15", typeof(nworkers), Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#94#95"{Int64}, Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#582 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch(t::Task) │ @ Base ./task.jl:525 │ [4] loop(eloop::Schedulers.ElasticLoop{var"#myaddprocs#myaddprocs##0", Schedulers.var"#14#15", typeof(nworkers), Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#94#95"{Int64}, Int64, Nothing}, journal::Dict{String, Any}, journal_task_callback::Function, tsk_map::Task, tsk_reduce::Task) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:735 │ [5] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1046 │ [6] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1041 │ [7] top-level scope │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:195 │ [8] macro expansion │ @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1833 [inlined] │ [9] macro expansion │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:218 [inlined] │ [10] include(mapexpr::Function, mod::Module, _path::String) │ @ Base ./Base.jl:310 │ [11] top-level scope │ @ none:6 │ [12] eval(m::Module, e::Any) │ @ Core ./boot.jl:489 │ [13] exec_options(opts::Base.JLOptions) │ @ Base ./client.jl:297 │ [14] _start() │ @ Base ./client.jl:564 │ │ nested task error: MethodError: no method matching addprocs(::Int64, ::String) │ The function `addprocs` exists, but no method is defined for this combination of argument types. │ │ Closest candidates are: │ addprocs(::Integer; restrict, kwargs...) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ addprocs(; ...) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ addprocs(!Matched::AbstractVector; kwargs...) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:158 │ ... │ │ Stacktrace: │ [1] (::var"#myaddprocs#myaddprocs##0")(n::Int64) │ @ Main ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:208 │ [2] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{var"#myaddprocs#myaddprocs##0", Schedulers.var"#14#15", typeof(nworkers), Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#94#95"{Int64}, Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 [ Info: nworkers()=1, i=2 [ Info: nworkers()=1, i=3 [ Info: running task 1 on process 51 (Schedulers-primary-qO2tilUH); 2 workers total; 99 tasks left in task-pool. [ Info: running task 2 on process 51 (Schedulers-primary-qO2tilUH); 2 workers total; 98 tasks left in task-pool. [ Info: running task 3 on process 51 (Schedulers-primary-qO2tilUH); 2 workers total; 97 tasks left in task-pool. [ Info: nworkers()=2, i=4 [ Info: running task 4 on process 51 (Schedulers-primary-qO2tilUH); 2 workers total; 96 tasks left in task-pool. [ Info: running task 5 on process 51 (Schedulers-primary-qO2tilUH); 2 workers total; 95 tasks left in task-pool. [ Info: running task 6 on process 51 (Schedulers-primary-qO2tilUH); 2 workers total; 94 tasks left in task-pool. [ Info: running task 7 on process 51 (Schedulers-primary-qO2tilUH); 2 workers total; 93 tasks left in task-pool. [ Info: running task 8 on process 51 (Schedulers-primary-qO2tilUH); 3 workers total; 92 tasks left in task-pool. [ Info: running task 9 on process 52 (Schedulers-primary-qO2tilUH); 3 workers total; 91 tasks left in task-pool. [ Info: running task 10 on process 51 (Schedulers-primary-qO2tilUH); 3 workers total; 90 tasks left in task-pool. [ Info: running task 11 on process 51 (Schedulers-primary-qO2tilUH); 3 workers total; 89 tasks left in task-pool. [ Info: running task 12 on process 52 (Schedulers-primary-qO2tilUH); 3 workers total; 88 tasks left in task-pool. [ Info: running task 13 on process 51 (Schedulers-primary-qO2tilUH); 3 workers total; 87 tasks left in task-pool. [ Info: running task 14 on process 52 (Schedulers-primary-qO2tilUH); 3 workers total; 86 tasks left in task-pool. [ Info: running task 15 on process 51 (Schedulers-primary-qO2tilUH); 3 workers total; 85 tasks left in task-pool. [ Info: running task 16 on process 52 (Schedulers-primary-qO2tilUH); 3 workers total; 84 tasks left in task-pool. [ Info: nworkers()=3, i=5 [ Info: running task 17 on process 51 (Schedulers-primary-qO2tilUH); 3 workers total; 83 tasks left in task-pool. [ Info: running task 18 on process 52 (Schedulers-primary-qO2tilUH); 3 workers total; 82 tasks left in task-pool. [ Info: running task 19 on process 51 (Schedulers-primary-qO2tilUH); 3 workers total; 81 tasks left in task-pool. [ Info: running task 20 on process 52 (Schedulers-primary-qO2tilUH); 3 workers total; 80 tasks left in task-pool. [ Info: running task 21 on process 51 (Schedulers-primary-qO2tilUH); 3 workers total; 79 tasks left in task-pool. [ Info: running task 22 on process 52 (Schedulers-primary-qO2tilUH); 3 workers total; 78 tasks left in task-pool. [ Info: running task 23 on process 53 (Schedulers-primary-qO2tilUH); 3 workers total; 77 tasks left in task-pool. [ Info: running task 24 on process 51 (Schedulers-primary-qO2tilUH); 3 workers total; 76 tasks left in task-pool. [ Info: running task 25 on process 52 (Schedulers-primary-qO2tilUH); 3 workers total; 75 tasks left in task-pool. [ Info: running task 26 on process 53 (Schedulers-primary-qO2tilUH); 3 workers total; 74 tasks left in task-pool. [ Info: running task 27 on process 51 (Schedulers-primary-qO2tilUH); 3 workers total; 73 tasks left in task-pool. [ Info: running task 28 on process 52 (Schedulers-primary-qO2tilUH); 3 workers total; 72 tasks left in task-pool. [ Info: running task 29 on process 53 (Schedulers-primary-qO2tilUH); 3 workers total; 71 tasks left in task-pool. [ Info: running task 30 on process 51 (Schedulers-primary-qO2tilUH); 3 workers total; 70 tasks left in task-pool. [ Info: running task 31 on process 52 (Schedulers-primary-qO2tilUH); 3 workers total; 69 tasks left in task-pool. [ Info: running task 32 on process 53 (Schedulers-primary-qO2tilUH); 3 workers total; 68 tasks left in task-pool. [ Info: running task 33 on process 51 (Schedulers-primary-qO2tilUH); 3 workers total; 67 tasks left in task-pool. [ Info: running task 34 on process 52 (Schedulers-primary-qO2tilUH); 3 workers total; 66 tasks left in task-pool. [ Info: running task 35 on process 53 (Schedulers-primary-qO2tilUH); 3 workers total; 65 tasks left in task-pool. [ Info: running task 36 on process 51 (Schedulers-primary-qO2tilUH); 3 workers total; 64 tasks left in task-pool. [ Info: running task 37 on process 52 (Schedulers-primary-qO2tilUH); 3 workers total; 63 tasks left in task-pool. [ Info: running task 38 on process 53 (Schedulers-primary-qO2tilUH); 3 workers total; 62 tasks left in task-pool. [ Info: running task 39 on process 51 (Schedulers-primary-qO2tilUH); 3 workers total; 61 tasks left in task-pool. [ Info: running task 40 on process 52 (Schedulers-primary-qO2tilUH); 3 workers total; 60 tasks left in task-pool. [ Info: running task 41 on process 53 (Schedulers-primary-qO2tilUH); 3 workers total; 59 tasks left in task-pool. [ Info: running task 42 on process 51 (Schedulers-primary-qO2tilUH); 3 workers total; 58 tasks left in task-pool. [ Info: running task 43 on process 52 (Schedulers-primary-qO2tilUH); 3 workers total; 57 tasks left in task-pool. [ Info: running task 44 on process 53 (Schedulers-primary-qO2tilUH); 3 workers total; 56 tasks left in task-pool. [ Info: running task 45 on process 51 (Schedulers-primary-qO2tilUH); 3 workers total; 55 tasks left in task-pool. [ Info: running task 46 on process 52 (Schedulers-primary-qO2tilUH); 3 workers total; 54 tasks left in task-pool. [ Info: running task 47 on process 53 (Schedulers-primary-qO2tilUH); 3 workers total; 53 tasks left in task-pool. [ Info: running task 48 on process 51 (Schedulers-primary-qO2tilUH); 3 workers total; 52 tasks left in task-pool. [ Info: running task 49 on process 52 (Schedulers-primary-qO2tilUH); 3 workers total; 51 tasks left in task-pool. [ Info: running task 50 on process 53 (Schedulers-primary-qO2tilUH); 3 workers total; 50 tasks left in task-pool. [ Info: running task 51 on process 51 (Schedulers-primary-qO2tilUH); 3 workers total; 49 tasks left in task-pool. [ Info: running task 52 on process 52 (Schedulers-primary-qO2tilUH); 3 workers total; 48 tasks left in task-pool. [ Info: running task 53 on process 53 (Schedulers-primary-qO2tilUH); 3 workers total; 47 tasks left in task-pool. [ Info: running task 54 on process 51 (Schedulers-primary-qO2tilUH); 3 workers total; 46 tasks left in task-pool. [ Info: running task 55 on process 52 (Schedulers-primary-qO2tilUH); 3 workers total; 45 tasks left in task-pool. [ Info: running task 56 on process 53 (Schedulers-primary-qO2tilUH); 3 workers total; 44 tasks left in task-pool. [ Info: running task 57 on process 51 (Schedulers-primary-qO2tilUH); 3 workers total; 43 tasks left in task-pool. [ Info: running task 58 on process 52 (Schedulers-primary-qO2tilUH); 3 workers total; 42 tasks left in task-pool. [ Info: running task 59 on process 53 (Schedulers-primary-qO2tilUH); 3 workers total; 41 tasks left in task-pool. [ Info: running task 60 on process 51 (Schedulers-primary-qO2tilUH); 3 workers total; 40 tasks left in task-pool. [ Info: running task 61 on process 52 (Schedulers-primary-qO2tilUH); 3 workers total; 39 tasks left in task-pool. [ Info: running task 62 on process 53 (Schedulers-primary-qO2tilUH); 3 workers total; 38 tasks left in task-pool. [ Info: running task 63 on process 51 (Schedulers-primary-qO2tilUH); 3 workers total; 37 tasks left in task-pool. [ Info: running task 64 on process 52 (Schedulers-primary-qO2tilUH); 3 workers total; 36 tasks left in task-pool. [ Info: running task 65 on process 53 (Schedulers-primary-qO2tilUH); 3 workers total; 35 tasks left in task-pool. [ Info: running task 66 on process 51 (Schedulers-primary-qO2tilUH); 3 workers total; 34 tasks left in task-pool. [ Info: running task 67 on process 52 (Schedulers-primary-qO2tilUH); 3 workers total; 33 tasks left in task-pool. [ Info: running task 68 on process 53 (Schedulers-primary-qO2tilUH); 3 workers total; 32 tasks left in task-pool. [ Info: running task 69 on process 51 (Schedulers-primary-qO2tilUH); 3 workers total; 31 tasks left in task-pool. [ Info: running task 70 on process 52 (Schedulers-primary-qO2tilUH); 3 workers total; 30 tasks left in task-pool. [ Info: running task 71 on process 53 (Schedulers-primary-qO2tilUH); 3 workers total; 29 tasks left in task-pool. [ Info: running task 72 on process 51 (Schedulers-primary-qO2tilUH); 3 workers total; 28 tasks left in task-pool. [ Info: running task 73 on process 52 (Schedulers-primary-qO2tilUH); 3 workers total; 27 tasks left in task-pool. [ Info: running task 74 on process 53 (Schedulers-primary-qO2tilUH); 3 workers total; 26 tasks left in task-pool. [ Info: running task 75 on process 51 (Schedulers-primary-qO2tilUH); 3 workers total; 25 tasks left in task-pool. [ Info: running task 76 on process 52 (Schedulers-primary-qO2tilUH); 3 workers total; 24 tasks left in task-pool. [ Info: running task 77 on process 53 (Schedulers-primary-qO2tilUH); 3 workers total; 23 tasks left in task-pool. [ Info: running task 78 on process 51 (Schedulers-primary-qO2tilUH); 3 workers total; 22 tasks left in task-pool. [ Info: running task 79 on process 52 (Schedulers-primary-qO2tilUH); 3 workers total; 21 tasks left in task-pool. [ Info: running task 80 on process 53 (Schedulers-primary-qO2tilUH); 3 workers total; 20 tasks left in task-pool. [ Info: running task 81 on process 51 (Schedulers-primary-qO2tilUH); 3 workers total; 19 tasks left in task-pool. [ Info: running task 82 on process 52 (Schedulers-primary-qO2tilUH); 3 workers total; 18 tasks left in task-pool. [ Info: running task 83 on process 53 (Schedulers-primary-qO2tilUH); 3 workers total; 17 tasks left in task-pool. [ Info: running task 84 on process 51 (Schedulers-primary-qO2tilUH); 3 workers total; 16 tasks left in task-pool. [ Info: running task 85 on process 52 (Schedulers-primary-qO2tilUH); 3 workers total; 15 tasks left in task-pool. [ Info: running task 86 on process 53 (Schedulers-primary-qO2tilUH); 3 workers total; 14 tasks left in task-pool. [ Info: running task 87 on process 51 (Schedulers-primary-qO2tilUH); 3 workers total; 13 tasks left in task-pool. [ Info: running task 88 on process 52 (Schedulers-primary-qO2tilUH); 3 workers total; 12 tasks left in task-pool. [ Info: running task 89 on process 53 (Schedulers-primary-qO2tilUH); 3 workers total; 11 tasks left in task-pool. [ Info: running task 90 on process 51 (Schedulers-primary-qO2tilUH); 3 workers total; 10 tasks left in task-pool. [ Info: running task 91 on process 52 (Schedulers-primary-qO2tilUH); 3 workers total; 9 tasks left in task-pool. [ Info: running task 92 on process 53 (Schedulers-primary-qO2tilUH); 3 workers total; 8 tasks left in task-pool. [ Info: running task 93 on process 51 (Schedulers-primary-qO2tilUH); 3 workers total; 7 tasks left in task-pool. [ Info: running task 94 on process 52 (Schedulers-primary-qO2tilUH); 3 workers total; 6 tasks left in task-pool. [ Info: running task 95 on process 53 (Schedulers-primary-qO2tilUH); 3 workers total; 5 tasks left in task-pool. [ Info: running task 96 on process 51 (Schedulers-primary-qO2tilUH); 3 workers total; 4 tasks left in task-pool. [ Info: running task 97 on process 52 (Schedulers-primary-qO2tilUH); 3 workers total; 3 tasks left in task-pool. [ Info: running task 98 on process 53 (Schedulers-primary-qO2tilUH); 3 workers total; 2 tasks left in task-pool. [ Info: running task 99 on process 51 (Schedulers-primary-qO2tilUH); 3 workers total; 1 tasks left in task-pool. [ Info: running task 100 on process 52 (Schedulers-primary-qO2tilUH); 3 workers total; 0 tasks left in task-pool. ┌ Warning: addprocs/rmprocs taking longer than expected, cancelling. └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:823 Test Summary: | Pass Total Time pmap with blocking addprocs | 103 103 2m12.2s ┌ Warning: addprocs/rmprocs taking longer than expected, cancelling. └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:776 ┌ Warning: problem adding or removing processes └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:737 ┌ Warning: TaskFailedException │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1213 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#193 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", typeof(nworkers), Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#582 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch(t::Task) │ @ Base ./task.jl:525 │ [4] loop(eloop::Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", typeof(nworkers), Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, journal::Dict{String, Any}, journal_task_callback::Function, tsk_map::Task, tsk_reduce::Task) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:735 │ [5] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1046 │ [6] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1041 │ [7] top-level scope │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:241 │ [8] macro expansion │ @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1833 [inlined] │ [9] macro expansion │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:256 [inlined] │ [10] include(mapexpr::Function, mod::Module, _path::String) │ @ Base ./Base.jl:310 │ [11] top-level scope │ @ none:6 │ [12] eval(m::Module, e::Any) │ @ Core ./boot.jl:489 │ [13] exec_options(opts::Base.JLOptions) │ @ Base ./client.jl:297 │ [14] _start() │ @ Base ./client.jl:564 │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1213 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#193 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", typeof(nworkers), Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 ┌ Warning: problem adding or removing processes └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:737 ┌ Warning: TaskFailedException │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1213 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#193 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", typeof(nworkers), Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#582 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch(t::Task) │ @ Base ./task.jl:525 │ [4] loop(eloop::Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", typeof(nworkers), Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, journal::Dict{String, Any}, journal_task_callback::Function, tsk_map::Task, tsk_reduce::Task) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:735 │ [5] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1046 │ [6] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1041 │ [7] top-level scope │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:241 │ [8] macro expansion │ @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1833 [inlined] │ [9] macro expansion │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:256 [inlined] │ [10] include(mapexpr::Function, mod::Module, _path::String) │ @ Base ./Base.jl:310 │ [11] top-level scope │ @ none:6 │ [12] eval(m::Module, e::Any) │ @ Core ./boot.jl:489 │ [13] exec_options(opts::Base.JLOptions) │ @ Base ./client.jl:297 │ [14] _start() │ @ Base ./client.jl:564 │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1213 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#193 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", typeof(nworkers), Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 ┌ Warning: problem adding or removing processes └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:737 ┌ Warning: TaskFailedException │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1213 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#193 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", typeof(nworkers), Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#582 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch(t::Task) │ @ Base ./task.jl:525 │ [4] loop(eloop::Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", typeof(nworkers), Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, journal::Dict{String, Any}, journal_task_callback::Function, tsk_map::Task, tsk_reduce::Task) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:735 │ [5] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1046 │ [6] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1041 │ [7] top-level scope │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:241 │ [8] macro expansion │ @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1833 [inlined] │ [9] macro expansion │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:256 [inlined] │ [10] include(mapexpr::Function, mod::Module, _path::String) │ @ Base ./Base.jl:310 │ [11] top-level scope │ @ none:6 │ [12] eval(m::Module, e::Any) │ @ Core ./boot.jl:489 │ [13] exec_options(opts::Base.JLOptions) │ @ Base ./client.jl:297 │ [14] _start() │ @ Base ./client.jl:564 │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1213 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#193 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", typeof(nworkers), Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 ┌ Warning: problem adding or removing processes └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:737 ┌ Warning: TaskFailedException │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1213 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#193 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", typeof(nworkers), Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#582 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch(t::Task) │ @ Base ./task.jl:525 │ [4] loop(eloop::Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", typeof(nworkers), Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, journal::Dict{String, Any}, journal_task_callback::Function, tsk_map::Task, tsk_reduce::Task) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:735 │ [5] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1046 │ [6] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1041 │ [7] top-level scope │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:241 │ [8] macro expansion │ @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1833 [inlined] │ [9] macro expansion │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:256 [inlined] │ [10] include(mapexpr::Function, mod::Module, _path::String) │ @ Base ./Base.jl:310 │ [11] top-level scope │ @ none:6 │ [12] eval(m::Module, e::Any) │ @ Core ./boot.jl:489 │ [13] exec_options(opts::Base.JLOptions) │ @ Base ./client.jl:297 │ [14] _start() │ @ Base ./client.jl:564 │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1213 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#193 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", typeof(nworkers), Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 From worker 57: ErrorException("Process(1) - Invalid connection credentials sent by remote.")CapturedException(ErrorException("Process(1) - Invalid connection credentials sent by remote."), Any[(error(s::String) at error.jl:44, 1), (process_hdr(s::Sockets.TCPSocket, validate_cookie::Bool) at process_messages.jl:266, 1), (message_handler_loop(r_stream::Sockets.TCPSocket, w_stream::Sockets.TCPSocket, incoming::Bool) at process_messages.jl:158, 1), (process_tcp_streams(r_stream::Sockets.TCPSocket, w_stream::Sockets.TCPSocket, incoming::Bool) at process_messages.jl:133, 1), ((::Distributed.var"#process_messages##0#process_messages##1"{Sockets.TCPSocket, Sockets.TCPSocket, Bool})() at process_messages.jl:121, 1)]) From worker 57: Process(1) - Unknown remote, closing connection. From worker 56: ErrorException("Process(1) - Invalid connection credentials sent by remote.")CapturedException(ErrorException("Process(1) - Invalid connection credentials sent by remote."), Any[(error(s::String) at error.jl:44, 1), (process_hdr(s::Sockets.TCPSocket, validate_cookie::Bool) at process_messages.jl:266, 1), (message_handler_loop(r_stream::Sockets.TCPSocket, w_stream::Sockets.TCPSocket, incoming::Bool) at process_messages.jl:158, 1), (process_tcp_streams(r_stream::Sockets.TCPSocket, w_stream::Sockets.TCPSocket, incoming::Bool) at process_messages.jl:133, 1), ((::Distributed.var"#process_messages##0#process_messages##1"{Sockets.TCPSocket, Sockets.TCPSocket, Bool})() at process_messages.jl:121, 1)]) Worker 57 terminated. From worker 56: Process(1) - Unknown remote, closing connection. Worker 56 terminated.UNHANDLED TASK ERROR: Version read failed. Connection closed by peer. Stacktrace: [1] error(s::String) @ Base ./error.jl:44 [2] process_hdr(s::Sockets.TCPSocket, validate_cookie::Bool) @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:276 [3] message_handler_loop(r_stream::Sockets.TCPSocket, w_stream::Sockets.TCPSocket, incoming::Bool) @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:158 [4] process_tcp_streams(r_stream::Sockets.TCPSocket, w_stream::Sockets.TCPSocket, incoming::Bool) @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:133 [5] (::Distributed.var"#process_messages##0#process_messages##1"{Sockets.TCPSocket, Sockets.TCPSocket, Bool})() @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:121 UNHANDLED TASK ERROR: Version read failed. Connection closed by peer. Stacktrace: [1] error(s::String) @ Base ./error.jl:44 [2] process_hdr(s::Sockets.TCPSocket, validate_cookie::Bool) @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:276 [3] message_handler_loop(r_stream::Sockets.TCPSocket, w_stream::Sockets.TCPSocket, incoming::Bool) @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:158 [4] process_tcp_streams(r_stream::Sockets.TCPSocket, w_stream::Sockets.TCPSocket, incoming::Bool) @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:133 [5] (::Distributed.var"#process_messages##0#process_messages##1"{Sockets.TCPSocket, Sockets.TCPSocket, Bool})() @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:121 ┌ Warning: problem adding or removing processes └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:737 ┌ Warning: TaskFailedException │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1213 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#193 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", typeof(nworkers), Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#582 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch(t::Task) │ @ Base ./task.jl:525 │ [4] loop(eloop::Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", typeof(nworkers), Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, journal::Dict{String, Any}, journal_task_callback::Function, tsk_map::Task, tsk_reduce::Task) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:735 │ [5] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1046 │ [6] epmap(options::SchedulerOptions{String}, f::Function, tasks::UnitRange{Int64}, args::String) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1041 │ [7] top-level scope │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:241 │ [8] macro expansion │ @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1833 [inlined] │ [9] macro expansion │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:256 [inlined] │ [10] include(mapexpr::Function, mod::Module, _path::String) │ @ Base ./Base.jl:310 │ [11] top-level scope │ @ none:6 │ [12] eval(m::Module, e::Any) │ @ Core ./boot.jl:489 │ [13] exec_options(opts::Base.JLOptions) │ @ Base ./client.jl:297 │ [14] _start() │ @ Base ./client.jl:564 │ │ nested task error: InterruptException: │ Stacktrace: │ [1] try_yieldto(undo::typeof(identity)) │ @ Base ./task.jl:1128 │ [2] yieldto(t::Task, x::Any) │ @ Base ./task.jl:1109 │ [3] yieldto │ @ ./task.jl:1095 [inlined] │ [4] wait() │ @ Base ./task.jl:1213 │ [5] wait(c::Base.GenericCondition{Base.Threads.SpinLock}; first::Bool) │ @ Base ./condition.jl:141 │ [6] wait │ @ ./condition.jl:136 [inlined] │ [7] _wait(t::Task) │ @ Base ./task.jl:312 │ [8] sync_end(c::Channel{Any}) │ @ Base ./task.jl:560 │ [9] macro expansion │ @ ./task.jl:634 [inlined] │ [10] addprocs_locked(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:496 │ [11] addprocs_locked │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:461 [inlined] │ [12] addprocs(manager::Distributed.LocalManager; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:455 │ [13] addprocs │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:448 [inlined] │ [14] #addprocs#193 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:472 [inlined] │ [15] addprocs(np::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/managers.jl:469 │ [16] (::Schedulers.var"#loop##20#loop##21"{Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", typeof(nworkers), Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:768 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 [ Info: running task 1 on process 54 (Schedulers-primary-qO2tilUH); 3 workers total; 19 tasks left in task-pool. [ Info: running task 2 on process 55 (Schedulers-primary-qO2tilUH); 3 workers total; 17 tasks left in task-pool. [ Info: running task 3 on process 58 (Schedulers-primary-qO2tilUH); 3 workers total; 17 tasks left in task-pool. [ Info: running task 4 on process 54 (Schedulers-primary-qO2tilUH); 3 workers total; 16 tasks left in task-pool. [ Info: running task 5 on process 55 (Schedulers-primary-qO2tilUH); 3 workers total; 15 tasks left in task-pool. [ Info: running task 6 on process 58 (Schedulers-primary-qO2tilUH); 3 workers total; 14 tasks left in task-pool. [ Info: running task 7 on process 54 (Schedulers-primary-qO2tilUH); 3 workers total; 13 tasks left in task-pool. [ Info: running task 8 on process 55 (Schedulers-primary-qO2tilUH); 3 workers total; 12 tasks left in task-pool. [ Info: running task 9 on process 58 (Schedulers-primary-qO2tilUH); 3 workers total; 11 tasks left in task-pool. [ Info: running task 10 on process 54 (Schedulers-primary-qO2tilUH); 3 workers total; 10 tasks left in task-pool. [ Info: running task 11 on process 55 (Schedulers-primary-qO2tilUH); 3 workers total; 9 tasks left in task-pool. [ Info: running task 12 on process 58 (Schedulers-primary-qO2tilUH); 3 workers total; 8 tasks left in task-pool. [ Info: running task 13 on process 54 (Schedulers-primary-qO2tilUH); 5 workers total; 7 tasks left in task-pool. [ Info: running task 14 on process 55 (Schedulers-primary-qO2tilUH); 5 workers total; 6 tasks left in task-pool. [ Info: running task 15 on process 58 (Schedulers-primary-qO2tilUH); 5 workers total; 5 tasks left in task-pool. [ Info: running task 16 on process 54 (Schedulers-primary-qO2tilUH); 5 workers total; 4 tasks left in task-pool. [ Info: running task 17 on process 55 (Schedulers-primary-qO2tilUH); 5 workers total; 3 tasks left in task-pool. [ Info: running task 18 on process 58 (Schedulers-primary-qO2tilUH); 5 workers total; 2 tasks left in task-pool. [ Info: running task 19 on process 54 (Schedulers-primary-qO2tilUH); 5 workers total; 1 tasks left in task-pool. [ Info: running task 20 on process 55 (Schedulers-primary-qO2tilUH); 5 workers total; 0 tasks left in task-pool. ┌ Warning: caught an exception, there have been 0 failure(s) on process 55 (Schedulers-primary-qO2tilUH)... └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1111 ┌ Warning: Schedulers.TimeoutException(55, 14.02000904083252) │ error type: Schedulers.TimeoutException │ 1-element ExceptionStack: │ Schedulers.TimeoutException(55, 14.02000904083252) │ Stacktrace: │ [1] remotecall_func_wait_timeout(::Vector{Float64}, ::Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", typeof(nworkers), Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, ::SchedulerOptions{String}, ::Nothing, ::Function, ::Function, ::Int64, ::var"#foo5c#foo5c##0", ::Int64, ::Int64, ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:343 │ [2] remotecall_func_wait_timeout(::Vector{Float64}, ::Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", typeof(nworkers), Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, ::SchedulerOptions{String}, ::Nothing, ::Function, ::Function, ::Int64, ::Function, ::Int64, ::Int64, ::Vararg{Any}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:338 │ [3] (::Schedulers.var"#107#108"{@Kwargs{}, SchedulerOptions{String}, var"#foo5c#foo5c##0", Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", typeof(nworkers), Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, Dict{String, Any}, Tuple{String}, Vector{Float64}, Int64})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1105 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 [ Info: running task 20 on process 54 (Schedulers-primary-qO2tilUH); 5 workers total; 0 tasks left in task-pool. Test Summary: | Pass Total Time pmap with timeout | 21 21 2m13.7s From worker 56: Master process (id 1) could not connect within 120.0 seconds. From worker 56: exiting. From worker 57: Master process (id 1) could not connect within 120.0 seconds. From worker 57: exiting. [ Info: running task 1 on process 54 (Schedulers-primary-qO2tilUH); 4 workers total; 19 tasks left in task-pool. [ Info: running task 2 on process 58 (Schedulers-primary-qO2tilUH); 4 workers total; 18 tasks left in task-pool. [ Info: running task 3 on process 59 (Schedulers-primary-qO2tilUH); 4 workers total; 17 tasks left in task-pool. [ Info: running task 4 on process 60 (Schedulers-primary-qO2tilUH); 4 workers total; 16 tasks left in task-pool. [ Info: running task 5 on process 54 (Schedulers-primary-qO2tilUH); 4 workers total; 15 tasks left in task-pool. [ Info: running task 6 on process 58 (Schedulers-primary-qO2tilUH); 4 workers total; 14 tasks left in task-pool. [ Info: running task 7 on process 54 (Schedulers-primary-qO2tilUH); 4 workers total; 13 tasks left in task-pool. [ Info: running task 8 on process 58 (Schedulers-primary-qO2tilUH); 4 workers total; 12 tasks left in task-pool. [ Info: running task 9 on process 54 (Schedulers-primary-qO2tilUH); 5 workers total; 11 tasks left in task-pool. [ Info: running task 10 on process 58 (Schedulers-primary-qO2tilUH); 5 workers total; 10 tasks left in task-pool. [ Info: running task 11 on process 59 (Schedulers-primary-qO2tilUH); 5 workers total; 9 tasks left in task-pool. [ Info: running task 12 on process 60 (Schedulers-primary-qO2tilUH); 5 workers total; 8 tasks left in task-pool. [ Info: running task 13 on process 54 (Schedulers-primary-qO2tilUH); 5 workers total; 7 tasks left in task-pool. [ Info: running task 14 on process 58 (Schedulers-primary-qO2tilUH); 5 workers total; 6 tasks left in task-pool. [ Info: running task 15 on process 59 (Schedulers-primary-qO2tilUH); 5 workers total; 5 tasks left in task-pool. [ Info: running task 16 on process 60 (Schedulers-primary-qO2tilUH); 5 workers total; 4 tasks left in task-pool. [ Info: running task 17 on process 54 (Schedulers-primary-qO2tilUH); 5 workers total; 3 tasks left in task-pool. [ Info: running task 18 on process 58 (Schedulers-primary-qO2tilUH); 5 workers total; 2 tasks left in task-pool. [ Info: running task 19 on process 59 (Schedulers-primary-qO2tilUH); 5 workers total; 1 tasks left in task-pool. [ Info: running task 20 on process 60 (Schedulers-primary-qO2tilUH); 5 workers total; 0 tasks left in task-pool. ┌ Warning: caught an exception, there have been 0 failure(s) on process 60 (Schedulers-primary-qO2tilUH)... └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1111 ┌ Warning: skipping task '20' that timed out └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1114 ┌ Warning: Schedulers.TimeoutException(60, 12.024520874023438) │ error type: Schedulers.TimeoutException │ 1-element ExceptionStack: │ Schedulers.TimeoutException(60, 12.024520874023438) │ Stacktrace: │ [1] remotecall_func_wait_timeout(::Vector{Float64}, ::Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", typeof(nworkers), Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, ::SchedulerOptions{String}, ::Nothing, ::Function, ::Function, ::Int64, ::var"#foo5c#foo5c##1", ::Int64, ::Int64, ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:343 │ [2] remotecall_func_wait_timeout(::Vector{Float64}, ::Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", typeof(nworkers), Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, ::SchedulerOptions{String}, ::Nothing, ::Function, ::Function, ::Int64, ::Function, ::Int64, ::Int64, ::Vararg{Any}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:338 │ [3] (::Schedulers.var"#107#108"{@Kwargs{}, SchedulerOptions{String}, var"#foo5c#foo5c##1", Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", typeof(nworkers), Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, Dict{String, Any}, Tuple{String}, Vector{Float64}, Int64})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1105 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 Test Summary: | Pass Total Time pmap with timeout and skip_tasks_that_timeout=true | 21 21 31.8s ┌ Warning: problem initializing 59, removing 59 from cluster. └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:675 ┌ Warning: no process with id 59 exists │ error type: ErrorException │ 1-element ExceptionStack: │ no process with id 59 exists │ Stacktrace: │ [1] error(s::String) │ @ Base ./error.jl:44 │ [2] worker_from_id(pg::Distributed.ProcessGroup, i::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:1114 │ [3] worker_from_id │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:1106 [inlined] │ [4] remotecall(::Function, ::Int64, ::Type, ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:447 │ [5] remotecall │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:447 [inlined] │ [6] init │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:299 [inlined] │ [7] #20 │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:340 [inlined] │ [8] (::Schedulers.var"#loop##12#loop##13"{Schedulers.ElasticLoop{typeof(addprocs), var"#20#21"{var"#init#init##0"{Dict{Int64, Future}}, Dict{Int64, Future}}, Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, Set{Int64}, Set{Int64}, Int64})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:668 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 ┌ Warning: problem initializing 58, removing 58 from cluster. └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:675 ┌ Warning: no process with id 58 exists │ error type: ErrorException │ 1-element ExceptionStack: │ no process with id 58 exists │ Stacktrace: │ [1] error(s::String) │ @ Base ./error.jl:44 │ [2] worker_from_id(pg::Distributed.ProcessGroup, i::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:1114 │ [3] worker_from_id │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:1106 [inlined] │ [4] remotecall(::Function, ::Int64, ::Type, ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:447 │ [5] remotecall │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:447 [inlined] │ [6] init │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:299 [inlined] │ [7] #20 │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:340 [inlined] │ [8] (::Schedulers.var"#loop##12#loop##13"{Schedulers.ElasticLoop{typeof(addprocs), var"#20#21"{var"#init#init##0"{Dict{Int64, Future}}, Dict{Int64, Future}}, Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, Set{Int64}, Set{Int64}, Int64})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:668 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 ┌ Warning: problem initializing 54, removing 54 from cluster. └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:675 ┌ Warning: no process with id 54 exists │ error type: ErrorException │ 1-element ExceptionStack: │ no process with id 54 exists │ Stacktrace: │ [1] error(s::String) │ @ Base ./error.jl:44 │ [2] worker_from_id(pg::Distributed.ProcessGroup, i::Int64) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:1114 │ [3] worker_from_id │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/cluster.jl:1106 [inlined] │ [4] remotecall(::Function, ::Int64, ::Type, ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:447 │ [5] remotecall │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:447 [inlined] │ [6] init │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:299 [inlined] │ [7] #20 │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:340 [inlined] │ [8] (::Schedulers.var"#loop##12#loop##13"{Schedulers.ElasticLoop{typeof(addprocs), var"#20#21"{var"#init#init##0"{Dict{Int64, Future}}, Dict{Int64, Future}}, Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, Set{Int64}, Set{Int64}, Int64})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:668 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 [ Info: running task 1 on process 60 (Schedulers-primary-qO2tilUH); 2 workers total; 3 tasks left in task-pool. [ Info: running task 2 on process 61 (Schedulers-primary-qO2tilUH); 2 workers total; 2 tasks left in task-pool. From worker 60: [ Info: restart task... From worker 60: ┌ Info: ...restart task, From worker 60: │ _s = From worker 60: │ 1-element Vector{Int64}: From worker 60: └ 1 From worker 60: [ Info: _s on pid=60 is 1 From worker 61: [ Info: restart task... From worker 61: ┌ Info: ...restart task, From worker 61: │ _s = From worker 61: │ 1-element Vector{Int64}: From worker 61: └ 1 From worker 61: [ Info: _s on pid=61 is 1 From worker 60: [ Info: _s on pid=60 is 2 From worker 61: [ Info: _s on pid=61 is 2 From worker 60: [ Info: checkpoint task... From worker 60: [ Info: ...checkpoint task with state=2. ┌ Warning: caught an exception, there have been 0 failure(s) on process 60 (Schedulers-primary-qO2tilUH)... └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1111 ┌ Warning: TaskFailedException │ │ nested task error: On worker 60: │ Schedulers.PreemptException() │ Stacktrace: │ [1] #default_threadpool_checkpoint_call#25 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:278 │ [2] default_threadpool_checkpoint_call │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:250 │ [3] #handle_msg##8 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:303 │ [4] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:70 │ [5] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:79 │ [6] #schedule_call##2 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:88 │ Stacktrace: │ [1] remotecall_wait(::Function, ::Distributed.Worker, ::Future, ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:507 │ [2] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:497 [inlined] │ [3] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:519 [inlined] │ [4] (::Schedulers.var"#43#44"{@Kwargs{}, Future, var"#22#23"{var"#checkpoint_task#checkpoint_task##0", Dict{Int64, Future}}, var"#24#25"{var"#restart_task!#restart_task!##0", Dict{Int64, Future}}, Int64, var"#30#31"{var"#foo5d#foo5d##0", String, Dict{Int64, Future}}, Int64, Tuple{Int64}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:339 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#582 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] remotecall_func_wait_timeout(tsk_times::Vector{Float64}, eloop::Schedulers.ElasticLoop{typeof(addprocs), var"#20#21"{var"#init#init##0"{Dict{Int64, Future}}, Dict{Int64, Future}}, Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, options::SchedulerOptions{String}, preempt_channel_future::Future, checkpoint_task::Function, restart_task::Function, tsk::Int64, f::var"#30#31"{var"#foo5d#foo5d##0", String, Dict{Int64, Future}}, pid::Int64, args::Int64; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:349 │ [5] remotecall_func_wait_timeout(tsk_times::Vector{Float64}, eloop::Schedulers.ElasticLoop{typeof(addprocs), var"#20#21"{var"#init#init##0"{Dict{Int64, Future}}, Dict{Int64, Future}}, Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, options::SchedulerOptions{String}, preempt_channel_future::Future, checkpoint_task::Function, restart_task::Function, tsk::Int64, f::Function, pid::Int64, args::Int64) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:338 │ [6] (::Schedulers.var"#107#108"{@Kwargs{}, SchedulerOptions{String}, var"#30#31"{var"#foo5d#foo5d##0", String, Dict{Int64, Future}}, Schedulers.ElasticLoop{typeof(addprocs), var"#20#21"{var"#init#init##0"{Dict{Int64, Future}}, Dict{Int64, Future}}, Schedulers.var"#90#91"{Int64}, Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, Nothing}, Dict{String, Any}, Tuple{}, Vector{Float64}, Int64})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1105 │ │ nested task error: On worker 60: │ Schedulers.PreemptException() │ Stacktrace: │ [1] #default_threadpool_checkpoint_call#25 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:278 │ [2] default_threadpool_checkpoint_call │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:250 │ [3] #handle_msg##8 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:303 │ [4] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:70 │ [5] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:79 │ [6] #schedule_call##2 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:88 │ Stacktrace: │ [1] remotecall_wait(::Function, ::Distributed.Worker, ::Future, ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:507 │ [2] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:497 [inlined] │ [3] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:519 [inlined] │ [4] (::Schedulers.var"#43#44"{@Kwargs{}, Future, var"#22#23"{var"#checkpoint_task#checkpoint_task##0", Dict{Int64, Future}}, var"#24#25"{var"#restart_task!#restart_task!##0", Dict{Int64, Future}}, Int64, var"#30#31"{var"#foo5d#foo5d##0", String, Dict{Int64, Future}}, Int64, Tuple{Int64}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:339 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 ┌ Warning: too many failures on process with id=60 (Schedulers-primary-qO2tilUH), removing from process list └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:393 From worker 61: [ Info: _s on pid=61 is 3 From worker 61: [ Info: _s on pid=61 is 4 From worker 61: [ Info: _s on pid=61 is 5 From worker 61: [ Info: _s on pid=61 is 6 [ Info: running task 3 on process 62 (Schedulers-primary-qO2tilUH); 2 workers total; 2 tasks left in task-pool. From worker 62: [ Info: restart task... From worker 62: ┌ Info: ...restart task, From worker 62: │ _s = From worker 62: │ 1-element Vector{Int64}: From worker 62: └ 1 From worker 62: [ Info: _s on pid=62 is 1 From worker 61: [ Info: _s on pid=61 is 7 From worker 62: [ Info: _s on pid=62 is 2 From worker 61: [ Info: _s on pid=61 is 8 From worker 62: [ Info: _s on pid=62 is 3 From worker 61: [ Info: _s on pid=61 is 9 From worker 62: [ Info: _s on pid=62 is 4 From worker 61: [ Info: _s on pid=61 is 10 From worker 62: [ Info: _s on pid=62 is 5 [ Info: running task 4 on process 61 (Schedulers-primary-qO2tilUH); 2 workers total; 1 tasks left in task-pool. From worker 61: [ Info: restart task... From worker 61: ┌ Info: ...restart task, From worker 61: │ _s = From worker 61: │ 1-element Vector{Int64}: From worker 61: └ 1 From worker 61: [ Info: _s on pid=61 is 1 From worker 62: [ Info: _s on pid=62 is 6 From worker 61: [ Info: _s on pid=61 is 2 From worker 62: [ Info: _s on pid=62 is 7 From worker 61: [ Info: _s on pid=61 is 3 From worker 62: [ Info: _s on pid=62 is 8 From worker 61: [ Info: _s on pid=61 is 4 From worker 62: [ Info: _s on pid=62 is 9 From worker 61: [ Info: _s on pid=61 is 5 From worker 62: [ Info: _s on pid=62 is 10 From worker 61: [ Info: _s on pid=61 is 6 [ Info: running task 1 on process 62 (Schedulers-primary-qO2tilUH); 2 workers total; 0 tasks left in task-pool. From worker 62: [ Info: restart task... From worker 62: ┌ Info: ...restart task, From worker 62: │ _s = From worker 62: │ 1-element Vector{Int64}: From worker 62: └ 2 From worker 62: [ Info: _s on pid=62 is 2 From worker 61: [ Info: _s on pid=61 is 7 From worker 62: [ Info: _s on pid=62 is 3 From worker 61: [ Info: _s on pid=61 is 8 From worker 62: [ Info: _s on pid=62 is 4 From worker 61: [ Info: _s on pid=61 is 9 From worker 62: [ Info: _s on pid=62 is 5 From worker 61: [ Info: _s on pid=61 is 10 From worker 62: [ Info: _s on pid=62 is 6 From worker 62: [ Info: _s on pid=62 is 7 From worker 62: [ Info: _s on pid=62 is 8 From worker 62: [ Info: _s on pid=62 is 9 From worker 62: [ Info: _s on pid=62 is 10 Test Summary: | Pass Total Time pmap with task checkpoint and restart | 1 1 2m27.6s [ Info: running task 1 on process 63 (Schedulers-primary-qO2tilUH); 5 workers total; 99 tasks left in task-pool. [ Info: running task 2 on process 64 (Schedulers-primary-qO2tilUH); 5 workers total; 98 tasks left in task-pool. [ Info: running task 3 on process 65 (Schedulers-primary-qO2tilUH); 5 workers total; 97 tasks left in task-pool. [ Info: running task 4 on process 66 (Schedulers-primary-qO2tilUH); 5 workers total; 95 tasks left in task-pool. [ Info: running task 5 on process 67 (Schedulers-primary-qO2tilUH); 5 workers total; 95 tasks left in task-pool. [ Info: running task 6 on process 64 (Schedulers-primary-qO2tilUH); 5 workers total; 94 tasks left in task-pool. [ Info: running task 7 on process 63 (Schedulers-primary-qO2tilUH); 5 workers total; 93 tasks left in task-pool. [ Info: running task 8 on process 65 (Schedulers-primary-qO2tilUH); 5 workers total; 92 tasks left in task-pool. [ Info: running task 9 on process 66 (Schedulers-primary-qO2tilUH); 5 workers total; 90 tasks left in task-pool. [ Info: running task 10 on process 67 (Schedulers-primary-qO2tilUH); 5 workers total; 90 tasks left in task-pool. [ Info: running task 11 on process 64 (Schedulers-primary-qO2tilUH); 5 workers total; 89 tasks left in task-pool. [ Info: running task 12 on process 63 (Schedulers-primary-qO2tilUH); 5 workers total; 88 tasks left in task-pool. [ Info: running task 13 on process 65 (Schedulers-primary-qO2tilUH); 5 workers total; 87 tasks left in task-pool. [ Info: running task 14 on process 66 (Schedulers-primary-qO2tilUH); 5 workers total; 86 tasks left in task-pool. [ Info: running task 15 on process 67 (Schedulers-primary-qO2tilUH); 5 workers total; 85 tasks left in task-pool. [ Info: running task 16 on process 64 (Schedulers-primary-qO2tilUH); 5 workers total; 84 tasks left in task-pool. [ Info: running task 17 on process 63 (Schedulers-primary-qO2tilUH); 5 workers total; 83 tasks left in task-pool. [ Info: running task 18 on process 65 (Schedulers-primary-qO2tilUH); 5 workers total; 82 tasks left in task-pool. [ Info: running task 19 on process 66 (Schedulers-primary-qO2tilUH); 5 workers total; 81 tasks left in task-pool. [ Info: running task 20 on process 67 (Schedulers-primary-qO2tilUH); 5 workers total; 80 tasks left in task-pool. [ Info: running task 21 on process 64 (Schedulers-primary-qO2tilUH); 5 workers total; 79 tasks left in task-pool. [ Info: running task 22 on process 63 (Schedulers-primary-qO2tilUH); 5 workers total; 78 tasks left in task-pool. [ Info: running task 23 on process 65 (Schedulers-primary-qO2tilUH); 5 workers total; 77 tasks left in task-pool. [ Info: running task 24 on process 66 (Schedulers-primary-qO2tilUH); 5 workers total; 76 tasks left in task-pool. [ Info: running task 25 on process 67 (Schedulers-primary-qO2tilUH); 5 workers total; 75 tasks left in task-pool. [ Info: running task 26 on process 64 (Schedulers-primary-qO2tilUH); 5 workers total; 74 tasks left in task-pool. [ Info: running task 27 on process 63 (Schedulers-primary-qO2tilUH); 5 workers total; 73 tasks left in task-pool. [ Info: running task 28 on process 65 (Schedulers-primary-qO2tilUH); 5 workers total; 72 tasks left in task-pool. [ Info: running task 29 on process 66 (Schedulers-primary-qO2tilUH); 5 workers total; 71 tasks left in task-pool. [ Info: running task 30 on process 67 (Schedulers-primary-qO2tilUH); 5 workers total; 70 tasks left in task-pool. [ Info: running task 31 on process 64 (Schedulers-primary-qO2tilUH); 5 workers total; 69 tasks left in task-pool. [ Info: running task 32 on process 63 (Schedulers-primary-qO2tilUH); 5 workers total; 68 tasks left in task-pool. [ Info: running task 33 on process 65 (Schedulers-primary-qO2tilUH); 5 workers total; 67 tasks left in task-pool. [ Info: running task 34 on process 66 (Schedulers-primary-qO2tilUH); 5 workers total; 66 tasks left in task-pool. [ Info: running task 35 on process 67 (Schedulers-primary-qO2tilUH); 5 workers total; 65 tasks left in task-pool. [ Info: running task 36 on process 64 (Schedulers-primary-qO2tilUH); 5 workers total; 64 tasks left in task-pool. [ Info: running task 37 on process 63 (Schedulers-primary-qO2tilUH); 5 workers total; 63 tasks left in task-pool. [ Info: running task 38 on process 65 (Schedulers-primary-qO2tilUH); 5 workers total; 62 tasks left in task-pool. [ Info: running task 39 on process 66 (Schedulers-primary-qO2tilUH); 5 workers total; 61 tasks left in task-pool. [ Info: running task 40 on process 67 (Schedulers-primary-qO2tilUH); 5 workers total; 60 tasks left in task-pool. [ Info: running task 41 on process 64 (Schedulers-primary-qO2tilUH); 5 workers total; 59 tasks left in task-pool. [ Info: running task 42 on process 63 (Schedulers-primary-qO2tilUH); 5 workers total; 58 tasks left in task-pool. [ Info: running task 43 on process 65 (Schedulers-primary-qO2tilUH); 5 workers total; 57 tasks left in task-pool. [ Info: running task 44 on process 66 (Schedulers-primary-qO2tilUH); 5 workers total; 56 tasks left in task-pool. [ Info: running task 45 on process 67 (Schedulers-primary-qO2tilUH); 5 workers total; 55 tasks left in task-pool. [ Info: running task 46 on process 64 (Schedulers-primary-qO2tilUH); 5 workers total; 54 tasks left in task-pool. [ Info: running task 47 on process 63 (Schedulers-primary-qO2tilUH); 5 workers total; 53 tasks left in task-pool. [ Info: running task 48 on process 65 (Schedulers-primary-qO2tilUH); 5 workers total; 52 tasks left in task-pool. [ Info: running task 49 on process 66 (Schedulers-primary-qO2tilUH); 5 workers total; 50 tasks left in task-pool. [ Info: running task 50 on process 67 (Schedulers-primary-qO2tilUH); 5 workers total; 50 tasks left in task-pool. [ Info: running task 51 on process 64 (Schedulers-primary-qO2tilUH); 5 workers total; 49 tasks left in task-pool. [ Info: running task 52 on process 63 (Schedulers-primary-qO2tilUH); 5 workers total; 48 tasks left in task-pool. [ Info: running task 53 on process 65 (Schedulers-primary-qO2tilUH); 5 workers total; 47 tasks left in task-pool. [ Info: running task 54 on process 66 (Schedulers-primary-qO2tilUH); 5 workers total; 46 tasks left in task-pool. [ Info: running task 55 on process 67 (Schedulers-primary-qO2tilUH); 5 workers total; 45 tasks left in task-pool. [ Info: running task 56 on process 64 (Schedulers-primary-qO2tilUH); 5 workers total; 44 tasks left in task-pool. [ Info: running task 57 on process 63 (Schedulers-primary-qO2tilUH); 5 workers total; 43 tasks left in task-pool. [ Info: running task 58 on process 65 (Schedulers-primary-qO2tilUH); 5 workers total; 42 tasks left in task-pool. [ Info: running task 59 on process 66 (Schedulers-primary-qO2tilUH); 5 workers total; 41 tasks left in task-pool. [ Info: running task 60 on process 67 (Schedulers-primary-qO2tilUH); 5 workers total; 40 tasks left in task-pool. [ Info: running task 61 on process 64 (Schedulers-primary-qO2tilUH); 5 workers total; 39 tasks left in task-pool. [ Info: running task 62 on process 63 (Schedulers-primary-qO2tilUH); 5 workers total; 38 tasks left in task-pool. [ Info: running task 63 on process 65 (Schedulers-primary-qO2tilUH); 5 workers total; 37 tasks left in task-pool. [ Info: running task 64 on process 66 (Schedulers-primary-qO2tilUH); 5 workers total; 36 tasks left in task-pool. [ Info: running task 65 on process 67 (Schedulers-primary-qO2tilUH); 5 workers total; 35 tasks left in task-pool. [ Info: running task 66 on process 64 (Schedulers-primary-qO2tilUH); 5 workers total; 34 tasks left in task-pool. [ Info: running task 67 on process 63 (Schedulers-primary-qO2tilUH); 5 workers total; 33 tasks left in task-pool. [ Info: running task 68 on process 65 (Schedulers-primary-qO2tilUH); 5 workers total; 32 tasks left in task-pool. [ Info: running task 69 on process 66 (Schedulers-primary-qO2tilUH); 5 workers total; 30 tasks left in task-pool. [ Info: running task 70 on process 67 (Schedulers-primary-qO2tilUH); 5 workers total; 30 tasks left in task-pool. [ Info: running task 71 on process 64 (Schedulers-primary-qO2tilUH); 5 workers total; 29 tasks left in task-pool. [ Info: running task 72 on process 63 (Schedulers-primary-qO2tilUH); 5 workers total; 28 tasks left in task-pool. [ Info: running task 73 on process 65 (Schedulers-primary-qO2tilUH); 5 workers total; 27 tasks left in task-pool. [ Info: running task 74 on process 66 (Schedulers-primary-qO2tilUH); 5 workers total; 25 tasks left in task-pool. [ Info: running task 75 on process 67 (Schedulers-primary-qO2tilUH); 5 workers total; 25 tasks left in task-pool. [ Info: running task 76 on process 64 (Schedulers-primary-qO2tilUH); 5 workers total; 24 tasks left in task-pool. [ Info: running task 77 on process 63 (Schedulers-primary-qO2tilUH); 5 workers total; 23 tasks left in task-pool. [ Info: running task 78 on process 65 (Schedulers-primary-qO2tilUH); 5 workers total; 22 tasks left in task-pool. [ Info: running task 79 on process 66 (Schedulers-primary-qO2tilUH); 5 workers total; 20 tasks left in task-pool. [ Info: running task 80 on process 67 (Schedulers-primary-qO2tilUH); 5 workers total; 20 tasks left in task-pool. [ Info: running task 81 on process 64 (Schedulers-primary-qO2tilUH); 5 workers total; 19 tasks left in task-pool. [ Info: running task 82 on process 63 (Schedulers-primary-qO2tilUH); 5 workers total; 18 tasks left in task-pool. [ Info: running task 83 on process 65 (Schedulers-primary-qO2tilUH); 5 workers total; 17 tasks left in task-pool. [ Info: running task 84 on process 66 (Schedulers-primary-qO2tilUH); 5 workers total; 16 tasks left in task-pool. [ Info: running task 85 on process 67 (Schedulers-primary-qO2tilUH); 5 workers total; 15 tasks left in task-pool. [ Info: running task 86 on process 64 (Schedulers-primary-qO2tilUH); 5 workers total; 14 tasks left in task-pool. [ Info: running task 87 on process 63 (Schedulers-primary-qO2tilUH); 5 workers total; 13 tasks left in task-pool. [ Info: running task 88 on process 65 (Schedulers-primary-qO2tilUH); 5 workers total; 12 tasks left in task-pool. [ Info: running task 89 on process 66 (Schedulers-primary-qO2tilUH); 5 workers total; 10 tasks left in task-pool. [ Info: running task 90 on process 67 (Schedulers-primary-qO2tilUH); 5 workers total; 10 tasks left in task-pool. [ Info: running task 91 on process 64 (Schedulers-primary-qO2tilUH); 5 workers total; 9 tasks left in task-pool. [ Info: running task 92 on process 63 (Schedulers-primary-qO2tilUH); 5 workers total; 8 tasks left in task-pool. [ Info: running task 93 on process 65 (Schedulers-primary-qO2tilUH); 5 workers total; 7 tasks left in task-pool. [ Info: running task 94 on process 66 (Schedulers-primary-qO2tilUH); 5 workers total; 5 tasks left in task-pool. [ Info: running task 95 on process 67 (Schedulers-primary-qO2tilUH); 5 workers total; 5 tasks left in task-pool. [ Info: running task 96 on process 64 (Schedulers-primary-qO2tilUH); 5 workers total; 4 tasks left in task-pool. [ Info: running task 97 on process 63 (Schedulers-primary-qO2tilUH); 5 workers total; 3 tasks left in task-pool. [ Info: running task 98 on process 65 (Schedulers-primary-qO2tilUH); 5 workers total; 2 tasks left in task-pool. [ Info: running task 99 on process 66 (Schedulers-primary-qO2tilUH); 5 workers total; 0 tasks left in task-pool. [ Info: running task 100 on process 67 (Schedulers-primary-qO2tilUH); 5 workers total; 0 tasks left in task-pool. [ Info: reducing from 5 checkpoints using process 63 (5 workers, 1 reduce workers). [ Info: reducing from 4 checkpoints using process 64 (5 workers, 2 reduce workers). [ Info: reducing from 2 checkpoints using process 63 (5 workers, 2 reduce workers). [ Info: reducing from 2 checkpoints using process 63 (5 workers, 2 reduce workers). Test Summary: | Pass Total Time pmapreduce, stable cluster test, backwards compatability | 36 36 3m18.3s [ Info: running task 1 on process 68 (Schedulers-primary-qO2tilUH); 5 workers total; 99 tasks left in task-pool. [ Info: running task 2 on process 69 (Schedulers-primary-qO2tilUH); 5 workers total; 98 tasks left in task-pool. [ Info: running task 3 on process 70 (Schedulers-primary-qO2tilUH); 5 workers total; 97 tasks left in task-pool. [ Info: running task 4 on process 71 (Schedulers-primary-qO2tilUH); 5 workers total; 95 tasks left in task-pool. [ Info: running task 5 on process 72 (Schedulers-primary-qO2tilUH); 5 workers total; 95 tasks left in task-pool. [ Info: running task 6 on process 68 (Schedulers-primary-qO2tilUH); 5 workers total; 94 tasks left in task-pool. [ Info: running task 7 on process 69 (Schedulers-primary-qO2tilUH); 5 workers total; 93 tasks left in task-pool. [ Info: running task 8 on process 70 (Schedulers-primary-qO2tilUH); 5 workers total; 92 tasks left in task-pool. [ Info: running task 9 on process 71 (Schedulers-primary-qO2tilUH); 5 workers total; 90 tasks left in task-pool. [ Info: running task 10 on process 72 (Schedulers-primary-qO2tilUH); 5 workers total; 90 tasks left in task-pool. [ Info: running task 11 on process 68 (Schedulers-primary-qO2tilUH); 5 workers total; 89 tasks left in task-pool. [ Info: running task 12 on process 69 (Schedulers-primary-qO2tilUH); 5 workers total; 88 tasks left in task-pool. [ Info: running task 13 on process 70 (Schedulers-primary-qO2tilUH); 5 workers total; 87 tasks left in task-pool. [ Info: running task 14 on process 71 (Schedulers-primary-qO2tilUH); 5 workers total; 86 tasks left in task-pool. [ Info: running task 15 on process 72 (Schedulers-primary-qO2tilUH); 5 workers total; 85 tasks left in task-pool. [ Info: running task 16 on process 68 (Schedulers-primary-qO2tilUH); 5 workers total; 84 tasks left in task-pool. [ Info: running task 17 on process 69 (Schedulers-primary-qO2tilUH); 5 workers total; 83 tasks left in task-pool. [ Info: running task 18 on process 70 (Schedulers-primary-qO2tilUH); 5 workers total; 82 tasks left in task-pool. [ Info: running task 19 on process 71 (Schedulers-primary-qO2tilUH); 5 workers total; 81 tasks left in task-pool. [ Info: running task 20 on process 72 (Schedulers-primary-qO2tilUH); 5 workers total; 80 tasks left in task-pool. [ Info: running task 21 on process 68 (Schedulers-primary-qO2tilUH); 5 workers total; 79 tasks left in task-pool. [ Info: running task 22 on process 69 (Schedulers-primary-qO2tilUH); 5 workers total; 78 tasks left in task-pool. [ Info: running task 23 on process 70 (Schedulers-primary-qO2tilUH); 5 workers total; 77 tasks left in task-pool. [ Info: running task 24 on process 71 (Schedulers-primary-qO2tilUH); 5 workers total; 76 tasks left in task-pool. [ Info: running task 25 on process 72 (Schedulers-primary-qO2tilUH); 5 workers total; 75 tasks left in task-pool. [ Info: running task 26 on process 68 (Schedulers-primary-qO2tilUH); 5 workers total; 74 tasks left in task-pool. [ Info: running task 27 on process 69 (Schedulers-primary-qO2tilUH); 5 workers total; 73 tasks left in task-pool. [ Info: running task 28 on process 70 (Schedulers-primary-qO2tilUH); 5 workers total; 72 tasks left in task-pool. [ Info: running task 29 on process 71 (Schedulers-primary-qO2tilUH); 5 workers total; 71 tasks left in task-pool. [ Info: running task 30 on process 72 (Schedulers-primary-qO2tilUH); 5 workers total; 70 tasks left in task-pool. [ Info: running task 31 on process 68 (Schedulers-primary-qO2tilUH); 5 workers total; 69 tasks left in task-pool. [ Info: running task 32 on process 69 (Schedulers-primary-qO2tilUH); 5 workers total; 68 tasks left in task-pool. [ Info: running task 33 on process 70 (Schedulers-primary-qO2tilUH); 5 workers total; 67 tasks left in task-pool. [ Info: running task 34 on process 71 (Schedulers-primary-qO2tilUH); 5 workers total; 65 tasks left in task-pool. [ Info: running task 35 on process 72 (Schedulers-primary-qO2tilUH); 5 workers total; 65 tasks left in task-pool. [ Info: running task 36 on process 68 (Schedulers-primary-qO2tilUH); 5 workers total; 64 tasks left in task-pool. [ Info: running task 37 on process 69 (Schedulers-primary-qO2tilUH); 5 workers total; 63 tasks left in task-pool. [ Info: running task 38 on process 70 (Schedulers-primary-qO2tilUH); 5 workers total; 62 tasks left in task-pool. [ Info: running task 39 on process 71 (Schedulers-primary-qO2tilUH); 5 workers total; 60 tasks left in task-pool. [ Info: running task 40 on process 72 (Schedulers-primary-qO2tilUH); 5 workers total; 60 tasks left in task-pool. [ Info: running task 41 on process 68 (Schedulers-primary-qO2tilUH); 5 workers total; 59 tasks left in task-pool. [ Info: running task 42 on process 69 (Schedulers-primary-qO2tilUH); 5 workers total; 58 tasks left in task-pool. [ Info: running task 43 on process 70 (Schedulers-primary-qO2tilUH); 5 workers total; 57 tasks left in task-pool. [ Info: running task 44 on process 71 (Schedulers-primary-qO2tilUH); 5 workers total; 55 tasks left in task-pool. [ Info: running task 45 on process 72 (Schedulers-primary-qO2tilUH); 5 workers total; 55 tasks left in task-pool. [ Info: running task 46 on process 68 (Schedulers-primary-qO2tilUH); 5 workers total; 54 tasks left in task-pool. [ Info: running task 47 on process 69 (Schedulers-primary-qO2tilUH); 5 workers total; 53 tasks left in task-pool. [ Info: running task 48 on process 70 (Schedulers-primary-qO2tilUH); 5 workers total; 52 tasks left in task-pool. [ Info: running task 49 on process 71 (Schedulers-primary-qO2tilUH); 5 workers total; 50 tasks left in task-pool. [ Info: running task 50 on process 72 (Schedulers-primary-qO2tilUH); 5 workers total; 50 tasks left in task-pool. [ Info: running task 51 on process 68 (Schedulers-primary-qO2tilUH); 5 workers total; 49 tasks left in task-pool. [ Info: running task 52 on process 69 (Schedulers-primary-qO2tilUH); 5 workers total; 48 tasks left in task-pool. [ Info: running task 53 on process 70 (Schedulers-primary-qO2tilUH); 5 workers total; 47 tasks left in task-pool. [ Info: running task 54 on process 71 (Schedulers-primary-qO2tilUH); 5 workers total; 45 tasks left in task-pool. [ Info: running task 55 on process 72 (Schedulers-primary-qO2tilUH); 5 workers total; 45 tasks left in task-pool. [ Info: running task 56 on process 68 (Schedulers-primary-qO2tilUH); 5 workers total; 44 tasks left in task-pool. [ Info: running task 57 on process 69 (Schedulers-primary-qO2tilUH); 5 workers total; 43 tasks left in task-pool. [ Info: running task 58 on process 70 (Schedulers-primary-qO2tilUH); 5 workers total; 42 tasks left in task-pool. [ Info: running task 59 on process 71 (Schedulers-primary-qO2tilUH); 5 workers total; 40 tasks left in task-pool. [ Info: running task 60 on process 72 (Schedulers-primary-qO2tilUH); 5 workers total; 40 tasks left in task-pool. [ Info: running task 61 on process 68 (Schedulers-primary-qO2tilUH); 5 workers total; 39 tasks left in task-pool. [ Info: running task 62 on process 69 (Schedulers-primary-qO2tilUH); 5 workers total; 38 tasks left in task-pool. [ Info: running task 63 on process 70 (Schedulers-primary-qO2tilUH); 5 workers total; 37 tasks left in task-pool. [ Info: running task 64 on process 71 (Schedulers-primary-qO2tilUH); 5 workers total; 35 tasks left in task-pool. [ Info: running task 65 on process 72 (Schedulers-primary-qO2tilUH); 5 workers total; 35 tasks left in task-pool. [ Info: running task 66 on process 68 (Schedulers-primary-qO2tilUH); 5 workers total; 34 tasks left in task-pool. [ Info: running task 67 on process 69 (Schedulers-primary-qO2tilUH); 5 workers total; 33 tasks left in task-pool. [ Info: running task 68 on process 70 (Schedulers-primary-qO2tilUH); 5 workers total; 32 tasks left in task-pool. [ Info: running task 69 on process 71 (Schedulers-primary-qO2tilUH); 5 workers total; 30 tasks left in task-pool. [ Info: running task 70 on process 72 (Schedulers-primary-qO2tilUH); 5 workers total; 30 tasks left in task-pool. [ Info: running task 71 on process 68 (Schedulers-primary-qO2tilUH); 5 workers total; 29 tasks left in task-pool. [ Info: running task 72 on process 69 (Schedulers-primary-qO2tilUH); 5 workers total; 28 tasks left in task-pool. [ Info: running task 73 on process 70 (Schedulers-primary-qO2tilUH); 5 workers total; 27 tasks left in task-pool. [ Info: running task 74 on process 71 (Schedulers-primary-qO2tilUH); 5 workers total; 25 tasks left in task-pool. [ Info: running task 75 on process 72 (Schedulers-primary-qO2tilUH); 5 workers total; 25 tasks left in task-pool. [ Info: running task 76 on process 68 (Schedulers-primary-qO2tilUH); 5 workers total; 24 tasks left in task-pool. [ Info: running task 77 on process 69 (Schedulers-primary-qO2tilUH); 5 workers total; 23 tasks left in task-pool. [ Info: running task 78 on process 70 (Schedulers-primary-qO2tilUH); 5 workers total; 22 tasks left in task-pool. [ Info: running task 79 on process 71 (Schedulers-primary-qO2tilUH); 5 workers total; 20 tasks left in task-pool. [ Info: running task 80 on process 72 (Schedulers-primary-qO2tilUH); 5 workers total; 20 tasks left in task-pool. [ Info: running task 81 on process 68 (Schedulers-primary-qO2tilUH); 5 workers total; 19 tasks left in task-pool. [ Info: running task 82 on process 69 (Schedulers-primary-qO2tilUH); 5 workers total; 18 tasks left in task-pool. [ Info: running task 83 on process 70 (Schedulers-primary-qO2tilUH); 5 workers total; 17 tasks left in task-pool. [ Info: running task 84 on process 71 (Schedulers-primary-qO2tilUH); 5 workers total; 15 tasks left in task-pool. [ Info: running task 85 on process 72 (Schedulers-primary-qO2tilUH); 5 workers total; 15 tasks left in task-pool. [ Info: running task 86 on process 68 (Schedulers-primary-qO2tilUH); 5 workers total; 14 tasks left in task-pool. [ Info: running task 87 on process 69 (Schedulers-primary-qO2tilUH); 5 workers total; 13 tasks left in task-pool. [ Info: running task 88 on process 70 (Schedulers-primary-qO2tilUH); 5 workers total; 12 tasks left in task-pool. [ Info: running task 89 on process 71 (Schedulers-primary-qO2tilUH); 5 workers total; 10 tasks left in task-pool. [ Info: running task 90 on process 72 (Schedulers-primary-qO2tilUH); 5 workers total; 10 tasks left in task-pool. [ Info: running task 91 on process 68 (Schedulers-primary-qO2tilUH); 5 workers total; 9 tasks left in task-pool. [ Info: running task 92 on process 69 (Schedulers-primary-qO2tilUH); 5 workers total; 8 tasks left in task-pool. [ Info: running task 93 on process 70 (Schedulers-primary-qO2tilUH); 5 workers total; 7 tasks left in task-pool. [ Info: running task 94 on process 71 (Schedulers-primary-qO2tilUH); 5 workers total; 5 tasks left in task-pool. [ Info: running task 95 on process 72 (Schedulers-primary-qO2tilUH); 5 workers total; 5 tasks left in task-pool. [ Info: running task 96 on process 68 (Schedulers-primary-qO2tilUH); 5 workers total; 4 tasks left in task-pool. [ Info: running task 97 on process 69 (Schedulers-primary-qO2tilUH); 5 workers total; 3 tasks left in task-pool. [ Info: running task 98 on process 70 (Schedulers-primary-qO2tilUH); 5 workers total; 2 tasks left in task-pool. [ Info: running task 99 on process 71 (Schedulers-primary-qO2tilUH); 5 workers total; 0 tasks left in task-pool. [ Info: running task 100 on process 72 (Schedulers-primary-qO2tilUH); 5 workers total; 0 tasks left in task-pool. [ Info: reducing from 5 checkpoints using process 68 (5 workers, 2 reduce workers). [ Info: reducing from 3 checkpoints using process 69 (5 workers, 2 reduce workers). [ Info: reducing from 3 checkpoints using process 68 (5 workers, 2 reduce workers). [ Info: reducing from 2 checkpoints using process 68 (5 workers, 1 reduce workers). Test Summary: | Pass Total Time pmapreduce, stable cluster test | 2 2 3m00.6s [ Info: running task 1 on process 73 (Schedulers-primary-qO2tilUH); 5 workers total; 99 tasks left in task-pool. [ Info: running task 2 on process 75 (Schedulers-primary-qO2tilUH); 5 workers total; 98 tasks left in task-pool. [ Info: running task 3 on process 76 (Schedulers-primary-qO2tilUH); 5 workers total; 97 tasks left in task-pool. [ Info: running task 4 on process 77 (Schedulers-primary-qO2tilUH); 5 workers total; 96 tasks left in task-pool. [ Info: running task 5 on process 75 (Schedulers-primary-qO2tilUH); 5 workers total; 95 tasks left in task-pool. [ Info: running task 6 on process 73 (Schedulers-primary-qO2tilUH); 5 workers total; 94 tasks left in task-pool. [ Info: running task 7 on process 76 (Schedulers-primary-qO2tilUH); 5 workers total; 93 tasks left in task-pool. [ Info: running task 8 on process 77 (Schedulers-primary-qO2tilUH); 5 workers total; 92 tasks left in task-pool. [ Info: running task 9 on process 75 (Schedulers-primary-qO2tilUH); 5 workers total; 91 tasks left in task-pool. [ Info: running task 10 on process 73 (Schedulers-primary-qO2tilUH); 5 workers total; 90 tasks left in task-pool. [ Info: running task 11 on process 76 (Schedulers-primary-qO2tilUH); 5 workers total; 89 tasks left in task-pool. [ Info: running task 12 on process 77 (Schedulers-primary-qO2tilUH); 5 workers total; 88 tasks left in task-pool. [ Info: running task 13 on process 75 (Schedulers-primary-qO2tilUH); 5 workers total; 87 tasks left in task-pool. [ Info: running task 14 on process 73 (Schedulers-primary-qO2tilUH); 5 workers total; 86 tasks left in task-pool. [ Info: running task 15 on process 76 (Schedulers-primary-qO2tilUH); 5 workers total; 85 tasks left in task-pool. [ Info: running task 16 on process 77 (Schedulers-primary-qO2tilUH); 5 workers total; 84 tasks left in task-pool. [ Info: running task 17 on process 75 (Schedulers-primary-qO2tilUH); 5 workers total; 83 tasks left in task-pool. [ Info: running task 18 on process 78 (Schedulers-primary-qO2tilUH); 5 workers total; 82 tasks left in task-pool. [ Info: running task 19 on process 73 (Schedulers-primary-qO2tilUH); 5 workers total; 81 tasks left in task-pool. [ Info: running task 20 on process 76 (Schedulers-primary-qO2tilUH); 5 workers total; 80 tasks left in task-pool. [ Info: running task 21 on process 77 (Schedulers-primary-qO2tilUH); 5 workers total; 79 tasks left in task-pool. [ Info: running task 22 on process 75 (Schedulers-primary-qO2tilUH); 5 workers total; 78 tasks left in task-pool. [ Info: running task 23 on process 78 (Schedulers-primary-qO2tilUH); 5 workers total; 77 tasks left in task-pool. [ Info: running task 24 on process 73 (Schedulers-primary-qO2tilUH); 5 workers total; 76 tasks left in task-pool. [ Info: running task 25 on process 76 (Schedulers-primary-qO2tilUH); 5 workers total; 75 tasks left in task-pool. [ Info: running task 26 on process 77 (Schedulers-primary-qO2tilUH); 5 workers total; 74 tasks left in task-pool. [ Info: running task 27 on process 75 (Schedulers-primary-qO2tilUH); 5 workers total; 73 tasks left in task-pool. [ Info: running task 28 on process 78 (Schedulers-primary-qO2tilUH); 5 workers total; 72 tasks left in task-pool. [ Info: running task 29 on process 73 (Schedulers-primary-qO2tilUH); 5 workers total; 71 tasks left in task-pool. [ Info: running task 30 on process 76 (Schedulers-primary-qO2tilUH); 5 workers total; 70 tasks left in task-pool. [ Info: running task 31 on process 77 (Schedulers-primary-qO2tilUH); 5 workers total; 69 tasks left in task-pool. [ Info: running task 32 on process 75 (Schedulers-primary-qO2tilUH); 5 workers total; 68 tasks left in task-pool. [ Info: running task 33 on process 78 (Schedulers-primary-qO2tilUH); 5 workers total; 67 tasks left in task-pool. [ Info: running task 34 on process 73 (Schedulers-primary-qO2tilUH); 5 workers total; 66 tasks left in task-pool. [ Info: running task 35 on process 76 (Schedulers-primary-qO2tilUH); 5 workers total; 65 tasks left in task-pool. [ Info: running task 36 on process 77 (Schedulers-primary-qO2tilUH); 5 workers total; 64 tasks left in task-pool. [ Info: running task 37 on process 75 (Schedulers-primary-qO2tilUH); 5 workers total; 63 tasks left in task-pool. [ Info: running task 38 on process 78 (Schedulers-primary-qO2tilUH); 5 workers total; 62 tasks left in task-pool. [ Info: running task 39 on process 73 (Schedulers-primary-qO2tilUH); 5 workers total; 61 tasks left in task-pool. [ Info: running task 40 on process 76 (Schedulers-primary-qO2tilUH); 5 workers total; 60 tasks left in task-pool. [ Info: running task 41 on process 77 (Schedulers-primary-qO2tilUH); 5 workers total; 59 tasks left in task-pool. [ Info: running task 42 on process 75 (Schedulers-primary-qO2tilUH); 5 workers total; 58 tasks left in task-pool. [ Info: running task 43 on process 78 (Schedulers-primary-qO2tilUH); 5 workers total; 57 tasks left in task-pool. [ Info: running task 44 on process 73 (Schedulers-primary-qO2tilUH); 5 workers total; 56 tasks left in task-pool. [ Info: running task 45 on process 76 (Schedulers-primary-qO2tilUH); 5 workers total; 55 tasks left in task-pool. [ Info: running task 46 on process 77 (Schedulers-primary-qO2tilUH); 5 workers total; 54 tasks left in task-pool. [ Info: running task 47 on process 75 (Schedulers-primary-qO2tilUH); 5 workers total; 53 tasks left in task-pool. [ Info: running task 48 on process 78 (Schedulers-primary-qO2tilUH); 5 workers total; 52 tasks left in task-pool. [ Info: running task 49 on process 73 (Schedulers-primary-qO2tilUH); 5 workers total; 51 tasks left in task-pool. [ Info: running task 50 on process 76 (Schedulers-primary-qO2tilUH); 5 workers total; 50 tasks left in task-pool. [ Info: running task 51 on process 77 (Schedulers-primary-qO2tilUH); 5 workers total; 49 tasks left in task-pool. [ Info: running task 52 on process 75 (Schedulers-primary-qO2tilUH); 5 workers total; 48 tasks left in task-pool. [ Info: running task 53 on process 78 (Schedulers-primary-qO2tilUH); 5 workers total; 47 tasks left in task-pool. [ Info: running task 54 on process 73 (Schedulers-primary-qO2tilUH); 5 workers total; 46 tasks left in task-pool. [ Info: running task 55 on process 76 (Schedulers-primary-qO2tilUH); 5 workers total; 45 tasks left in task-pool. [ Info: running task 56 on process 77 (Schedulers-primary-qO2tilUH); 5 workers total; 44 tasks left in task-pool. [ Info: running task 57 on process 75 (Schedulers-primary-qO2tilUH); 5 workers total; 43 tasks left in task-pool. [ Info: running task 58 on process 78 (Schedulers-primary-qO2tilUH); 5 workers total; 42 tasks left in task-pool. [ Info: running task 59 on process 73 (Schedulers-primary-qO2tilUH); 5 workers total; 41 tasks left in task-pool. [ Info: running task 60 on process 76 (Schedulers-primary-qO2tilUH); 5 workers total; 40 tasks left in task-pool. [ Info: running task 61 on process 77 (Schedulers-primary-qO2tilUH); 5 workers total; 39 tasks left in task-pool. [ Info: running task 62 on process 75 (Schedulers-primary-qO2tilUH); 5 workers total; 38 tasks left in task-pool. [ Info: running task 63 on process 78 (Schedulers-primary-qO2tilUH); 5 workers total; 37 tasks left in task-pool. [ Info: running task 64 on process 73 (Schedulers-primary-qO2tilUH); 5 workers total; 36 tasks left in task-pool. [ Info: running task 65 on process 76 (Schedulers-primary-qO2tilUH); 5 workers total; 35 tasks left in task-pool. [ Info: running task 66 on process 77 (Schedulers-primary-qO2tilUH); 5 workers total; 34 tasks left in task-pool. [ Info: running task 67 on process 75 (Schedulers-primary-qO2tilUH); 5 workers total; 33 tasks left in task-pool. [ Info: running task 68 on process 78 (Schedulers-primary-qO2tilUH); 5 workers total; 32 tasks left in task-pool. [ Info: running task 69 on process 73 (Schedulers-primary-qO2tilUH); 5 workers total; 31 tasks left in task-pool. [ Info: running task 70 on process 76 (Schedulers-primary-qO2tilUH); 5 workers total; 30 tasks left in task-pool. [ Info: running task 71 on process 77 (Schedulers-primary-qO2tilUH); 5 workers total; 29 tasks left in task-pool. [ Info: running task 72 on process 75 (Schedulers-primary-qO2tilUH); 5 workers total; 28 tasks left in task-pool. [ Info: running task 73 on process 78 (Schedulers-primary-qO2tilUH); 5 workers total; 27 tasks left in task-pool. [ Info: running task 74 on process 73 (Schedulers-primary-qO2tilUH); 5 workers total; 26 tasks left in task-pool. [ Info: running task 75 on process 76 (Schedulers-primary-qO2tilUH); 5 workers total; 25 tasks left in task-pool. [ Info: running task 76 on process 77 (Schedulers-primary-qO2tilUH); 5 workers total; 24 tasks left in task-pool. [ Info: running task 77 on process 75 (Schedulers-primary-qO2tilUH); 5 workers total; 23 tasks left in task-pool. [ Info: running task 78 on process 78 (Schedulers-primary-qO2tilUH); 5 workers total; 22 tasks left in task-pool. [ Info: running task 79 on process 73 (Schedulers-primary-qO2tilUH); 5 workers total; 21 tasks left in task-pool. [ Info: running task 80 on process 76 (Schedulers-primary-qO2tilUH); 5 workers total; 20 tasks left in task-pool. [ Info: running task 81 on process 77 (Schedulers-primary-qO2tilUH); 5 workers total; 19 tasks left in task-pool. [ Info: running task 82 on process 75 (Schedulers-primary-qO2tilUH); 5 workers total; 18 tasks left in task-pool. [ Info: running task 83 on process 78 (Schedulers-primary-qO2tilUH); 5 workers total; 17 tasks left in task-pool. [ Info: running task 84 on process 73 (Schedulers-primary-qO2tilUH); 5 workers total; 16 tasks left in task-pool. [ Info: running task 85 on process 76 (Schedulers-primary-qO2tilUH); 5 workers total; 15 tasks left in task-pool. [ Info: running task 86 on process 77 (Schedulers-primary-qO2tilUH); 5 workers total; 14 tasks left in task-pool. [ Info: running task 87 on process 75 (Schedulers-primary-qO2tilUH); 5 workers total; 13 tasks left in task-pool. [ Info: running task 88 on process 78 (Schedulers-primary-qO2tilUH); 5 workers total; 12 tasks left in task-pool. [ Info: running task 89 on process 73 (Schedulers-primary-qO2tilUH); 5 workers total; 11 tasks left in task-pool. [ Info: running task 90 on process 76 (Schedulers-primary-qO2tilUH); 5 workers total; 10 tasks left in task-pool. [ Info: running task 91 on process 77 (Schedulers-primary-qO2tilUH); 5 workers total; 9 tasks left in task-pool. [ Info: running task 92 on process 75 (Schedulers-primary-qO2tilUH); 5 workers total; 8 tasks left in task-pool. [ Info: running task 93 on process 78 (Schedulers-primary-qO2tilUH); 5 workers total; 7 tasks left in task-pool. [ Info: running task 94 on process 73 (Schedulers-primary-qO2tilUH); 5 workers total; 6 tasks left in task-pool. [ Info: running task 95 on process 76 (Schedulers-primary-qO2tilUH); 5 workers total; 5 tasks left in task-pool. [ Info: running task 96 on process 77 (Schedulers-primary-qO2tilUH); 5 workers total; 4 tasks left in task-pool. [ Info: running task 97 on process 75 (Schedulers-primary-qO2tilUH); 5 workers total; 3 tasks left in task-pool. [ Info: running task 98 on process 78 (Schedulers-primary-qO2tilUH); 5 workers total; 2 tasks left in task-pool. [ Info: running task 99 on process 73 (Schedulers-primary-qO2tilUH); 5 workers total; 1 tasks left in task-pool. [ Info: running task 100 on process 76 (Schedulers-primary-qO2tilUH); 5 workers total; 0 tasks left in task-pool. [ Info: reducing from 5 checkpoints using process 77 (5 workers, 1 reduce workers). [ Info: reducing from 4 checkpoints using process 73 (5 workers, 2 reduce workers). [ Info: reducing from 2 checkpoints using process 77 (5 workers, 2 reduce workers). [ Info: reducing from 2 checkpoints using process 77 (5 workers, 2 reduce workers). Test Summary: | Pass Total Time pmapreduce, cluster with ProcessExitedException during tasks | 4 4 3m32.4s [ Info: running task 1 on process 79 (Schedulers-primary-qO2tilUH); 5 workers total; 99 tasks left in task-pool. [ Info: running task 2 on process 80 (Schedulers-primary-qO2tilUH); 5 workers total; 98 tasks left in task-pool. [ Info: running task 3 on process 81 (Schedulers-primary-qO2tilUH); 5 workers total; 97 tasks left in task-pool. [ Info: running task 4 on process 82 (Schedulers-primary-qO2tilUH); 5 workers total; 95 tasks left in task-pool. [ Info: running task 5 on process 83 (Schedulers-primary-qO2tilUH); 5 workers total; 95 tasks left in task-pool. [ Info: running task 6 on process 79 (Schedulers-primary-qO2tilUH); 5 workers total; 94 tasks left in task-pool. [ Info: running task 7 on process 80 (Schedulers-primary-qO2tilUH); 5 workers total; 93 tasks left in task-pool. [ Info: running task 8 on process 82 (Schedulers-primary-qO2tilUH); 5 workers total; 92 tasks left in task-pool. [ Info: running task 9 on process 79 (Schedulers-primary-qO2tilUH); 5 workers total; 91 tasks left in task-pool. ┌ Warning: pid=81 (Schedulers-primary-qO2tilUH), checkpoint=nothing, task loop, caught exception during save_checkpoint └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1410 ┌ Warning: TaskFailedException │ │ nested task error: On worker 81: │ TaskFailedException │ Stacktrace: │ [1] #wait#582 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] #default_threadpool_checkpoint_call#25 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:284 │ [5] default_threadpool_checkpoint_call │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:250 │ [6] #handle_msg##8 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:303 │ [7] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:70 │ [8] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:79 │ [9] #schedule_call##2 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:88 │ │ nested task error: foo,x=0.9798366964095246 │ Stacktrace: │ [1] error │ @ ./error.jl:44 │ [2] test_save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:448 │ [3] save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1785 │ [4] #28 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:262 │ Stacktrace: │ [1] remotecall_wait(::Function, ::Distributed.Worker, ::Nothing, ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:507 │ [2] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:497 [inlined] │ [3] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:519 [inlined] │ [4] (::Schedulers.var"#34#35"{@Kwargs{}, Nothing, Schedulers.var"#140#141", Schedulers.var"#142#143", Int64, typeof(Schedulers.save_checkpoint), Int64, Tuple{typeof(test_save_checkpoint), typeof(fetch), String, Future, DataType}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:288 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#582 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] remotecall_wait_timeout(::Vector{Float64}, ::Int64, ::Float64, ::Nothing, ::Function, ::Function, ::Int64, ::typeof(Schedulers.save_checkpoint), ::Int64, ::typeof(test_save_checkpoint), ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:298 │ [5] remotecall_wait_timeout(::Vector{Float64}, ::Int64, ::Float64, ::Nothing, ::Function, ::Function, ::Int64, ::Function, ::Int64, ::typeof(test_save_checkpoint), ::Vararg{Any}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:287 │ [6] (::Schedulers.var"#134#135"{Vector{Float32}, @Kwargs{}, typeof(foo7b), Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", typeof(nworkers), Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, String}, Dict{String, Any}, SchedulerOptions{String}, Tuple{Int64, Int64}, Vector{Float64}, Vector{Float64}, Vector{Float64}, Vector{Any}, Dict{Int64, Future}, Int64})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1405 │ │ nested task error: On worker 81: │ TaskFailedException │ Stacktrace: │ [1] #wait#582 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] #default_threadpool_checkpoint_call#25 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:284 │ [5] default_threadpool_checkpoint_call │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:250 │ [6] #handle_msg##8 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:303 │ [7] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:70 │ [8] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:79 │ [9] #schedule_call##2 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:88 │ │ nested task error: foo,x=0.9798366964095246 │ Stacktrace: │ [1] error │ @ ./error.jl:44 │ [2] test_save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:448 │ [3] save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1785 │ [4] #28 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:262 │ Stacktrace: │ [1] remotecall_wait(::Function, ::Distributed.Worker, ::Nothing, ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:507 │ [2] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:497 [inlined] │ [3] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:519 [inlined] │ [4] (::Schedulers.var"#34#35"{@Kwargs{}, Nothing, Schedulers.var"#140#141", Schedulers.var"#142#143", Int64, typeof(Schedulers.save_checkpoint), Int64, Tuple{typeof(test_save_checkpoint), typeof(fetch), String, Future, DataType}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:288 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 ┌ Warning: too many failures on process with id=81 (Schedulers-primary-qO2tilUH), removing from process list └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:393 ┌ Warning: pid=83 (Schedulers-primary-qO2tilUH), checkpoint=nothing, task loop, caught exception during save_checkpoint └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1410 ┌ Warning: TaskFailedException │ │ nested task error: On worker 83: │ TaskFailedException │ Stacktrace: │ [1] #wait#582 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] #default_threadpool_checkpoint_call#25 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:284 │ [5] default_threadpool_checkpoint_call │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:250 │ [6] #handle_msg##8 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:303 │ [7] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:70 │ [8] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:79 │ [9] #schedule_call##2 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:88 │ │ nested task error: foo,x=0.910048941731727 │ Stacktrace: │ [1] error │ @ ./error.jl:44 │ [2] test_save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:448 │ [3] save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1785 │ [4] #28 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:262 │ Stacktrace: │ [1] remotecall_wait(::Function, ::Distributed.Worker, ::Nothing, ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:507 │ [2] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:497 [inlined] │ [3] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:519 [inlined] │ [4] (::Schedulers.var"#34#35"{@Kwargs{}, Nothing, Schedulers.var"#140#141", Schedulers.var"#142#143", Int64, typeof(Schedulers.save_checkpoint), Int64, Tuple{typeof(test_save_checkpoint), typeof(fetch), String, Future, DataType}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:288 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#582 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] remotecall_wait_timeout(::Vector{Float64}, ::Int64, ::Float64, ::Nothing, ::Function, ::Function, ::Int64, ::typeof(Schedulers.save_checkpoint), ::Int64, ::typeof(test_save_checkpoint), ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:298 │ [5] remotecall_wait_timeout(::Vector{Float64}, ::Int64, ::Float64, ::Nothing, ::Function, ::Function, ::Int64, ::Function, ::Int64, ::typeof(test_save_checkpoint), ::Vararg{Any}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:287 │ [6] (::Schedulers.var"#134#135"{Vector{Float32}, @Kwargs{}, typeof(foo7b), Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", typeof(nworkers), Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, String}, Dict{String, Any}, SchedulerOptions{String}, Tuple{Int64, Int64}, Vector{Float64}, Vector{Float64}, Vector{Float64}, Vector{Any}, Dict{Int64, Future}, Int64})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1405 │ │ nested task error: On worker 83: │ TaskFailedException │ Stacktrace: │ [1] #wait#582 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] #default_threadpool_checkpoint_call#25 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:284 │ [5] default_threadpool_checkpoint_call │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:250 │ [6] #handle_msg##8 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:303 │ [7] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:70 │ [8] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:79 │ [9] #schedule_call##2 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:88 │ │ nested task error: foo,x=0.910048941731727 │ Stacktrace: │ [1] error │ @ ./error.jl:44 │ [2] test_save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:448 │ [3] save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1785 │ [4] #28 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:262 │ Stacktrace: │ [1] remotecall_wait(::Function, ::Distributed.Worker, ::Nothing, ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:507 │ [2] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:497 [inlined] │ [3] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:519 [inlined] │ [4] (::Schedulers.var"#34#35"{@Kwargs{}, Nothing, Schedulers.var"#140#141", Schedulers.var"#142#143", Int64, typeof(Schedulers.save_checkpoint), Int64, Tuple{typeof(test_save_checkpoint), typeof(fetch), String, Future, DataType}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:288 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 ┌ Warning: too many failures on process with id=83 (Schedulers-primary-qO2tilUH), removing from process list └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:393 [ Info: running task 10 on process 80 (Schedulers-primary-qO2tilUH); 5 workers total; 92 tasks left in task-pool. [ Info: running task 11 on process 82 (Schedulers-primary-qO2tilUH); 3 workers total; 91 tasks left in task-pool. [ Info: running task 12 on process 79 (Schedulers-primary-qO2tilUH); 3 workers total; 90 tasks left in task-pool. [ Info: running task 13 on process 80 (Schedulers-primary-qO2tilUH); 3 workers total; 89 tasks left in task-pool. [ Info: running task 14 on process 79 (Schedulers-primary-qO2tilUH); 3 workers total; 88 tasks left in task-pool. ┌ Warning: pid=82 (Schedulers-primary-qO2tilUH), checkpoint=/tmp/jl_Ivn5eM/checkpoint-5oRfHi-324, task loop, caught exception during save_checkpoint └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1410 ┌ Warning: TaskFailedException │ │ nested task error: On worker 82: │ TaskFailedException │ Stacktrace: │ [1] #wait#582 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] #default_threadpool_checkpoint_call#25 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:284 │ [5] default_threadpool_checkpoint_call │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:250 │ [6] #handle_msg##8 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:303 │ [7] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:70 │ [8] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:79 │ [9] #schedule_call##2 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:88 │ │ nested task error: foo,x=0.9247325117368228 │ Stacktrace: │ [1] error │ @ ./error.jl:44 │ [2] test_save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:448 │ [3] save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1785 │ [4] #28 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:262 │ Stacktrace: │ [1] remotecall_wait(::Function, ::Distributed.Worker, ::Nothing, ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:507 │ [2] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:497 [inlined] │ [3] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:519 [inlined] │ [4] (::Schedulers.var"#34#35"{@Kwargs{}, Nothing, Schedulers.var"#140#141", Schedulers.var"#142#143", Int64, typeof(Schedulers.save_checkpoint), Int64, Tuple{typeof(test_save_checkpoint), typeof(fetch), String, Future, DataType}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:288 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#582 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] remotecall_wait_timeout(::Vector{Float64}, ::Int64, ::Float64, ::Nothing, ::Function, ::Function, ::Int64, ::typeof(Schedulers.save_checkpoint), ::Int64, ::typeof(test_save_checkpoint), ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:298 │ [5] remotecall_wait_timeout(::Vector{Float64}, ::Int64, ::Float64, ::Nothing, ::Function, ::Function, ::Int64, ::Function, ::Int64, ::typeof(test_save_checkpoint), ::Vararg{Any}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:287 │ [6] (::Schedulers.var"#134#135"{Vector{Float32}, @Kwargs{}, typeof(foo7b), Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", typeof(nworkers), Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, String}, Dict{String, Any}, SchedulerOptions{String}, Tuple{Int64, Int64}, Vector{Float64}, Vector{Float64}, Vector{Float64}, Vector{Any}, Dict{Int64, Future}, Int64})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1405 │ │ nested task error: On worker 82: │ TaskFailedException │ Stacktrace: │ [1] #wait#582 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] #default_threadpool_checkpoint_call#25 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:284 │ [5] default_threadpool_checkpoint_call │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:250 │ [6] #handle_msg##8 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:303 │ [7] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:70 │ [8] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:79 │ [9] #schedule_call##2 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:88 │ │ nested task error: foo,x=0.9247325117368228 │ Stacktrace: │ [1] error │ @ ./error.jl:44 │ [2] test_save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:448 │ [3] save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1785 │ [4] #28 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:262 │ Stacktrace: │ [1] remotecall_wait(::Function, ::Distributed.Worker, ::Nothing, ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:507 │ [2] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:497 [inlined] │ [3] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:519 [inlined] │ [4] (::Schedulers.var"#34#35"{@Kwargs{}, Nothing, Schedulers.var"#140#141", Schedulers.var"#142#143", Int64, typeof(Schedulers.save_checkpoint), Int64, Tuple{typeof(test_save_checkpoint), typeof(fetch), String, Future, DataType}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:288 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 ┌ Warning: too many failures on process with id=82 (Schedulers-primary-qO2tilUH), removing from process list └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:393 [ Info: running task 15 on process 80 (Schedulers-primary-qO2tilUH); 3 workers total; 88 tasks left in task-pool. [ Info: running task 16 on process 79 (Schedulers-primary-qO2tilUH); 3 workers total; 87 tasks left in task-pool. ┌ Warning: pid=80 (Schedulers-primary-qO2tilUH), checkpoint=/tmp/jl_Ivn5eM/checkpoint-5oRfHi-329, task loop, caught exception during save_checkpoint └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1410 ┌ Warning: TaskFailedException │ │ nested task error: On worker 80: │ TaskFailedException │ Stacktrace: │ [1] #wait#582 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] #default_threadpool_checkpoint_call#25 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:284 │ [5] default_threadpool_checkpoint_call │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:250 │ [6] #handle_msg##8 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:303 │ [7] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:70 │ [8] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:79 │ [9] #schedule_call##2 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:88 │ │ nested task error: foo,x=0.9822371987357681 │ Stacktrace: │ [1] error │ @ ./error.jl:44 │ [2] test_save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:448 │ [3] save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1785 │ [4] #28 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:262 │ Stacktrace: │ [1] remotecall_wait(::Function, ::Distributed.Worker, ::Nothing, ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:507 │ [2] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:497 [inlined] │ [3] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:519 [inlined] │ [4] (::Schedulers.var"#34#35"{@Kwargs{}, Nothing, Schedulers.var"#140#141", Schedulers.var"#142#143", Int64, typeof(Schedulers.save_checkpoint), Int64, Tuple{typeof(test_save_checkpoint), typeof(fetch), String, Future, DataType}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:288 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#582 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] remotecall_wait_timeout(::Vector{Float64}, ::Int64, ::Float64, ::Nothing, ::Function, ::Function, ::Int64, ::typeof(Schedulers.save_checkpoint), ::Int64, ::typeof(test_save_checkpoint), ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:298 │ [5] remotecall_wait_timeout(::Vector{Float64}, ::Int64, ::Float64, ::Nothing, ::Function, ::Function, ::Int64, ::Function, ::Int64, ::typeof(test_save_checkpoint), ::Vararg{Any}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:287 │ [6] (::Schedulers.var"#134#135"{Vector{Float32}, @Kwargs{}, typeof(foo7b), Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", typeof(nworkers), Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, String}, Dict{String, Any}, SchedulerOptions{String}, Tuple{Int64, Int64}, Vector{Float64}, Vector{Float64}, Vector{Float64}, Vector{Any}, Dict{Int64, Future}, Int64})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1405 │ │ nested task error: On worker 80: │ TaskFailedException │ Stacktrace: │ [1] #wait#582 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] #default_threadpool_checkpoint_call#25 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:284 │ [5] default_threadpool_checkpoint_call │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:250 │ [6] #handle_msg##8 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:303 │ [7] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:70 │ [8] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:79 │ [9] #schedule_call##2 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:88 │ │ nested task error: foo,x=0.9822371987357681 │ Stacktrace: │ [1] error │ @ ./error.jl:44 │ [2] test_save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:448 │ [3] save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1785 │ [4] #28 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:262 │ Stacktrace: │ [1] remotecall_wait(::Function, ::Distributed.Worker, ::Nothing, ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:507 │ [2] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:497 [inlined] │ [3] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:519 [inlined] │ [4] (::Schedulers.var"#34#35"{@Kwargs{}, Nothing, Schedulers.var"#140#141", Schedulers.var"#142#143", Int64, typeof(Schedulers.save_checkpoint), Int64, Tuple{typeof(test_save_checkpoint), typeof(fetch), String, Future, DataType}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:288 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 ┌ Warning: too many failures on process with id=80 (Schedulers-primary-qO2tilUH), removing from process list └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:393 [ Info: running task 17 on process 79 (Schedulers-primary-qO2tilUH); 5 workers total; 87 tasks left in task-pool. [ Info: running task 18 on process 79 (Schedulers-primary-qO2tilUH); 5 workers total; 86 tasks left in task-pool. [ Info: running task 19 on process 79 (Schedulers-primary-qO2tilUH); 5 workers total; 85 tasks left in task-pool. [ Info: running task 20 on process 79 (Schedulers-primary-qO2tilUH); 5 workers total; 84 tasks left in task-pool. ┌ Warning: pid=79 (Schedulers-primary-qO2tilUH), checkpoint=/tmp/jl_Ivn5eM/checkpoint-5oRfHi-335, task loop, caught exception during save_checkpoint └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1410 ┌ Warning: TaskFailedException │ │ nested task error: On worker 79: │ TaskFailedException │ Stacktrace: │ [1] #wait#582 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] #default_threadpool_checkpoint_call#25 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:284 │ [5] default_threadpool_checkpoint_call │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:250 │ [6] #handle_msg##8 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:303 │ [7] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:70 │ [8] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:79 │ [9] #schedule_call##2 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:88 │ │ nested task error: foo,x=0.8379436529080888 │ Stacktrace: │ [1] error │ @ ./error.jl:44 │ [2] test_save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:448 │ [3] save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1785 │ [4] #28 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:262 │ Stacktrace: │ [1] remotecall_wait(::Function, ::Distributed.Worker, ::Nothing, ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:507 │ [2] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:497 [inlined] │ [3] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:519 [inlined] │ [4] (::Schedulers.var"#34#35"{@Kwargs{}, Nothing, Schedulers.var"#140#141", Schedulers.var"#142#143", Int64, typeof(Schedulers.save_checkpoint), Int64, Tuple{typeof(test_save_checkpoint), typeof(fetch), String, Future, DataType}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:288 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#582 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] remotecall_wait_timeout(::Vector{Float64}, ::Int64, ::Float64, ::Nothing, ::Function, ::Function, ::Int64, ::typeof(Schedulers.save_checkpoint), ::Int64, ::typeof(test_save_checkpoint), ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:298 │ [5] remotecall_wait_timeout(::Vector{Float64}, ::Int64, ::Float64, ::Nothing, ::Function, ::Function, ::Int64, ::Function, ::Int64, ::typeof(test_save_checkpoint), ::Vararg{Any}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:287 │ [6] (::Schedulers.var"#134#135"{Vector{Float32}, @Kwargs{}, typeof(foo7b), Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", typeof(nworkers), Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, String}, Dict{String, Any}, SchedulerOptions{String}, Tuple{Int64, Int64}, Vector{Float64}, Vector{Float64}, Vector{Float64}, Vector{Any}, Dict{Int64, Future}, Int64})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1405 │ │ nested task error: On worker 79: │ TaskFailedException │ Stacktrace: │ [1] #wait#582 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] #default_threadpool_checkpoint_call#25 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:284 │ [5] default_threadpool_checkpoint_call │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:250 │ [6] #handle_msg##8 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:303 │ [7] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:70 │ [8] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:79 │ [9] #schedule_call##2 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:88 │ │ nested task error: foo,x=0.8379436529080888 │ Stacktrace: │ [1] error │ @ ./error.jl:44 │ [2] test_save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:448 │ [3] save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1785 │ [4] #28 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:262 │ Stacktrace: │ [1] remotecall_wait(::Function, ::Distributed.Worker, ::Nothing, ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:507 │ [2] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:497 [inlined] │ [3] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:519 [inlined] │ [4] (::Schedulers.var"#34#35"{@Kwargs{}, Nothing, Schedulers.var"#140#141", Schedulers.var"#142#143", Int64, typeof(Schedulers.save_checkpoint), Int64, Tuple{typeof(test_save_checkpoint), typeof(fetch), String, Future, DataType}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:288 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 ┌ Warning: too many failures on process with id=79 (Schedulers-primary-qO2tilUH), removing from process list └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:393 [ Info: running task 21 on process 84 (Schedulers-primary-qO2tilUH); 5 workers total; 84 tasks left in task-pool. [ Info: running task 22 on process 85 (Schedulers-primary-qO2tilUH); 5 workers total; 83 tasks left in task-pool. [ Info: running task 23 on process 84 (Schedulers-primary-qO2tilUH); 5 workers total; 82 tasks left in task-pool. [ Info: running task 24 on process 85 (Schedulers-primary-qO2tilUH); 5 workers total; 81 tasks left in task-pool. [ Info: running task 25 on process 84 (Schedulers-primary-qO2tilUH); 5 workers total; 80 tasks left in task-pool. [ Info: running task 26 on process 85 (Schedulers-primary-qO2tilUH); 5 workers total; 79 tasks left in task-pool. [ Info: running task 27 on process 84 (Schedulers-primary-qO2tilUH); 4 workers total; 78 tasks left in task-pool. [ Info: running task 28 on process 85 (Schedulers-primary-qO2tilUH); 4 workers total; 77 tasks left in task-pool. [ Info: running task 29 on process 85 (Schedulers-primary-qO2tilUH); 4 workers total; 76 tasks left in task-pool. ┌ Warning: pid=84 (Schedulers-primary-qO2tilUH), checkpoint=/tmp/jl_Ivn5eM/checkpoint-5oRfHi-341, task loop, caught exception during save_checkpoint └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1410 ┌ Warning: TaskFailedException │ │ nested task error: On worker 84: │ TaskFailedException │ Stacktrace: │ [1] #wait#582 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] #default_threadpool_checkpoint_call#25 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:284 │ [5] default_threadpool_checkpoint_call │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:250 │ [6] #handle_msg##8 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:303 │ [7] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:70 │ [8] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:79 │ [9] #schedule_call##2 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:88 │ │ nested task error: foo,x=0.8832459870940537 │ Stacktrace: │ [1] error │ @ ./error.jl:44 │ [2] test_save_checkpoint │ @ ./none:-1 [inlined] │ [3] test_save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:0 │ [4] save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1785 │ [5] #28 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:262 │ Stacktrace: │ [1] remotecall_wait(::Function, ::Distributed.Worker, ::Nothing, ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:507 │ [2] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:497 [inlined] │ [3] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:519 [inlined] │ [4] (::Schedulers.var"#34#35"{@Kwargs{}, Nothing, Schedulers.var"#140#141", Schedulers.var"#142#143", Int64, typeof(Schedulers.save_checkpoint), Int64, Tuple{typeof(test_save_checkpoint), typeof(fetch), String, Future, DataType}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:288 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#582 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] remotecall_wait_timeout(::Vector{Float64}, ::Int64, ::Float64, ::Nothing, ::Function, ::Function, ::Int64, ::typeof(Schedulers.save_checkpoint), ::Int64, ::typeof(test_save_checkpoint), ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:298 │ [5] remotecall_wait_timeout(::Vector{Float64}, ::Int64, ::Float64, ::Nothing, ::Function, ::Function, ::Int64, ::Function, ::Int64, ::typeof(test_save_checkpoint), ::Vararg{Any}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:287 │ [6] (::Schedulers.var"#134#135"{Vector{Float32}, @Kwargs{}, typeof(foo7b), Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", typeof(nworkers), Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, String}, Dict{String, Any}, SchedulerOptions{String}, Tuple{Int64, Int64}, Vector{Float64}, Vector{Float64}, Vector{Float64}, Vector{Any}, Dict{Int64, Future}, Int64})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1405 │ │ nested task error: On worker 84: │ TaskFailedException │ Stacktrace: │ [1] #wait#582 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] #default_threadpool_checkpoint_call#25 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:284 │ [5] default_threadpool_checkpoint_call │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:250 │ [6] #handle_msg##8 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:303 │ [7] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:70 │ [8] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:79 │ [9] #schedule_call##2 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:88 │ │ nested task error: foo,x=0.8832459870940537 │ Stacktrace: │ [1] error │ @ ./error.jl:44 │ [2] test_save_checkpoint │ @ ./none:-1 [inlined] │ [3] test_save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:0 │ [4] save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1785 │ [5] #28 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:262 │ Stacktrace: │ [1] remotecall_wait(::Function, ::Distributed.Worker, ::Nothing, ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:507 │ [2] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:497 [inlined] │ [3] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:519 [inlined] │ [4] (::Schedulers.var"#34#35"{@Kwargs{}, Nothing, Schedulers.var"#140#141", Schedulers.var"#142#143", Int64, typeof(Schedulers.save_checkpoint), Int64, Tuple{typeof(test_save_checkpoint), typeof(fetch), String, Future, DataType}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:288 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 ┌ Warning: too many failures on process with id=84 (Schedulers-primary-qO2tilUH), removing from process list └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:393 ┌ Warning: pid=85 (Schedulers-primary-qO2tilUH), checkpoint=/tmp/jl_Ivn5eM/checkpoint-5oRfHi-344, task loop, caught exception during save_checkpoint └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1410 ┌ Warning: TaskFailedException │ │ nested task error: On worker 85: │ TaskFailedException │ Stacktrace: │ [1] #wait#582 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] #default_threadpool_checkpoint_call#25 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:284 │ [5] default_threadpool_checkpoint_call │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:250 │ [6] #handle_msg##8 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:303 │ [7] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:70 │ [8] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:79 │ [9] #schedule_call##2 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:88 │ │ nested task error: foo,x=0.9207017944341208 │ Stacktrace: │ [1] error │ @ ./error.jl:44 │ [2] test_save_checkpoint │ @ ./none:-1 [inlined] │ [3] test_save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:0 │ [4] save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1785 │ [5] #28 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:262 │ Stacktrace: │ [1] remotecall_wait(::Function, ::Distributed.Worker, ::Nothing, ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:507 │ [2] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:497 [inlined] │ [3] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:519 [inlined] │ [4] (::Schedulers.var"#34#35"{@Kwargs{}, Nothing, Schedulers.var"#140#141", Schedulers.var"#142#143", Int64, typeof(Schedulers.save_checkpoint), Int64, Tuple{typeof(test_save_checkpoint), typeof(fetch), String, Future, DataType}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:288 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#582 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] remotecall_wait_timeout(::Vector{Float64}, ::Int64, ::Float64, ::Nothing, ::Function, ::Function, ::Int64, ::typeof(Schedulers.save_checkpoint), ::Int64, ::typeof(test_save_checkpoint), ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:298 │ [5] remotecall_wait_timeout(::Vector{Float64}, ::Int64, ::Float64, ::Nothing, ::Function, ::Function, ::Int64, ::Function, ::Int64, ::typeof(test_save_checkpoint), ::Vararg{Any}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:287 │ [6] (::Schedulers.var"#134#135"{Vector{Float32}, @Kwargs{}, typeof(foo7b), Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", typeof(nworkers), Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, String}, Dict{String, Any}, SchedulerOptions{String}, Tuple{Int64, Int64}, Vector{Float64}, Vector{Float64}, Vector{Float64}, Vector{Any}, Dict{Int64, Future}, Int64})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1405 │ │ nested task error: On worker 85: │ TaskFailedException │ Stacktrace: │ [1] #wait#582 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] #default_threadpool_checkpoint_call#25 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:284 │ [5] default_threadpool_checkpoint_call │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:250 │ [6] #handle_msg##8 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:303 │ [7] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:70 │ [8] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:79 │ [9] #schedule_call##2 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:88 │ │ nested task error: foo,x=0.9207017944341208 │ Stacktrace: │ [1] error │ @ ./error.jl:44 │ [2] test_save_checkpoint │ @ ./none:-1 [inlined] │ [3] test_save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:0 │ [4] save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1785 │ [5] #28 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:262 │ Stacktrace: │ [1] remotecall_wait(::Function, ::Distributed.Worker, ::Nothing, ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:507 │ [2] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:497 [inlined] │ [3] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:519 [inlined] │ [4] (::Schedulers.var"#34#35"{@Kwargs{}, Nothing, Schedulers.var"#140#141", Schedulers.var"#142#143", Int64, typeof(Schedulers.save_checkpoint), Int64, Tuple{typeof(test_save_checkpoint), typeof(fetch), String, Future, DataType}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:288 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 ┌ Warning: too many failures on process with id=85 (Schedulers-primary-qO2tilUH), removing from process list └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:393 [ Info: running task 30 on process 86 (Schedulers-primary-qO2tilUH); 5 workers total; 77 tasks left in task-pool. [ Info: running task 31 on process 87 (Schedulers-primary-qO2tilUH); 5 workers total; 76 tasks left in task-pool. [ Info: running task 32 on process 86 (Schedulers-primary-qO2tilUH); 5 workers total; 75 tasks left in task-pool. [ Info: running task 33 on process 87 (Schedulers-primary-qO2tilUH); 5 workers total; 74 tasks left in task-pool. [ Info: running task 34 on process 87 (Schedulers-primary-qO2tilUH); 3 workers total; 73 tasks left in task-pool. ┌ Warning: pid=86 (Schedulers-primary-qO2tilUH), checkpoint=/tmp/jl_Ivn5eM/checkpoint-5oRfHi-346, task loop, caught exception during save_checkpoint └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1410 ┌ Warning: TaskFailedException │ │ nested task error: On worker 86: │ TaskFailedException │ Stacktrace: │ [1] #wait#582 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] #default_threadpool_checkpoint_call#25 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:284 │ [5] default_threadpool_checkpoint_call │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:250 │ [6] #handle_msg##8 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:303 │ [7] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:70 │ [8] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:79 │ [9] #schedule_call##2 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:88 │ │ nested task error: foo,x=0.8538002492603152 │ Stacktrace: │ [1] error │ @ ./error.jl:44 │ [2] test_save_checkpoint │ @ ./none:-1 [inlined] │ [3] test_save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:0 │ [4] save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1785 │ [5] #28 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:262 │ Stacktrace: │ [1] remotecall_wait(::Function, ::Distributed.Worker, ::Nothing, ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:507 │ [2] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:497 [inlined] │ [3] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:519 [inlined] │ [4] (::Schedulers.var"#34#35"{@Kwargs{}, Nothing, Schedulers.var"#140#141", Schedulers.var"#142#143", Int64, typeof(Schedulers.save_checkpoint), Int64, Tuple{typeof(test_save_checkpoint), typeof(fetch), String, Future, DataType}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:288 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#582 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] remotecall_wait_timeout(::Vector{Float64}, ::Int64, ::Float64, ::Nothing, ::Function, ::Function, ::Int64, ::typeof(Schedulers.save_checkpoint), ::Int64, ::typeof(test_save_checkpoint), ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:298 │ [5] remotecall_wait_timeout(::Vector{Float64}, ::Int64, ::Float64, ::Nothing, ::Function, ::Function, ::Int64, ::Function, ::Int64, ::typeof(test_save_checkpoint), ::Vararg{Any}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:287 │ [6] (::Schedulers.var"#134#135"{Vector{Float32}, @Kwargs{}, typeof(foo7b), Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", typeof(nworkers), Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, String}, Dict{String, Any}, SchedulerOptions{String}, Tuple{Int64, Int64}, Vector{Float64}, Vector{Float64}, Vector{Float64}, Vector{Any}, Dict{Int64, Future}, Int64})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1405 │ │ nested task error: On worker 86: │ TaskFailedException │ Stacktrace: │ [1] #wait#582 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] #default_threadpool_checkpoint_call#25 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:284 │ [5] default_threadpool_checkpoint_call │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:250 │ [6] #handle_msg##8 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:303 │ [7] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:70 │ [8] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:79 │ [9] #schedule_call##2 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:88 │ │ nested task error: foo,x=0.8538002492603152 │ Stacktrace: │ [1] error │ @ ./error.jl:44 │ [2] test_save_checkpoint │ @ ./none:-1 [inlined] │ [3] test_save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:0 │ [4] save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1785 │ [5] #28 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:262 │ Stacktrace: │ [1] remotecall_wait(::Function, ::Distributed.Worker, ::Nothing, ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:507 │ [2] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:497 [inlined] │ [3] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:519 [inlined] │ [4] (::Schedulers.var"#34#35"{@Kwargs{}, Nothing, Schedulers.var"#140#141", Schedulers.var"#142#143", Int64, typeof(Schedulers.save_checkpoint), Int64, Tuple{typeof(test_save_checkpoint), typeof(fetch), String, Future, DataType}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:288 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 ┌ Warning: too many failures on process with id=86 (Schedulers-primary-qO2tilUH), removing from process list └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:393 [ Info: running task 35 on process 87 (Schedulers-primary-qO2tilUH); 3 workers total; 73 tasks left in task-pool. [ Info: running task 36 on process 87 (Schedulers-primary-qO2tilUH); 3 workers total; 72 tasks left in task-pool. [ Info: running task 37 on process 87 (Schedulers-primary-qO2tilUH); 3 workers total; 71 tasks left in task-pool. [ Info: running task 38 on process 87 (Schedulers-primary-qO2tilUH); 5 workers total; 70 tasks left in task-pool. [ Info: running task 39 on process 88 (Schedulers-primary-qO2tilUH); 5 workers total; 69 tasks left in task-pool. [ Info: running task 40 on process 87 (Schedulers-primary-qO2tilUH); 5 workers total; 68 tasks left in task-pool. ┌ Warning: pid=87 (Schedulers-primary-qO2tilUH), checkpoint=/tmp/jl_Ivn5eM/checkpoint-5oRfHi-354, task loop, caught exception during save_checkpoint └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1410 ┌ Warning: TaskFailedException │ │ nested task error: On worker 87: │ TaskFailedException │ Stacktrace: │ [1] #wait#582 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] #default_threadpool_checkpoint_call#25 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:284 │ [5] default_threadpool_checkpoint_call │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:250 │ [6] #handle_msg##8 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:303 │ [7] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:70 │ [8] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:79 │ [9] #schedule_call##2 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:88 │ │ nested task error: foo,x=0.8277026883132625 │ Stacktrace: │ [1] error │ @ ./error.jl:44 │ [2] test_save_checkpoint │ @ ./none:-1 [inlined] │ [3] test_save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:0 │ [4] save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1785 │ [5] #28 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:262 │ Stacktrace: │ [1] remotecall_wait(::Function, ::Distributed.Worker, ::Nothing, ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:507 │ [2] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:497 [inlined] │ [3] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:519 [inlined] │ [4] (::Schedulers.var"#34#35"{@Kwargs{}, Nothing, Schedulers.var"#140#141", Schedulers.var"#142#143", Int64, typeof(Schedulers.save_checkpoint), Int64, Tuple{typeof(test_save_checkpoint), typeof(fetch), String, Future, DataType}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:288 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#582 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] remotecall_wait_timeout(::Vector{Float64}, ::Int64, ::Float64, ::Nothing, ::Function, ::Function, ::Int64, ::typeof(Schedulers.save_checkpoint), ::Int64, ::typeof(test_save_checkpoint), ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:298 │ [5] remotecall_wait_timeout(::Vector{Float64}, ::Int64, ::Float64, ::Nothing, ::Function, ::Function, ::Int64, ::Function, ::Int64, ::typeof(test_save_checkpoint), ::Vararg{Any}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:287 │ [6] (::Schedulers.var"#134#135"{Vector{Float32}, @Kwargs{}, typeof(foo7b), Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", typeof(nworkers), Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, String}, Dict{String, Any}, SchedulerOptions{String}, Tuple{Int64, Int64}, Vector{Float64}, Vector{Float64}, Vector{Float64}, Vector{Any}, Dict{Int64, Future}, Int64})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1405 │ │ nested task error: On worker 87: │ TaskFailedException │ Stacktrace: │ [1] #wait#582 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] #default_threadpool_checkpoint_call#25 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:284 │ [5] default_threadpool_checkpoint_call │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:250 │ [6] #handle_msg##8 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:303 │ [7] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:70 │ [8] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:79 │ [9] #schedule_call##2 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:88 │ │ nested task error: foo,x=0.8277026883132625 │ Stacktrace: │ [1] error │ @ ./error.jl:44 │ [2] test_save_checkpoint │ @ ./none:-1 [inlined] │ [3] test_save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:0 │ [4] save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1785 │ [5] #28 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:262 │ Stacktrace: │ [1] remotecall_wait(::Function, ::Distributed.Worker, ::Nothing, ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:507 │ [2] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:497 [inlined] │ [3] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:519 [inlined] │ [4] (::Schedulers.var"#34#35"{@Kwargs{}, Nothing, Schedulers.var"#140#141", Schedulers.var"#142#143", Int64, typeof(Schedulers.save_checkpoint), Int64, Tuple{typeof(test_save_checkpoint), typeof(fetch), String, Future, DataType}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:288 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 ┌ Warning: too many failures on process with id=87 (Schedulers-primary-qO2tilUH), removing from process list └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:393 From worker 90: From worker 90: ====================================================================================== From worker 90: Information request received. A stacktrace will print followed by a 1.0 second profile From worker 90: ====================================================================================== From worker 90: From worker 90: cmd: /opt/julia/bin/julia 459 running 1 of 2 From worker 90: From worker 90: signal (10): User defined signal 1 From worker 90: unknown function (ip: 0x726a111b4f14) at /lib/x86_64-linux-gnu/libc.so.6 From worker 90: pthread_cond_wait at /lib/x86_64-linux-gnu/libc.so.6 (unknown line) From worker 90: uv_cond_wait at /workspace/srcdir/libuv/src/unix/thread.c:822 From worker 90: ijl_task_get_next at /source/src/scheduler.c:520 From worker 90: wait at ./task.jl:1216 From worker 90: wait_forever at ./task.jl:1150 From worker 90: jfptr_wait_forever_30682.1 at /opt/julia/lib/julia/sys.so (unknown line) From worker 90: _jl_invoke at /source/src/gf.c:3490 [inlined] From worker 90: ijl_apply_generic at /source/src/gf.c:3690 From worker 90: jl_apply at /source/src/julia.h:2336 [inlined] From worker 90: start_task at /source/src/task.c:1249 From worker 90: unknown function (ip: (nil)) at (unknown file) From worker 90: _ZNK4llvm9LiveRange8overlapsERKS0_RKNS_13CoalescerPairERKNS_11SlotIndexesE at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) From worker 90: _ZN4llvm13LiveRegMatrix24checkRegUnitInterferenceERKNS_12LiveIntervalENS_10MCRegisterE at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) From worker 90: _ZN4llvm13LiveRegMatrix17checkInterferenceERKNS_12LiveIntervalENS_10MCRegisterE at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) From worker 90: _ZN4llvm8RAGreedy9tryAssignERKNS_12LiveIntervalERNS_15AllocationOrderERNS_15SmallVectorImplINS_8RegisterEEERKNS_8SmallSetIS7_Lj16ESt4lessIS7_EEE at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) From worker 90: _ZN4llvm8RAGreedy17selectOrSplitImplERKNS_12LiveIntervalERNS_15SmallVectorImplINS_8RegisterEEERNS_8SmallSetIS5_Lj16ESt4lessIS5_EEERNS_11SmallVectorISt4pairIPS2_NS_10MCRegisterEELj8EEEj at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) From worker 90: _ZN4llvm8RAGreedy13selectOrSplitERKNS_12LiveIntervalERNS_15SmallVectorImplINS_8RegisterEEE at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) From worker 90: _ZN4llvm12RegAllocBase16allocatePhysRegsEv at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) From worker 90: _ZN4llvm8RAGreedy20runOnMachineFunctionERNS_15MachineFunctionE at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) From worker 90: _ZN4llvm19MachineFunctionPass13runOnFunctionERNS_8FunctionE.part.0 at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) From worker 90: _ZN4llvm13FPPassManager13runOnFunctionERNS_8FunctionE at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) From worker 90: _ZN4llvm13FPPassManager11runOnModuleERNS_6ModuleE at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) From worker 90: _ZN4llvm6legacy15PassManagerImpl3runERNS_6ModuleE at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) From worker 90: _ZN4llvm3orc14SimpleCompilerclERNS_6ModuleE at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) From worker 90: operator() at /source/src/jitlayers.cpp:1577 From worker 90: addModule at /source/src/jitlayers.cpp:2058 From worker 90: jl_compile_codeinst_now at /source/src/jitlayers.cpp:649 From worker 90: jl_compile_codeinst_impl at /source/src/jitlayers.cpp:840 From worker 90: jl_compile_method_internal at /source/src/gf.c:3004 From worker 90: _jl_invoke at /source/src/gf.c:3482 [inlined] From worker 90: ijl_apply_generic at /source/src/gf.c:3690 From worker 90: jl_apply at /source/src/julia.h:2336 [inlined] From worker 90: start_task at /source/src/task.c:1249 From worker 90: unknown function (ip: (nil)) at (unknown file) From worker 90: From worker 90: ============================================================== From worker 90: Profile collected. A report will print at the next yield point From worker 90: ============================================================== From worker 90: [ Info: running task 41 on process 88 (Schedulers-primary-qO2tilUH); 5 workers total; 68 tasks left in task-pool. [ Info: running task 42 on process 88 (Schedulers-primary-qO2tilUH); 5 workers total; 67 tasks left in task-pool. From worker 89: From worker 89: ====================================================================================== From worker 89: Information request received. A stacktrace will print followed by a 1.0 second profile From worker 89: ====================================================================================== From worker 89: From worker 89: cmd: /opt/julia/bin/julia 457 running 1 of 2 From worker 89: From worker 89: signal (10): User defined signal 1 From worker 89: unknown function (ip: 0x7c5c528a7f14) at /lib/x86_64-linux-gnu/libc.so.6 From worker 89: pthread_cond_wait at /lib/x86_64-linux-gnu/libc.so.6 (unknown line) From worker 89: uv_cond_wait at /workspace/srcdir/libuv/src/unix/thread.c:822 From worker 89: ijl_task_get_next at /source/src/scheduler.c:520 From worker 89: wait at ./task.jl:1216 From worker 89: wait_forever at ./task.jl:1150 From worker 89: jfptr_wait_forever_30682.1 at /opt/julia/lib/julia/sys.so (unknown line) From worker 89: _jl_invoke at /source/src/gf.c:3490 [inlined] From worker 89: ijl_apply_generic at /source/src/gf.c:3690 From worker 89: jl_apply at /source/src/julia.h:2336 [inlined] From worker 89: start_task at /source/src/task.c:1249 From worker 89: unknown function (ip: (nil)) at (unknown file) From worker 89: _ZNK4llvm5Value30assertModuleIsMaterializedImplEv at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) From worker 89: unknown function (ip: (nil)) at (unknown file) From worker 89: From worker 89: ============================================================== From worker 89: Profile collected. A report will print at the next yield point From worker 89: ============================================================== From worker 89: ┌ Warning: pid=88 (Schedulers-primary-qO2tilUH), checkpoint=/tmp/jl_Ivn5eM/checkpoint-5oRfHi-357, task loop, caught exception during save_checkpoint └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1410 ┌ Warning: TaskFailedException │ │ nested task error: On worker 88: │ TaskFailedException │ Stacktrace: │ [1] #wait#582 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] #default_threadpool_checkpoint_call#25 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:284 │ [5] default_threadpool_checkpoint_call │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:250 │ [6] #handle_msg##8 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:303 │ [7] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:70 │ [8] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:79 │ [9] #schedule_call##2 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:88 │ │ nested task error: foo,x=0.8224625244752827 │ Stacktrace: │ [1] error │ @ ./error.jl:44 │ [2] test_save_checkpoint │ @ ./none:-1 [inlined] │ [3] test_save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:0 │ [4] save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1785 │ [5] #28 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:262 │ Stacktrace: │ [1] remotecall_wait(::Function, ::Distributed.Worker, ::Nothing, ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:507 │ [2] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:497 [inlined] │ [3] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:519 [inlined] │ [4] (::Schedulers.var"#34#35"{@Kwargs{}, Nothing, Schedulers.var"#140#141", Schedulers.var"#142#143", Int64, typeof(Schedulers.save_checkpoint), Int64, Tuple{typeof(test_save_checkpoint), typeof(fetch), String, Future, DataType}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:288 │ error type: TaskFailedException │ 1-element ExceptionStack: │ TaskFailedException │ Stacktrace: │ [1] #wait#582 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] remotecall_wait_timeout(::Vector{Float64}, ::Int64, ::Float64, ::Nothing, ::Function, ::Function, ::Int64, ::typeof(Schedulers.save_checkpoint), ::Int64, ::typeof(test_save_checkpoint), ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:298 │ [5] remotecall_wait_timeout(::Vector{Float64}, ::Int64, ::Float64, ::Nothing, ::Function, ::Function, ::Int64, ::Function, ::Int64, ::typeof(test_save_checkpoint), ::Vararg{Any}) │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:287 │ [6] (::Schedulers.var"#134#135"{Vector{Float32}, @Kwargs{}, typeof(foo7b), Schedulers.ElasticLoop{typeof(addprocs), Schedulers.var"#14#15", typeof(nworkers), Schedulers.var"#92#93"{Int64}, Schedulers.var"#SchedulerOptions##14#SchedulerOptions##15", Schedulers.var"#SchedulerOptions##24#SchedulerOptions##25", Schedulers.var"#SchedulerOptions##26#SchedulerOptions##27", Schedulers.var"#SchedulerOptions##16#SchedulerOptions##17", Int64, String}, Dict{String, Any}, SchedulerOptions{String}, Tuple{Int64, Int64}, Vector{Float64}, Vector{Float64}, Vector{Float64}, Vector{Any}, Dict{Int64, Future}, Int64})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1405 │ │ nested task error: On worker 88: │ TaskFailedException │ Stacktrace: │ [1] #wait#582 │ @ ./task.jl:363 [inlined] │ [2] wait │ @ ./task.jl:360 [inlined] │ [3] fetch │ @ ./task.jl:525 [inlined] │ [4] #default_threadpool_checkpoint_call#25 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:284 │ [5] default_threadpool_checkpoint_call │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:250 │ [6] #handle_msg##8 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:303 │ [7] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:70 │ [8] run_work_thunk │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:79 │ [9] #schedule_call##2 │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/process_messages.jl:88 │ │ nested task error: foo,x=0.8224625244752827 │ Stacktrace: │ [1] error │ @ ./error.jl:44 │ [2] test_save_checkpoint │ @ ./none:-1 [inlined] │ [3] test_save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/test/runtests.jl:0 │ [4] save_checkpoint │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:1785 │ [5] #28 │ @ ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:262 │ Stacktrace: │ [1] remotecall_wait(::Function, ::Distributed.Worker, ::Nothing, ::Vararg{Any}; kwargs::@Kwargs{}) │ @ Distributed /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:507 │ [2] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:497 [inlined] │ [3] remotecall_wait │ @ /opt/julia/share/julia/stdlib/v1.13/Distributed/src/remotecall.jl:519 [inlined] │ [4] (::Schedulers.var"#34#35"{@Kwargs{}, Nothing, Schedulers.var"#140#141", Schedulers.var"#142#143", Int64, typeof(Schedulers.save_checkpoint), Int64, Tuple{typeof(test_save_checkpoint), typeof(fetch), String, Future, DataType}})() │ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:288 └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:23 ┌ Warning: too many failures on process with id=88 (Schedulers-primary-qO2tilUH), removing from process list └ @ Schedulers ~/.julia/packages/Schedulers/JBB7J/src/Schedulers.jl:393 From worker 88: From worker 88: ====================================================================================== From worker 88: Information request received. A stacktrace will print followed by a 1.0 second profile From worker 88: ====================================================================================== From worker 88: From worker 88: cmd: /opt/julia/bin/julia 452 running 0 of 2 From worker 88: From worker 88: signal (10): User defined signal 1 From worker 88: unknown function (ip: 0x712f292baf14) at /lib/x86_64-linux-gnu/libc.so.6 From worker 88: pthread_cond_wait at /lib/x86_64-linux-gnu/libc.so.6 (unknown line) From worker 88: uv_cond_wait at /workspace/srcdir/libuv/src/unix/thread.c:822 From worker 88: ijl_task_get_next at /source/src/scheduler.c:520 From worker 88: wait at ./task.jl:1216 From worker 88: wait_forever at ./task.jl:1150 From worker 88: jfptr_wait_forever_30682.1 at /opt/julia/lib/julia/sys.so (unknown line) From worker 88: _jl_invoke at /source/src/gf.c:3490 [inlined] From worker 88: ijl_apply_generic at /source/src/gf.c:3690 From worker 88: jl_apply at /source/src/julia.h:2336 [inlined] From worker 88: start_task at /source/src/task.c:1249 From worker 88: unknown function (ip: (nil)) at (unknown file) From worker 88: epoll_pwait at /lib/x86_64-linux-gnu/libc.so.6 (unknown line) From worker 88: uv__io_poll at /workspace/srcdir/libuv/src/unix/linux.c:1404 From worker 88: uv_run at /workspace/srcdir/libuv/src/unix/core.c:430 From worker 88: ijl_task_get_next at /source/src/scheduler.c:454 From worker 88: wait at ./task.jl:1216 From worker 88: wait_forever at ./task.jl:1150 From worker 88: jfptr_wait_forever_30682.1 at /opt/julia/lib/julia/sys.so (unknown line) From worker 88: _jl_invoke at /source/src/gf.c:3490 [inlined] From worker 88: ijl_apply_generic at /source/src/gf.c:3690 From worker 88: jl_apply at /source/src/julia.h:2336 [inlined] From worker 88: start_task at /source/src/task.c:1249 From worker 88: unknown function (ip: (nil)) at (unknown file) From worker 88: From worker 88: ============================================================== From worker 88: Profile collected. A report will print at the next yield point From worker 88: ============================================================== From worker 88: From worker 90: ┌ Warning: There were no samples collected in one or more groups. From worker 90: │ This may be due to idle threads, or you may need to run your From worker 90: │ program longer (perhaps by running it multiple times), From worker 90: │ or adjust the delay between samples with `Profile.init()`. From worker 90: └ @ Profile /opt/julia/share/julia/stdlib/v1.13/Profile/src/Profile.jl:1362 From worker 90: Overhead ╎ [+additional indent] Count File:Line Function From worker 90: ========================================================= From worker 90: Thread 1 (interactive) Task 0x00007269f9caafe0 Total snapshots: 1. Utilization: 100% From worker 90: From worker 90: Thread 2 (default) Task 0x00007269fa5fc100 Total snapshots: 1. Utilization: 0% From worker 90: ╎1 @Base/task.jl:1150 wait_forever() From worker 90: ╎ 1 @Base/task.jl:1216 wait() From worker 90: ====================================================================================== Information request received. A stacktrace will print followed by a 1.0 second profile ====================================================================================== cmd: /opt/julia/bin/julia 17 running 0 of 2 signal (10): User defined signal 1 unknown function (ip: 0x7a8e1cd52f14) at /lib/x86_64-linux-gnu/libc.so.6 pthread_cond_wait at /lib/x86_64-linux-gnu/libc.so.6 (unknown line) uv_cond_wait at /workspace/srcdir/libuv/src/unix/thread.c:822 ijl_task_get_next at /source/src/scheduler.c:520 wait at ./task.jl:1216 wait_forever at ./task.jl:1150 jfptr_wait_forever_30682.1 at /opt/julia/lib/julia/sys.so (unknown line) _jl_invoke at /source/src/gf.c:3490 [inlined] ijl_apply_generic at /source/src/gf.c:3690 jl_apply at /source/src/julia.h:2336 [inlined] start_task at /source/src/task.c:1249 unknown function (ip: (nil)) at (unknown file) epoll_pwait at /lib/x86_64-linux-gnu/libc.so.6 (unknown line) uv__io_poll at /workspace/srcdir/libuv/src/unix/linux.c:1404 uv_run at /workspace/srcdir/libuv/src/unix/core.c:430 ijl_task_get_next at /source/src/scheduler.c:454 wait at ./task.jl:1216 wait_forever at ./task.jl:1150 jfptr_wait_forever_30682.1 at /opt/julia/lib/julia/sys.so (unknown line) _jl_invoke at /source/src/gf.c:3490 [inlined] ijl_apply_generic at /source/src/gf.c:3690 jl_apply at /source/src/julia.h:2336 [inlined] start_task at /source/src/task.c:1249 unknown function (ip: (nil)) at (unknown file) ============================================================== Profile collected. A report will print at the next yield point ============================================================== From worker 88: ┌ Warning: There were no samples collected in one or more groups. From worker 88: │ This may be due to idle threads, or you may need to run your From worker 88: │ program longer (perhaps by running it multiple times), From worker 88: │ or adjust the delay between samples with `Profile.init()`. From worker 88: └ @ Profile /opt/julia/share/julia/stdlib/v1.13/Profile/src/Profile.jl:1362 From worker 88: Overhead ╎ [+additional indent] Count File:Line Function From worker 88: ========================================================= From worker 88: Thread 1 (interactive) Task 0x0000712f10ea1e10 Total snapshots: 1. Utilization: 0% From worker 88: ╎1 @Base/task.jl:1150 wait_forever() From worker 88: ╎ 1 @Base/task.jl:1216 wait() From worker 88: From worker 88: Thread 2 (default) Task 0x0000712f127fc100 Total snapshots: 1. Utilization: 0% From worker 88: ╎1 @Base/task.jl:1150 wait_forever() From worker 88: ╎ 1 @Base/task.jl:1216 wait() From worker 88: From worker 89: ┌ Warning: There were no samples collected in one or more groups. From worker 89: │ This may be due to idle threads, or you may need to run your From worker 89: │ program longer (perhaps by running it multiple times), From worker 89: │ or adjust the delay between samples with `Profile.init()`. From worker 89: └ @ Profile /opt/julia/share/julia/stdlib/v1.13/Profile/src/Profile.jl:1362 From worker 89: Overhead ╎ [+additional indent] Count File:Line Function From worker 89: ========================================================= From worker 89: Thread 1 (interactive) Task 0x00007c5c3b1646a0 Total snapshots: 1. Utilization: 100% From worker 89: From worker 89: Thread 2 (default) Task 0x00007c5c3bdfc100 Total snapshots: 1. Utilization: 0% From worker 89: ╎1 @Base/task.jl:1150 wait_forever() From worker 89: ╎ 1 @Base/task.jl:1216 wait() From worker 89: ====================================================================================== Information request received. A stacktrace will print followed by a 1.0 second profile ====================================================================================== cmd: /opt/julia/bin/julia 1 running 0 of 2 signal (10): User defined signal 1 unknown function (ip: 0x71d2f90d4f14) at /lib/x86_64-linux-gnu/libc.so.6 pthread_cond_wait at /lib/x86_64-linux-gnu/libc.so.6 (unknown line) uv_cond_wait at /workspace/srcdir/libuv/src/unix/thread.c:822 ijl_task_get_next at /source/src/scheduler.c:520 wait at ./task.jl:1216 wait_forever at ./task.jl:1150 jfptr_wait_forever_30682.1 at /opt/julia/lib/julia/sys.so (unknown line) _jl_invoke at /source/src/gf.c:3490 [inlined] ijl_apply_generic at /source/src/gf.c:3690 jl_apply at /source/src/julia.h:2336 [inlined] start_task at /source/src/task.c:1249 unknown function (ip: (nil)) at (unknown file) epoll_pwait at /lib/x86_64-linux-gnu/libc.so.6 (unknown line) uv__io_poll at /workspace/srcdir/libuv/src/unix/linux.c:1404 uv_run at /workspace/srcdir/libuv/src/unix/core.c:430 ijl_task_get_next at /source/src/scheduler.c:454 wait at ./task.jl:1216 wait_forever at ./task.jl:1150 [ Info: running task 43 on process 90 (Schedulers-primary-qO2tilUH); 5 workers total; 67 tasks left in task-pool. jfptr_wait_forever_30682.1 at /opt/julia/lib/julia/sys.so (unknown line) _jl_invoke at /source/src/gf.c:3490 [inlined] ijl_apply_generic at /source/src/gf.c:3690 jl_apply at /source/src/julia.h:2336 [inlined] start_task at /source/src/task.c:1249 unknown function (ip: (nil)) at (unknown file) ============================================================== Profile collected. A report will print at the next yield point ============================================================== [ Info: running task 44 on process 89 (Schedulers-primary-qO2tilUH); 5 workers total; 66 tasks left in task-pool. From worker 88: From worker 88: [452] signal 15: Terminated From worker 88: in expression starting at none:0 From worker 88: unknown function (ip: 0x712f292baf14) at /lib/x86_64-linux-gnu/libc.so.6 From worker 88: pthread_cond_wait at /lib/x86_64-linux-gnu/libc.so.6 (unknown line) From worker 88: uv_cond_wait at /workspace/srcdir/libuv/src/unix/thread.c:822 From worker 88: ijl_task_get_next at /source/src/scheduler.c:520 [1] signal 15: Terminated in expression starting at /PkgEval.jl/scripts/evaluate.jl:210 From worker 90: From worker 90: [459] signal 15: Terminated From worker 90: in expression starting at none:0 From worker 89: From worker 89: [457] signal 15: Terminated From worker 89: in expression starting at none:0 From worker 92: From worker 92: [469] signal 15: Terminated From worker 92: in expression starting at none:0 From worker 92: unknown function (ip: 0x7a8f3732af14) at /lib/x86_64-linux-gnu/libc.so.6 From worker 92: pthread_cond_wait at /lib/x86_64-linux-gnu/libc.so.6 (unknown line) From worker 92: uv_cond_wait at /workspace/srcdir/libuv/src/unix/thread.c:822 From worker 92: ijl_task_get_next at /source/src/scheduler.c:520 From worker 91: From worker 91: [467] signal 15: Terminated From worker 91: in expression starting at none:0 From worker 91: unknown function (ip: 0x72ef68d74f14) at /lib/x86_64-linux-gnu/libc.so.6 From worker 91: pthread_cond_wait at /lib/x86_64-linux-gnu/libc.so.6 (unknown line) From worker 91: uv_cond_wait at /workspace/srcdir/libuv/src/unix/thread.c:822 From worker 91: ijl_task_get_next at /source/src/scheduler.c:520 [17] signal 15: Terminated in expression starting at /home/pkgeval/.julia/packages/Schedulers/JBB7J/test/runtests.jl:433 _ZL28formLCSSAForInstructionsImplRN4llvm15SmallVectorImplIPNS_11InstructionEEERKNS_13DominatorTreeERKNS_8LoopInfoEPNS_15ScalarEvolutionEPNS0_IPNS_7PHINodeEEESG_RNS_13SmallDenseMapIPNS_4LoopENS_11SmallVectorIPNS_10BasicBlockELj1EEELj4ENS_12DenseMapInfoISJ_vEENS_6detail12DenseMapPairISJ_SN_EEEE at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) _ZN4llvm12DenseMapBaseINS_8DenseMapIPNS_13slpvectorizer7BoUpSLP9TreeEntryENS_11SmallVectorISt4pairIjS5_ELj3EEENS_12DenseMapInfoIS5_vEENS_6detail12DenseMapPairIS5_S9_EEEES5_S9_SB_SE_E15LookupBucketForIS5_EEbRKT_RPSE_ at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) _ZN4llvm13slpvectorizer7BoUpSLP18reorderBottomToTopEb at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) _ZN4llvm17SLPVectorizerPass18tryToVectorizeListENS_8ArrayRefIPNS_5ValueEEERNS_13slpvectorizer7BoUpSLPEb at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) _ZL13formLCSSAImplRN4llvm4LoopERKNS_13DominatorTreeEPKNS_8LoopInfoEPNS_15ScalarEvolutionERNS_13SmallDenseMapIPS0_NS_11SmallVectorIPNS_10BasicBlockELj1EEELj4ENS_12DenseMapInfoISB_vEENS_6detail12DenseMapPairISB_SF_EEEE at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) _Z22tryToVectorizeSequenceIN4llvm5ValueEEbRNS0_15SmallVectorImplIPT_EENS0_12function_refIFbS4_S4_EEES9_NS7_IFbNS0_8ArrayRefIS4_EEbEEEbRNS0_13slpvectorizer7BoUpSLPE.constprop.0 at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) _ZN4llvm17SLPVectorizerPass22vectorizeChainsInBlockEPNS_10BasicBlockERNS_13slpvectorizer7BoUpSLPE at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) _ZL24formLCSSARecursivelyImplRN4llvm4LoopERKNS_13DominatorTreeEPKNS_8LoopInfoEPNS_15ScalarEvolutionERNS_13SmallDenseMapIPS0_NS_11SmallVectorIPNS_10BasicBlockELj1EEELj4ENS_12DenseMapInfoISB_vEENS_6detail12DenseMapPairISB_SF_EEEE at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) _ZN4llvm9LCSSAPass3runERNS_8FunctionERNS_15AnalysisManagerIS1_JEEE at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) ┌ Warning: There were no samples collected in one or more groups. │ This may be due to idle threads, or you may need to run your │ program longer (perhaps by running it multiple times), │ or adjust the delay between samples with `Profile.init()`. └ @ Profile /opt/julia/share/julia/stdlib/v1.13/Profile/src/Profile.jl:1362 _ZN4llvm17SLPVectorizerPass7runImplERNS_8FunctionEPNS_15ScalarEvolutionEPNS_19TargetTransformInfoEPNS_17TargetLibraryInfoEPNS_9AAResultsEPNS_8LoopInfoEPNS_13DominatorTreeEPNS_15AssumptionCacheEPNS_12DemandedBitsEPNS_25OptimizationRemarkEmitterE.part.0 at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) _ZN4llvm17SLPVectorizerPass3runERNS_8FunctionERNS_15AnalysisManagerIS1_JEEE at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) run at /source/usr/include/llvm/IR/PassManagerInternal.h:91 _ZN4llvm11PassManagerINS_8FunctionENS_15AnalysisManagerIS1_JEEEJEE3runERS1_RS3_ at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) _ZN4llvm25FunctionToLoopPassAdaptor3runERNS_8FunctionERNS_15AnalysisManagerIS1_JEEE at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) run at /source/usr/include/llvm/IR/PassManagerInternal.h:91 _ZN4llvm11PassManagerINS_8FunctionENS_15AnalysisManagerIS1_JEEEJEE3runERS1_RS3_ at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) run at /source/usr/include/llvm/IR/PassManagerInternal.h:91 _ZN4llvm27ModuleToFunctionPassAdaptor3runERNS_6ModuleERNS_15AnalysisManagerIS1_JEEE at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) run at /source/usr/include/llvm/IR/PassManagerInternal.h:91 _ZN4llvm11PassManagerINS_6ModuleENS_15AnalysisManagerIS1_JEEEJEE3runERS1_RS3_ at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) run at /source/src/pipeline.cpp:791 run at /source/usr/include/llvm/IR/PassManagerInternal.h:91 _ZN4llvm11PassManagerINS_8FunctionENS_15AnalysisManagerIS1_JEEEJEE3runERS1_RS3_ at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) run at /source/usr/include/llvm/IR/PassManagerInternal.h:91 _ZN4llvm27ModuleToFunctionPassAdaptor3runERNS_6ModuleERNS_15AnalysisManagerIS1_JEEE at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) run at /source/usr/include/llvm/IR/PassManagerInternal.h:91 _ZN4llvm11PassManagerINS_6ModuleENS_15AnalysisManagerIS1_JEEEJEE3runERS1_RS3_ at /opt/julia/bin/../lib/julia/libLLVM.so.20.1jl (unknown line) run at /source/src/pipeline.cpp:791 operator() at /source/src/jitlayers.cpp:1475 withModuleDo<(anonymous namespace)::sizedOptimizerT::operator()(llvm::orc::ThreadSafeModule) [with long unsigned int N = 4]:: > at /source/usr/include/llvm/ExecutionEngine/Orc/ThreadSafeModule.h:136 [inlined] operator() at /source/src/jitlayers.cpp:1436 [inlined] operator() at /source/src/jitlayers.cpp:1588 [inlined] addModule at /source/src/jitlayers.cpp:2045 jl_compile_codeinst_now at /source/src/jitlayers.cpp:649 jl_compile_codeinst_impl at /source/src/jitlayers.cpp:840 jl_compile_method_internal at /source/src/gf.c:3004 _jl_invoke at /source/src/gf.c:3482 [inlined] ijl_apply_generic at /source/src/gf.c:3690 jl_apply at /source/src/julia.h:2336 [inlined] jl_f_invoke_in_world at /source/src/builtins.c:896 ansi_write_ at ./strings/annotated_io.jl:216 [inlined] print at ./strings/annotated_io.jl:241 [inlined] print at ./strings/io.jl:46 println at ./strings/io.jl:77 [inlined] print_tree at /source/usr/share/julia/stdlib/v1.13/Profile/src/Profile.jl:1249 tree at /source/usr/share/julia/stdlib/v1.13/Profile/src/Profile.jl:1309 print_group at /source/usr/share/julia/stdlib/v1.13/Profile/src/Profile.jl:389 #print#5 at /source/usr/share/julia/stdlib/v1.13/Profile/src/Profile.jl:331 operator() at /source/src/jitlayers.cpp:1475 print at /source/usr/share/julia/stdlib/v1.13/Profile/src/Profile.jl:271 [inlined] print at /source/usr/share/julia/stdlib/v1.13/Profile/src/Profile.jl:271 [inlined] print at /source/usr/share/julia/stdlib/v1.13/Profile/src/Profile.jl:271 withModuleDo<(anonymous namespace)::sizedOptimizerT::operator()(llvm::orc::ThreadSafeModule) [with long unsigned int N = 4]:: > at /source/usr/include/llvm/ExecutionEngine/Orc/ThreadSafeModule.h:136 [inlined] operator() at /source/src/jitlayers.cpp:1436 [inlined] operator() at /source/src/jitlayers.cpp:1588 [inlined] addModule at /source/src/jitlayers.cpp:2045 jl_compile_codeinst_now at /source/src/jitlayers.cpp:649 jl_compile_codeinst_impl at /source/src/jitlayers.cpp:840 jl_compile_method_internal at /source/src/gf.c:3004 _jl_invoke at /source/src/gf.c:3482 [inlined] ijl_apply_generic at /source/src/gf.c:3690 _peek_report at /source/usr/share/julia/stdlib/v1.13/Profile/src/Profile.jl:94 jfptr__peek_report_4475.1 at /opt/julia/share/julia/compiled/v1.13/Profile/nGhxz_9HAIE.so (unknown line) jl_apply at /source/src/julia.h:2336 [inlined] _jl_invoke at /source/src/gf.c:3490 [inlined] ijl_apply_generic at /source/src/gf.c:3690 jl_apply at /source/src/julia.h:2336 [inlined] jl_f_invokelatest at /source/src/builtins.c:877 jl_f_invoke_in_world at /source/src/builtins.c:896 ansi_write_ at ./strings/annotated_io.jl:216 [inlined] print at ./strings/annotated_io.jl:241 [inlined] print at ./strings/io.jl:46 println at ./strings/io.jl:77 [inlined] print_tree at /source/usr/share/julia/stdlib/v1.13/Profile/src/Profile.jl:1249 tree at /source/usr/share/julia/stdlib/v1.13/Profile/src/Profile.jl:1309 print_group at /source/usr/share/julia/stdlib/v1.13/Profile/src/Profile.jl:389 #print#5 at /source/usr/share/julia/stdlib/v1.13/Profile/src/Profile.jl:331 print at /source/usr/share/julia/stdlib/v1.13/Profile/src/Profile.jl:271 [inlined] print at /source/usr/share/julia/stdlib/v1.13/Profile/src/Profile.jl:271 [inlined] print at /source/usr/share/julia/stdlib/v1.13/Profile/src/Profile.jl:271 _peek_report at /source/usr/share/julia/stdlib/v1.13/Profile/src/Profile.jl:94 jfptr__peek_report_4438.1 at /opt/julia/share/julia/compiled/v1.13/Profile/nGhxz_X6ROR.so (unknown line) _jl_invoke at /source/src/gf.c:3490 [inlined] ijl_apply_generic at /source/src/gf.c:3690 jl_apply at /source/src/julia.h:2336 [inlined] jl_f_invokelatest at /source/src/builtins.c:877 ┌ Warning: There were no samples collected in one or more groups. │ This may be due to idle threads, or you may need to run your │ program longer (perhaps by running it multiple times), │ or adjust the delay between samples with `Profile.init()`. └ @ Profile /opt/julia/share/julia/stdlib/v1.13/Profile/src/Profile.jl:1362 profile_printing_listener at ./Base.jl:337 profile_printing_listener at ./Base.jl:337 #start_profile_listener##0 at ./Base.jl:355 #start_profile_listener##0 at ./Base.jl:355 jfptr_YY.start_profile_listenerYY.YY.0_16770.1 at /opt/julia/lib/julia/sys.so (unknown line) _jl_invoke at /source/src/gf.c:3490 [inlined] ijl_apply_generic at /source/src/gf.c:3690 jl_apply at /source/src/julia.h:2336 [inlined] start_task at /source/src/task.c:1249 unknown function (ip: (nil)) at (unknown file) epoll_pwait at /lib/x86_64-linux-gnu/libc.so.6 (unknown line) uv__io_poll at /workspace/srcdir/libuv/src/unix/linux.c:1404 uv_run at /workspace/srcdir/libuv/src/unix/core.c:430 ijl_task_get_next at /source/src/scheduler.c:454 jfptr_YY.start_profile_listenerYY.YY.0_16770.1 at /opt/julia/lib/julia/sys.so (unknown line) _jl_invoke at /source/src/gf.c:3490 [inlined] ijl_apply_generic at /source/src/gf.c:3690 jl_apply at /source/src/julia.h:2336 [inlined] start_task at /source/src/task.c:1249 unknown function (ip: (nil)) at (unknown file) epoll_pwait at /lib/x86_64-linux-gnu/libc.so.6 (unknown line) uv__io_poll at /workspace/srcdir/libuv/src/unix/linux.c:1404 uv_run at /workspace/srcdir/libuv/src/unix/core.c:430 ijl_task_get_next at /source/src/scheduler.c:454 wait at ./task.jl:1216 wait at ./task.jl:1216 wait_forever at ./task.jl:1150 wait_forever at ./task.jl:1150 jfptr_wait_forever_30682.1 at /opt/julia/lib/julia/sys.so (unknown line) _jl_invoke at /source/src/gf.c:3490 [inlined] ijl_apply_generic at /source/src/gf.c:3690 jl_apply at /source/src/julia.h:2336 [inlined] start_task at /source/src/task.c:1249 unknown function (ip: (nil)) at (unknown file) Allocations: 52264406 (Pool: 52263412; Big: 994); GC: 60 jfptr_wait_forever_30682.1 at /opt/julia/lib/julia/sys.so (unknown line) _jl_invoke at /source/src/gf.c:3490 [inlined] ijl_apply_generic at /source/src/gf.c:3690 jl_apply at /source/src/julia.h:2336 [inlined] start_task at /source/src/task.c:1249 unknown function (ip: (nil)) at (unknown file) Allocations: 10800088 (Pool: 10799948; Big: 140); GC: 9 Overhead ╎ [+additional indent] Count File:Line Function ========================================================= Thread 1 (interactive) Task 0x000071d2e14641f0 Total snapshots: 1. Utilization: 0% ╎1 @Base/task.jl:1150 wait_forever() ╎ 1 @Base/task.jl:1216 wait() Thread 2 (default) Task 0x000071d2e25fc100 Total snapshots: 1. Utilization: 0% ╎1 @Base/task.jl:1150 wait_forever() ╎ 1 @Base/task.jl:1216 wait() PkgEval terminated after 2760.03s: test duration exceeded the time limit