Package evaluation of GridapPETSc on Julia 1.13.0-DEV.733 (c0ecfe9bcc*) started at 2025-06-12T21:57:55.253 ################################################################################ # Set-up # Installing PkgEval dependencies (TestEnv)... Set-up completed after 8.2s ################################################################################ # Installation # Installing GridapPETSc... Resolving package versions... Installed GridapPETSc ─ v0.5.3 Updating `~/.julia/environments/v1.13/Project.toml` [bcdc36c2] + GridapPETSc v0.5.3 Updating `~/.julia/environments/v1.13/Manifest.toml` [47edcb42] + ADTypes v1.14.0 [621f4979] + AbstractFFTs v1.5.0 [1520ce14] + AbstractTrees v0.4.5 [79e6a3ab] + Adapt v4.3.0 [dce04be8] + ArgCheck v2.5.0 [4fba245c] + ArrayInterface v7.19.0 [4c555306] + ArrayLayouts v1.11.1 [15f4f7f2] + AutoHashEquals v2.2.0 [fbb218c0] + BSON v0.3.9 [8e7c35d0] + BlockArrays v1.6.3 [7a955b69] + CircularArrays v1.4.0 [944b1d66] + CodecZlib v0.7.8 [861a8166] + Combinatorics v1.0.3 [bbf7d656] + CommonSubexpressions v0.3.1 [34da2185] + Compat v4.16.0 [187b0558] + ConstructionBase v1.6.0 [864edb3b] + DataStructures v0.18.22 [163ba53b] + DiffResults v1.1.0 [b552c78f] + DiffRules v1.15.1 [a0c0ee7d] + DifferentiationInterface v0.7.1 [b4f34e82] + Distances v0.10.12 [ffbed154] + DocStringExtensions v0.9.5 [7a1cc6ca] + FFTW v1.9.0 [442a2c76] + FastGaussQuadrature v1.0.2 [5789e2e9] + FileIO v1.17.0 [1a297f60] + FillArrays v1.13.0 [6a86dc24] + FiniteDiff v2.27.0 ⌅ [f6369f11] + ForwardDiff v0.10.38 ⌅ [56d4f2e9] + Gridap v0.18.12 ⌃ [f9701e48] + GridapDistributed v0.4.7 [bcdc36c2] + GridapPETSc v0.5.3 [92d709cd] + IrrationalConstants v0.2.4 [42fd0dbc] + IterativeSolvers v0.9.4 [033835bb] + JLD2 v0.5.13 [692b3bcd] + JLLWrappers v1.7.0 [682c06a0] + JSON v0.21.4 [9c8b4983] + LightXML v0.9.1 [d3d80556] + LineSearches v7.4.0 [2ab3a3ac] + LogExpFunctions v0.3.29 [da04e1cc] + MPI v0.20.22 [3da0fdf6] + MPIPreferences v0.1.11 [1914dd2f] + MacroTools v0.5.16 [d41bc354] + NLSolversBase v7.10.0 [2774e3e8] + NLsolve v4.5.1 [77ba4419] + NaNMath v1.1.3 [b8a86587] + NearestNeighbors v0.4.21 [6fe1bfb0] + OffsetArrays v1.17.0 [bac558e1] + OrderedCollections v1.8.1 [d96e819e] + Parameters v0.12.3 [69de0a69] + Parsers v2.8.3 ⌅ [5a9dfac6] + PartitionedArrays v0.3.4 [eebad327] + PkgVersion v0.3.3 [c74db56a] + PolynomialBases v0.4.22 [aea7be01] + PrecompileTools v1.3.2 [21216c6a] + Preferences v1.4.3 [1fd47b50] + QuadGK v2.11.2 [3cdcf5f2] + RecipesBase v1.3.4 [189a3867] + Reexport v1.2.2 [ae029012] + Requires v1.3.1 [efcf1570] + Setfield v1.1.2 [ce78b400] + SimpleUnPack v1.1.0 [a0a7dd2c] + SparseMatricesCSR v0.6.9 [276daf66] + SpecialFunctions v2.5.1 [90137ffa] + StaticArrays v1.9.13 [1e83bf80] + StaticArraysCore v1.4.3 [10745b16] + Statistics v1.11.1 [82ae8749] + StatsAPI v1.7.1 [3bb67fe8] + TranscodingStreams v0.11.3 [3a884ed6] + UnPack v1.0.2 [4004b06d] + VTKBase v1.0.1 [64499a7a] + WriteVTK v1.21.2 [f5851436] + FFTW_jll v3.3.11+0 [e33a78d0] + Hwloc_jll v2.12.1+0 [1d5cc7b8] + IntelOpenMP_jll v2025.0.4+0 [94ce4f54] + Libiconv_jll v1.18.0+0 [856f044c] + MKL_jll v2025.0.1+1 [7cb0a576] + MPICH_jll v4.3.0+1 [f1f71cc9] + MPItrampoline_jll v5.5.3+0 [9237b28f] + MicrosoftMPI_jll v10.1.4+3 [656ef2d0] + OpenBLAS32_jll v0.3.29+0 [fe0851c0] + OpenMPI_jll v5.0.7+2 [efe28fd5] + OpenSpecFun_jll v0.5.6+0 ⌅ [8fa3689e] + PETSc_jll v3.15.2+0 ⌅ [02c8fc9c] + XML2_jll v2.13.6+1 [1317d2d5] + oneTBB_jll v2022.0.0+0 [0dad84c5] + ArgTools v1.1.2 [56f22d72] + Artifacts v1.11.0 [2a0f44e3] + Base64 v1.11.0 [ade2ca70] + Dates v1.11.0 [8ba89e20] + Distributed v1.11.0 [f43a241f] + Downloads v1.7.0 [7b1f6079] + FileWatching v1.11.0 [9fa8497b] + Future v1.11.0 [b77e0a4c] + InteractiveUtils v1.11.0 [ac6e5ff7] + JuliaSyntaxHighlighting v1.12.0 [4af54fe1] + LazyArtifacts v1.11.0 [b27032c2] + LibCURL v0.6.4 [76f85450] + LibGit2 v1.11.0 [8f399da3] + Libdl v1.11.0 [37e2e46d] + LinearAlgebra v1.12.0 [56ddb016] + Logging v1.11.0 [d6f4376e] + Markdown v1.11.0 [a63ad114] + Mmap v1.11.0 [ca575930] + NetworkOptions v1.3.0 [44cfe95a] + Pkg v1.13.0 [de0858da] + Printf v1.11.0 [9a3f8284] + Random v1.11.0 [ea8e919c] + SHA v0.7.0 [9e88b42a] + Serialization v1.11.0 [6462fe0b] + Sockets v1.11.0 [2f01184e] + SparseArrays v1.12.0 [f489334b] + StyledStrings v1.11.0 [4607b0f0] + SuiteSparse [fa267f1f] + TOML v1.0.3 [a4e569a6] + Tar v1.10.0 [8dfed614] + Test v1.11.0 [cf7118a7] + UUIDs v1.11.0 [4ec0a83e] + Unicode v1.11.0 [e66e0078] + CompilerSupportLibraries_jll v1.3.0+1 [deac9b47] + LibCURL_jll v8.12.1+1 [e37daf67] + LibGit2_jll v1.9.0+0 [29816b5a] + LibSSH2_jll v1.11.3+1 [14a3606d] + MozillaCACerts_jll v2025.5.20 [4536629a] + OpenBLAS_jll v0.3.29+0 [05823500] + OpenLibm_jll v0.8.5+0 [458c3c95] + OpenSSL_jll v3.5.0+0 [bea87d4a] + SuiteSparse_jll v7.10.1+0 [83775a58] + Zlib_jll v1.3.1+2 [8e850b90] + libblastrampoline_jll v5.12.0+0 [8e850ede] + nghttp2_jll v1.65.0+0 [3f19e933] + p7zip_jll v17.5.0+2 Info Packages marked with ⌃ and ⌅ have new versions available. Those with ⌃ may be upgradable, but those with ⌅ are restricted by compatibility constraints from upgrading. To see why use `status --outdated -m` Building GridapPETSc → `~/.julia/scratchspaces/44cfe95a-1eb2-52ea-b672-e2afdf69b78f/601899a547ac1f473be900e6b3157e62cb2db8fc/build.log` Installation completed after 24.54s ################################################################################ # Precompilation # Precompiling PkgEval dependencies... Precompiling package dependencies... Precompilation completed after 104.95s ################################################################################ # Testing # Testing GridapPETSc Status `/tmp/jl_2zmb76/Project.toml` ⌅ [56d4f2e9] Gridap v0.18.12 ⌃ [f9701e48] GridapDistributed v0.4.7 [bcdc36c2] GridapPETSc v0.5.3 [da04e1cc] MPI v0.20.22 ⌅ [5a9dfac6] PartitionedArrays v0.3.4 [a0a7dd2c] SparseMatricesCSR v0.6.9 ⌅ [8fa3689e] PETSc_jll v3.15.2+0 [8f399da3] Libdl v1.11.0 [37e2e46d] LinearAlgebra v1.12.0 [9a3f8284] Random v1.11.0 [2f01184e] SparseArrays v1.12.0 [8dfed614] Test v1.11.0 Status `/tmp/jl_2zmb76/Manifest.toml` [47edcb42] ADTypes v1.14.0 [621f4979] AbstractFFTs v1.5.0 [1520ce14] AbstractTrees v0.4.5 [79e6a3ab] Adapt v4.3.0 [dce04be8] ArgCheck v2.5.0 [4fba245c] ArrayInterface v7.19.0 [4c555306] ArrayLayouts v1.11.1 [15f4f7f2] AutoHashEquals v2.2.0 [fbb218c0] BSON v0.3.9 [8e7c35d0] BlockArrays v1.6.3 [7a955b69] CircularArrays v1.4.0 [944b1d66] CodecZlib v0.7.8 [861a8166] Combinatorics v1.0.3 [bbf7d656] CommonSubexpressions v0.3.1 [34da2185] Compat v4.16.0 [187b0558] ConstructionBase v1.6.0 [864edb3b] DataStructures v0.18.22 [163ba53b] DiffResults v1.1.0 [b552c78f] DiffRules v1.15.1 [a0c0ee7d] DifferentiationInterface v0.7.1 [b4f34e82] Distances v0.10.12 [ffbed154] DocStringExtensions v0.9.5 [7a1cc6ca] FFTW v1.9.0 [442a2c76] FastGaussQuadrature v1.0.2 [5789e2e9] FileIO v1.17.0 [1a297f60] FillArrays v1.13.0 [6a86dc24] FiniteDiff v2.27.0 ⌅ [f6369f11] ForwardDiff v0.10.38 ⌅ [56d4f2e9] Gridap v0.18.12 ⌃ [f9701e48] GridapDistributed v0.4.7 [bcdc36c2] GridapPETSc v0.5.3 [92d709cd] IrrationalConstants v0.2.4 [42fd0dbc] IterativeSolvers v0.9.4 [033835bb] JLD2 v0.5.13 [692b3bcd] JLLWrappers v1.7.0 [682c06a0] JSON v0.21.4 [9c8b4983] LightXML v0.9.1 [d3d80556] LineSearches v7.4.0 [2ab3a3ac] LogExpFunctions v0.3.29 [da04e1cc] MPI v0.20.22 [3da0fdf6] MPIPreferences v0.1.11 [1914dd2f] MacroTools v0.5.16 [d41bc354] NLSolversBase v7.10.0 [2774e3e8] NLsolve v4.5.1 [77ba4419] NaNMath v1.1.3 [b8a86587] NearestNeighbors v0.4.21 [6fe1bfb0] OffsetArrays v1.17.0 [bac558e1] OrderedCollections v1.8.1 [d96e819e] Parameters v0.12.3 [69de0a69] Parsers v2.8.3 ⌅ [5a9dfac6] PartitionedArrays v0.3.4 [eebad327] PkgVersion v0.3.3 [c74db56a] PolynomialBases v0.4.22 [aea7be01] PrecompileTools v1.3.2 [21216c6a] Preferences v1.4.3 [1fd47b50] QuadGK v2.11.2 [3cdcf5f2] RecipesBase v1.3.4 [189a3867] Reexport v1.2.2 [ae029012] Requires v1.3.1 [efcf1570] Setfield v1.1.2 [ce78b400] SimpleUnPack v1.1.0 [a0a7dd2c] SparseMatricesCSR v0.6.9 [276daf66] SpecialFunctions v2.5.1 [90137ffa] StaticArrays v1.9.13 [1e83bf80] StaticArraysCore v1.4.3 [10745b16] Statistics v1.11.1 [82ae8749] StatsAPI v1.7.1 [3bb67fe8] TranscodingStreams v0.11.3 [3a884ed6] UnPack v1.0.2 [4004b06d] VTKBase v1.0.1 [64499a7a] WriteVTK v1.21.2 [f5851436] FFTW_jll v3.3.11+0 [e33a78d0] Hwloc_jll v2.12.1+0 [1d5cc7b8] IntelOpenMP_jll v2025.0.4+0 [94ce4f54] Libiconv_jll v1.18.0+0 [856f044c] MKL_jll v2025.0.1+1 [7cb0a576] MPICH_jll v4.3.0+1 [f1f71cc9] MPItrampoline_jll v5.5.3+0 [9237b28f] MicrosoftMPI_jll v10.1.4+3 [656ef2d0] OpenBLAS32_jll v0.3.29+0 [fe0851c0] OpenMPI_jll v5.0.7+2 [efe28fd5] OpenSpecFun_jll v0.5.6+0 ⌅ [8fa3689e] PETSc_jll v3.15.2+0 ⌅ [02c8fc9c] XML2_jll v2.13.6+1 [1317d2d5] oneTBB_jll v2022.0.0+0 [0dad84c5] ArgTools v1.1.2 [56f22d72] Artifacts v1.11.0 [2a0f44e3] Base64 v1.11.0 [ade2ca70] Dates v1.11.0 [8ba89e20] Distributed v1.11.0 [f43a241f] Downloads v1.7.0 [7b1f6079] FileWatching v1.11.0 [9fa8497b] Future v1.11.0 [b77e0a4c] InteractiveUtils v1.11.0 [ac6e5ff7] JuliaSyntaxHighlighting v1.12.0 [4af54fe1] LazyArtifacts v1.11.0 [b27032c2] LibCURL v0.6.4 [76f85450] LibGit2 v1.11.0 [8f399da3] Libdl v1.11.0 [37e2e46d] LinearAlgebra v1.12.0 [56ddb016] Logging v1.11.0 [d6f4376e] Markdown v1.11.0 [a63ad114] Mmap v1.11.0 [ca575930] NetworkOptions v1.3.0 [44cfe95a] Pkg v1.13.0 [de0858da] Printf v1.11.0 [9a3f8284] Random v1.11.0 [ea8e919c] SHA v0.7.0 [9e88b42a] Serialization v1.11.0 [6462fe0b] Sockets v1.11.0 [2f01184e] SparseArrays v1.12.0 [f489334b] StyledStrings v1.11.0 [4607b0f0] SuiteSparse [fa267f1f] TOML v1.0.3 [a4e569a6] Tar v1.10.0 [8dfed614] Test v1.11.0 [cf7118a7] UUIDs v1.11.0 [4ec0a83e] Unicode v1.11.0 [e66e0078] CompilerSupportLibraries_jll v1.3.0+1 [deac9b47] LibCURL_jll v8.12.1+1 [e37daf67] LibGit2_jll v1.9.0+0 [29816b5a] LibSSH2_jll v1.11.3+1 [14a3606d] MozillaCACerts_jll v2025.5.20 [4536629a] OpenBLAS_jll v0.3.29+0 [05823500] OpenLibm_jll v0.8.5+0 [458c3c95] OpenSSL_jll v3.5.0+0 [bea87d4a] SuiteSparse_jll v7.10.1+0 [83775a58] Zlib_jll v1.3.1+2 [8e850b90] libblastrampoline_jll v5.12.0+0 [8e850ede] nghttp2_jll v1.65.0+0 [3f19e933] p7zip_jll v17.5.0+2 Info Packages marked with ⌃ and ⌅ have new versions available. Those with ⌃ may be upgradable, but those with ⌅ are restricted by compatibility constraints from upgrading. Testing Running tests... [0] PetscDetermineInitialFPTrap(): Floating point trapping is off by default 0 [0] PetscInitialize(): PETSc successfully started: number of processors = 1 [0] PetscGetHostName(): Rejecting domainname, likely is NIS GridapPETSc-primary-0D8uaB42.(none) [0] PetscInitialize(): Running on machine: GridapPETSc-primary-0D8uaB42 [0] PetscCommDuplicate(): Duplicating a communicator 1140850689 -2080374784 max tags = 1073741823 [0] PetscCommDuplicate(): Using internal PETSc communicator 1140850689 -2080374784 [0] PetscGetHostName(): Rejecting domainname, likely is NIS GridapPETSc-primary-0D8uaB42.(none) Vec Object: 1 MPI processes type: seq 1. 2. 4. 1. [0] PetscCommDuplicate(): Duplicating a communicator 1140850688 -2080374783 max tags = 1073741823 [0] PetscGetHostName(): Rejecting domainname, likely is NIS GridapPETSc-primary-0D8uaB42.(none) Vec Object: 1 MPI processes type: seq 1. 2. 4. 1. Vec Object: 1 MPI processes type: seq 20. 40. 4. 60. [0] PetscCommDuplicate(): Using internal PETSc communicator 1140850689 -2080374784 [0] MatAssemblyEnd_SeqAIJ(): Matrix size: 4 X 5; storage space: 10 unneeded,10 used [0] MatAssemblyEnd_SeqAIJ(): Number of mallocs during MatSetValues() is 0 [0] MatAssemblyEnd_SeqAIJ(): Maximum nonzeros in any row is 3 [0] MatCheckCompressedRow(): Found the ratio (num_zerorows 0)/(num_localrows 4) < 0.6. Do not use CompressedRow routines. [0] MatSeqAIJCheckInode(): Found 3 nodes of 4. Limit used: 5. Using Inode routines Mat Object: 1 MPI processes type: seqaij row 0: (1, 2.) (3, 3.) (4, 1.) row 1: (1, 6.) (3, 11.) (4, 5.) row 2: (1, 4.) (3, 3.) row 3: (3, 4.) (4, 3.) [0] PetscCommDuplicate(): Using internal PETSc communicator 1140850689 -2080374784 [0] MatAssemblyEnd_SeqAIJ(): Matrix size: 4 X 4; storage space: 0 unneeded,10 used [0] MatAssemblyEnd_SeqAIJ(): Number of mallocs during MatSetValues() is 0 [0] MatAssemblyEnd_SeqAIJ(): Maximum nonzeros in any row is 3 [0] MatCheckCompressedRow(): Found the ratio (num_zerorows 0)/(num_localrows 4) < 0.6. Do not use CompressedRow routines. [0] MatSeqAIJCheckInode(): Found 4 nodes out of 4 rows. Not using Inode routines Mat Object: 1 MPI processes type: seqaij row 0: (0, 4.) (1, -2.) row 1: (0, -1.) (1, 6.) (2, -2.) row 2: (1, -1.) (2, 6.) (3, -2.) row 3: (2, -1.) (3, 4.) [0] PetscCommDuplicate(): Using internal PETSc communicator 1140850689 -2080374784 [0] PetscCommDuplicate(): Using internal PETSc communicator 1140850689 -2080374784 [0] PetscCommDuplicate(): Using internal PETSc communicator 1140850689 -2080374784 [0] PCSetUp(): Setting up PC for first time [0] PCSetUp(): Leaving PC with identical preconditioner since operator is unchanged [0] PCSetUp(): Leaving PC with identical preconditioner since operator is unchanged [0] PCSetUp(): Leaving PC with identical preconditioner since operator is unchanged [0] PCSetUp(): Leaving PC with identical preconditioner since operator is unchanged [0] PCSetUp(): Leaving PC with identical preconditioner since operator is unchanged [0] PCSetUp(): Leaving PC with identical preconditioner since operator is unchanged [0] PCSetUp(): Leaving PC with identical preconditioner since operator is unchanged [0] PCSetUp(): Leaving PC with identical preconditioner since operator is unchanged [0] PCSetUp(): Leaving PC with identical preconditioner since operator is unchanged [0] KSPConvergedDefault(): Linear solver has converged. Residual norm 1.260351938337e-16 is less than relative tolerance 1.000000000000e-05 times initial right hand side norm 1.145643923739e+00 at iteration 4 [0] PCSetUp(): Leaving PC with identical preconditioner since operator is unchanged [0] PCSetUp(): Leaving PC with identical preconditioner since operator is unchanged [0] PCSetUp(): Leaving PC with identical preconditioner since operator is unchanged [0] PCSetUp(): Leaving PC with identical preconditioner since operator is unchanged [0] PCSetUp(): Leaving PC with identical preconditioner since operator is unchanged [0] PCSetUp(): Leaving PC with identical preconditioner since operator is unchanged [0] PCSetUp(): Leaving PC with identical preconditioner since operator is unchanged [0] PCSetUp(): Leaving PC with identical preconditioner since operator is unchanged [0] PCSetUp(): Leaving PC with identical preconditioner since operator is unchanged [0] PCSetUp(): Leaving PC with identical preconditioner since operator is unchanged [0] KSPConvergedDefault(): Linear solver has converged. Residual norm 1.260351938337e-16 is less than relative tolerance 1.000000000000e-05 times initial right hand side norm 1.145643923739e+00 at iteration 4 [0] PCSetUp(): Leaving PC with identical preconditioner since operator is unchanged [0] PCSetUp(): Leaving PC with identical preconditioner since operator is unchanged [0] PCSetUp(): Leaving PC with identical preconditioner since operator is unchanged [0] PCSetUp(): Leaving PC with identical preconditioner since operator is unchanged [0] PCSetUp(): Leaving PC with identical preconditioner since operator is unchanged [0] PCSetUp(): Leaving PC with identical preconditioner since operator is unchanged [0] PCSetUp(): Leaving PC with identical preconditioner since operator is unchanged [0] PCSetUp(): Leaving PC with identical preconditioner since operator is unchanged [0] PCSetUp(): Leaving PC with identical preconditioner since operator is unchanged [0] KSPConvergedDefault(): Linear solver has converged. Residual norm 1.315348017250e-16 is less than relative tolerance 1.000000000000e-05 times initial right hand side norm 1.145643923739e+00 at iteration 4 KSP Object: (p_) 1 MPI processes type: gmres restart=30, using Classical (unmodified) Gram-Schmidt Orthogonalization with no iterative refinement happy breakdown tolerance 1e-30 maximum iterations=10000, initial guess is zero tolerances: relative=1e-05, absolute=1e-50, divergence=10000. left preconditioning using PRECONDITIONED norm type for convergence test PC Object: (p_) 1 MPI processes type: jacobi type DIAGONAL linear system matrix = precond matrix: Mat Object: 1 MPI processes type: seqaij rows=4, cols=4 total: nonzeros=10, allocated nonzeros=0 total number of mallocs used during MatSetValues calls=0 not using I-node routines [0] PetscFinalize(): PetscFinalize() called [0] Petsc_DelViewer(): Removing viewer data attribute in an MPI_Comm -2080374784 [0] Petsc_OuterComm_Attr_Delete_Fn(): Removing reference to PETSc communicator embedded in a user MPI_Comm -2080374784 [0] Petsc_InnerComm_Attr_Delete_Fn(): User MPI_Comm 1140850689 is being unlinked from inner PETSc comm -2080374784 [0] PetscCommDestroy(): Deleting PETSc MPI_Comm -2080374784 [0] Petsc_Counter_Attr_Delete_Fn(): Deleting counter data in an MPI_Comm -2080374784 [0] Petsc_DelViewer(): Removing viewer data attribute in an MPI_Comm -2080374783 [0] Petsc_OuterComm_Attr_Delete_Fn(): Removing reference to PETSc communicator embedded in a user MPI_Comm -2080374783 [0] Petsc_InnerComm_Attr_Delete_Fn(): User MPI_Comm 1140850688 is being unlinked from inner PETSc comm -2080374783 [0] PetscCommDestroy(): Deleting PETSc MPI_Comm -2080374783 [0] Petsc_Counter_Attr_Delete_Fn(): Deleting counter data in an MPI_Comm -2080374783 23.835427 seconds (7.83 M allocations: 457.483 MiB, 3.30% gc time, 53.55% compilation time: 49% of which was recompilation) [0] PetscDetermineInitialFPTrap(): Floating point trapping is off by default 0 [0] PetscInitialize(): PETSc successfully started: number of processors = 1 [0] PetscGetHostName(): Rejecting domainname, likely is NIS GridapPETSc-primary-0D8uaB42.(none) [0] PetscInitialize(): Running on machine: GridapPETSc-primary-0D8uaB42 [0] PetscCommDuplicate(): Duplicating a communicator 1140850689 -2080374783 max tags = 1073741823 [0] PetscCommDuplicate(): Using internal PETSc communicator 1140850689 -2080374783 [0] PetscCommDuplicate(): Using internal PETSc communicator 1140850689 -2080374783 [0] PetscCommDuplicate(): Using internal PETSc communicator 1140850689 -2080374783 [0] PetscCommDuplicate(): Using internal PETSc communicator 1140850689 -2080374783 [0] PetscCommDuplicate(): Using internal PETSc communicator 1140850689 -2080374783 [0] MatAssemblyEnd_SeqAIJ(): Matrix size: 4 X 5; storage space: 20 unneeded,0 used [0] MatAssemblyEnd_SeqAIJ(): Number of mallocs during MatSetValues() is 0 [0] MatAssemblyEnd_SeqAIJ(): Maximum nonzeros in any row is 0 [0] MatCheckCompressedRow(): Found the ratio (num_zerorows 4)/(num_localrows 4) > 0.6. Use CompressedRow routines. [0] MatSeqAIJCheckInode(): Found 1 nodes of 4. Limit used: 5. Using Inode routines [0] MatAssemblyEnd_SeqAIJ(): Matrix size: 4 X 5; storage space: 14 unneeded,1 used [0] MatAssemblyEnd_SeqAIJ(): Number of mallocs during MatSetValues() is 1 [0] MatAssemblyEnd_SeqAIJ(): Maximum nonzeros in any row is 1 [0] MatCheckCompressedRow(): Found the ratio (num_zerorows 3)/(num_localrows 4) > 0.6. Use CompressedRow routines. [0] MatSeqAIJCheckInode(): Found 2 nodes of 4. Limit used: 5. Using Inode routines [0] MatAssemblyEnd_SeqAIJ(): Matrix size: 4 X 5; storage space: 14 unneeded,2 used [0] MatAssemblyEnd_SeqAIJ(): Number of mallocs during MatSetValues() is 1 [0] MatAssemblyEnd_SeqAIJ(): Maximum nonzeros in any row is 1 [0] MatCheckCompressedRow(): Found the ratio (num_zerorows 2)/(num_localrows 4) < 0.6. Do not use CompressedRow routines. [0] MatSeqAIJCheckInode(): Found 4 nodes out of 4 rows. Not using Inode routines 4×5 GridapPETSc.PETScMatrix: 0.0 0.0 5.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 7.0 0.0 0.0 0.0 0.0 0.0 [0] PetscCommDuplicate(): Using internal PETSc communicator 1140850689 -2080374783 [0] MatAssemblyEnd_SeqAIJ(): Matrix size: 3 X 2; storage space: 6 unneeded,0 used [0] MatAssemblyEnd_SeqAIJ(): Number of mallocs during MatSetValues() is 0 [0] MatAssemblyEnd_SeqAIJ(): Maximum nonzeros in any row is 0 [0] MatCheckCompressedRow(): Found the ratio (num_zerorows 3)/(num_localrows 3) > 0.6. Use CompressedRow routines. [0] MatSeqAIJCheckInode(): Found 1 nodes of 3. Limit used: 5. Using Inode routines [0] PetscCommDuplicate(): Using internal PETSc communicator 1140850689 -2080374783 [0] MatAssemblyEnd_SeqAIJ(): Matrix size: 4 X 4; storage space: 0 unneeded,10 used [0] MatAssemblyEnd_SeqAIJ(): Number of mallocs during MatSetValues() is 0 [0] MatAssemblyEnd_SeqAIJ(): Maximum nonzeros in any row is 3 [0] MatCheckCompressedRow(): Found the ratio (num_zerorows 0)/(num_localrows 4) < 0.6. Do not use CompressedRow routines. [0] MatSeqAIJCheckInode(): Found 4 nodes out of 4 rows. Not using Inode routines [0] PetscCommDuplicate(): Using internal PETSc communicator 1140850689 -2080374783 [0] PetscCommDuplicate(): Using internal PETSc communicator 1140850689 -2080374783 [0] MatConvert(): Calling duplicate for initial matrix seqaij 0 1 [0] MatConvert(): Calling duplicate for initial matrix seqaij 0 1 [0] PetscCommDuplicate(): Using internal PETSc communicator 1140850689 -2080374783 [0] PetscCommDuplicate(): Using internal PETSc communicator 1140850689 -2080374783 [0] PetscCommDuplicate(): Using internal PETSc communicator 1140850689 -2080374783 [0] PetscCommDuplicate(): Using internal PETSc communicator 1140850689 -2080374783 [0] PetscCommDuplicate(): Using internal PETSc communicator 1140850689 -2080374783 [0] PetscCommDuplicate(): Using internal PETSc communicator 1140850689 -2080374783 [0] PetscCommDuplicate(): Using internal PETSc communicator 1140850689 -2080374783 [0] Petsc_OuterComm_Attr_Delete_Fn(): Removing reference to PETSc communicator embedded in a user MPI_Comm -2080374783 [0] Petsc_InnerComm_Attr_Delete_Fn(): User MPI_Comm 1140850689 is being unlinked from inner PETSc comm -2080374783 [0] PetscCommDestroy(): Deleting PETSc MPI_Comm -2080374783 [0] Petsc_Counter_Attr_Delete_Fn(): Deleting counter data in an MPI_Comm -2080374783 [0] PetscFinalize(): PetscFinalize() called 7.724618 seconds (2.85 M allocations: 153.497 MiB, 7.74% gc time, 92.10% compilation time: 1% of which was recompilation) [1, 2, 3, 5, 6] [4, 5, 7, 3, 6] [6, 7, 4] Vec Object: 1 MPI processes type: seq 10. 20. 30. 40. 50. 60. 70. 3-element PartitionedArrays.DebugArray{SparseArrays.SparseMatrixCSC{Float64, Int64}, 1}: [1] = sparse([1, 2, 3, 1, 3], [1, 2, 3, 4, 5], [9.0, 9.0, 9.0, 1.0, 1.0], 5, 5) [2] = sparse([1, 2, 2, 2, 1], [1, 2, 3, 4, 5], [9.0, 9.0, 1.0, 9.0, 1.0], 5, 5) [3] = sparse([1, 2, 1], [1, 2, 3], [9.0, 9.0, 1.0], 3, 3) Mat Object: 1 MPI processes type: seqaij row 0: (0, 9.) (4, 1.) row 1: (1, 9.) row 2: (2, 9.) (5, 1.) row 3: (3, 9.) (5, 1.) row 4: (2, 9.) (4, 9.) (6, 1.) row 5: (3, 1.) (5, 9.) row 6: (6, 9.) Linear solve converged due to CONVERGED_RTOL iterations 5 Linear solve converged due to CONVERGED_RTOL iterations 5 Vec Object: 1 MPI processes type: seq 140. 180. 330. 420. 790. 580. 630. Vec Object: 1 MPI processes type: seq 10. 20. 30. 40. 50. 60. 70. Linear solve converged due to CONVERGED_RTOL iterations 5 Linear solve converged due to CONVERGED_RTOL iterations 5 Vec Object: 1 MPI processes type: seq 140. 180. 330. 420. 790. 580. 630. Vec Object: 1 MPI processes type: seq 10. 20. 30. 40. 50. 60. 70. 67.184613 seconds (42.10 M allocations: 2.198 GiB, 2.58% gc time, 83.71% compilation time) 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 KSP Object: 1 MPI processes type: gmres restart=30, using Classical (unmodified) Gram-Schmidt Orthogonalization with no iterative refinement happy breakdown tolerance 1e-30 maximum iterations=10000, initial guess is zero tolerances: relative=1e-05, absolute=1e-50, divergence=10000. left preconditioning using DEFAULT norm type for convergence test PC Object: 1 MPI processes type: jacobi PC has not been set up so information may be incomplete type DIAGONAL linear system matrix = precond matrix: Mat Object: 1 MPI processes type: seqaij rows=4, cols=4 total: nonzeros=10, allocated nonzeros=0 total number of mallocs used during MatSetValues calls=0 not using I-node routines 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 8.989370 seconds (4.64 M allocations: 250.646 MiB, 13.67% gc time, 85.62% compilation time) 0 SNES Function norm 3.605551275464e+00 1 SNES Function norm 4.444444444444e-01 2 SNES Function norm 7.111111111111e-02 3 SNES Function norm 3.936947327951e-03 4 SNES Function norm 1.525925473445e-05 5 SNES Function norm 2.328306437081e-10 6 SNES Function norm 0.000000000000e+00 Nonlinear solve converged due to CONVERGED_FNORM_RELATIVE iterations 6 0 SNES Function norm 0.000000000000e+00 1 SNES Function norm 0.000000000000e+00 Nonlinear solve converged due to CONVERGED_FNORM_RELATIVE iterations 1 0 SNES Function norm 0.000000000000e+00 1 SNES Function norm 0.000000000000e+00 Nonlinear solve converged due to CONVERGED_FNORM_RELATIVE iterations 1 0 SNES Function norm 1.486606874732e-01 1 SNES Function norm 8.402777777778e-03 2 SNES Function norm 6.831067663990e-05 3 SNES Function norm 4.665073682466e-09 4 SNES Function norm 0.000000000000e+00 Nonlinear solve converged due to CONVERGED_FNORM_RELATIVE iterations 4 0 SNES Function norm 0.000000000000e+00 1 SNES Function norm 0.000000000000e+00 Nonlinear solve converged due to CONVERGED_FNORM_RELATIVE iterations 1 0 SNES Function norm 0.000000000000e+00 1 SNES Function norm 0.000000000000e+00 Nonlinear solve converged due to CONVERGED_FNORM_RELATIVE iterations 1 0 SNES Function norm 3.605551275464e+00 1 SNES Function norm 4.444444444444e-01 2 SNES Function norm 7.111111111111e-02 3 SNES Function norm 3.936947327951e-03 4 SNES Function norm 1.525925473445e-05 5 SNES Function norm 2.328306437081e-10 6 SNES Function norm 0.000000000000e+00 Nonlinear solve converged due to CONVERGED_FNORM_RELATIVE iterations 6 2.611573 seconds (1.30 M allocations: 69.602 MiB, 22.37% gc time, 77.05% compilation time) [0] PetscDetermineInitialFPTrap(): Floating point trapping is off by default 0 [0] PetscInitialize(): PETSc successfully started: number of processors = 1 [0] PetscGetHostName(): Rejecting domainname, likely is NIS GridapPETSc-primary-0D8uaB42.(none) [0] PetscInitialize(): Running on machine: GridapPETSc-primary-0D8uaB42 [0] PetscCommDuplicate(): Duplicating a communicator 1140850689 -2080374784 max tags = 1073741823 [0] MatAssemblyEnd_SeqAIJ(): Matrix size: 4 X 3; storage space: 2 unneeded,1 used [0] MatAssemblyEnd_SeqAIJ(): Number of mallocs during MatSetValues() is 0 [0] MatAssemblyEnd_SeqAIJ(): Maximum nonzeros in any row is 1 [0] MatCheckCompressedRow(): Found the ratio (num_zerorows 3)/(num_localrows 4) > 0.6. Use CompressedRow routines. [0] MatSeqAIJCheckInode(): Found 2 nodes of 4. Limit used: 5. Using Inode routines 4×3 GridapPETSc.PETScMatrix: -4.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 [0] PetscCommDuplicate(): Using internal PETSc communicator 1140850689 -2080374784 [0] MatAssemblyEnd_SeqAIJ(): Matrix size: 4 X 3; storage space: 11 unneeded,1 used [0] MatAssemblyEnd_SeqAIJ(): Number of mallocs during MatSetValues() is 0 [0] MatAssemblyEnd_SeqAIJ(): Maximum nonzeros in any row is 1 [0] MatCheckCompressedRow(): Found the ratio (num_zerorows 3)/(num_localrows 4) > 0.6. Use CompressedRow routines. [0] MatSeqAIJCheckInode(): Found 2 nodes of 4. Limit used: 5. Using Inode routines 4×3 GridapPETSc.PETScMatrix: -4.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 [0] PetscCommDuplicate(): Using internal PETSc communicator 1140850689 -2080374784 [0] MatAssemblyEnd_SeqAIJ(): Matrix size: 4 X 3; storage space: 0 unneeded,1 used [0] MatAssemblyEnd_SeqAIJ(): Number of mallocs during MatSetValues() is 0 [0] MatAssemblyEnd_SeqAIJ(): Maximum nonzeros in any row is 1 [0] MatCheckCompressedRow(): Found the ratio (num_zerorows 3)/(num_localrows 4) > 0.6. Use CompressedRow routines. [0] MatSeqAIJCheckInode(): Found 2 nodes of 4. Limit used: 5. Using Inode routines 4×3 GridapPETSc.PETScMatrix: -2.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 [0] MatConvert(): Calling duplicate for initial matrix seqaij 0 1 [0] MatConvert(): Calling duplicate for initial matrix seqaij 0 1 [0] PetscCommDuplicate(): Using internal PETSc communicator 1140850689 -2080374784 4-element GridapPETSc.PETScVector: 6.0 0.0 0.0 1.0 ┌ Warning: 1 objects still not finalized before calling GridapPETSc.Finalize() └ @ GridapPETSc ~/.julia/packages/GridapPETSc/X0jXh/src/Environment.jl:28 [0] PetscFinalize(): PetscFinalize() called 8.221547 seconds (2.73 M allocations: 145.192 MiB, 7.21% gc time, 92.34% compilation time: 14% of which was recompilation) [0]PETSC ERROR: --------------------- Error Message -------------------------------------------------------------- [0]PETSC ERROR: Invalid argument [0]PETSC ERROR: Invalid object classid 1211222 This could happen if you compile with PETSC_HAVE_DYNAMIC_LIBRARIES, but link with static libraries. [0]PETSC ERROR: See https://www.mcs.anl.gov/petsc/documentation/faq.html for trouble shooting. [0]PETSC ERROR: Petsc Release Version 3.15.2, Jul 10, 2021 [0]PETSC ERROR: GridapPETSc on a named GridapPETSc-primary-0D8uaB42 by Unknown Thu Jun 12 22:02:49 2025 [0]PETSC ERROR: Configure options --prefix=/workspace/destdir CC=cc FC=gfortran CXX=c++ COPTFLAGS=-O3 CXXOPTFLAGS=-O3 CFLAGS=-fno-stack-protector FOPTFLAGS=-O3 --with-64-bit-indices=0 --with-debugging=0 --with-batch --PETSC_ARCH=x86_64-linux-gnu_double_real_Int32 --with-blaslapack-lib=/workspace/destdir/lib/libopenblas.so --with-blaslapack-suffix= --known-64-bit-blas-indices=0 --with-mpi-lib="[/workspace/destdir/lib/libmpifort.so,/workspace/destdir/lib/libmpi.so]" --known-mpi-int64_t=0 --with-mpi-include=/workspace/destdir/include --with-sowing=0 --with-precision=double --with-scalar-type=real [0]PETSC ERROR: #1 PetscClassRegLogGetClass() at /workspace/srcdir/petsc-3.15.2/src/sys/logging/utils/classlog.c:263 [0]PETSC ERROR: #2 PetscLogObjDestroyDefault() at /workspace/srcdir/petsc-3.15.2/src/sys/logging/utils/classlog.c:363 [0]PETSC ERROR: #3 PetscHeaderDestroy_Private() at /workspace/srcdir/petsc-3.15.2/src/sys/objects/inherit.c:103 [0]PETSC ERROR: #4 VecDestroy() at /workspace/srcdir/petsc-3.15.2/src/vec/vec/interface/vector.c:403 error in running finalizer: GridapPETSc.PETSC.PetscError(code=Int32(1)) macro expansion at /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/src/Config.jl:88 [inlined] Finalize at /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/src/PETScArrays.jl:25 unknown function (ip: 0x778994920cf2) at (unknown file) _jl_invoke at /source/src/gf.c:3692 [inlined] ijl_apply_generic at /source/src/gf.c:3892 run_finalizer at /source/src/gc-common.c:180 jl_gc_run_finalizers_in_list at /source/src/gc-common.c:272 run_finalizers at /source/src/gc-common.c:316 finish_nocycle at ./../usr/share/julia/Compiler/src/typeinfer.jl:241 jfptr_finish_nocycle_86571.1 at /opt/julia/lib/julia/sys.so (unknown line) _jl_invoke at /source/src/gf.c:3692 [inlined] ijl_apply_generic at /source/src/gf.c:3892 typeinf at ./../usr/share/julia/Compiler/src/abstractinterpretation.jl:4490 typeinf_ext at ./../usr/share/julia/Compiler/src/typeinfer.jl:1391 typeinf_ext_toplevel at ./../usr/share/julia/Compiler/src/typeinfer.jl:1560 typeinf_ext_toplevel at ./../usr/share/julia/Compiler/src/typeinfer.jl:1568 jfptr_typeinf_ext_toplevel_86753.1 at /opt/julia/lib/julia/sys.so (unknown line) _jl_invoke at /source/src/gf.c:3692 [inlined] ijl_apply_generic at /source/src/gf.c:3892 jl_apply at /source/src/julia.h:2340 [inlined] jl_type_infer at /source/src/gf.c:461 jl_compile_method_internal at /source/src/gf.c:3199 _jl_invoke at /source/src/gf.c:3684 [inlined] ijl_apply_generic at /source/src/gf.c:3892 _tensor_product_legendre at /home/pkgeval/.julia/packages/Gridap/20Ftr/src/ReferenceFEs/TensorProductQuadratures.jl:21 #Quadrature#117 at /home/pkgeval/.julia/packages/Gridap/20Ftr/src/ReferenceFEs/TensorProductQuadratures.jl:11 unknown function (ip: 0x778975d5e1f1) at (unknown file) _jl_invoke at /source/src/gf.c:3692 [inlined] ijl_apply_generic at /source/src/gf.c:3892 Quadrature at /home/pkgeval/.julia/packages/Gridap/20Ftr/src/ReferenceFEs/TensorProductQuadratures.jl:6 #Quadrature#118 at /home/pkgeval/.julia/packages/Gridap/20Ftr/src/ReferenceFEs/TensorProductQuadratures.jl:16 unknown function (ip: 0x778975d5d6e4) at (unknown file) _jl_invoke at /source/src/gf.c:3692 [inlined] ijl_apply_generic at /source/src/gf.c:3892 Quadrature at /home/pkgeval/.julia/packages/Gridap/20Ftr/src/ReferenceFEs/TensorProductQuadratures.jl:14 #Quadrature#116 at /home/pkgeval/.julia/packages/Gridap/20Ftr/src/ReferenceFEs/Quadratures.jl:127 Quadrature at /home/pkgeval/.julia/packages/Gridap/20Ftr/src/ReferenceFEs/Quadratures.jl:125 [inlined] #3 at /home/pkgeval/.julia/packages/Gridap/20Ftr/src/Geometry/Grids.jl:221 [inlined] iterate at ./generator.jl:48 [inlined] _collect at ./array.jl:816 collect_similar at ./array.jl:738 unknown function (ip: 0x778975d5bf51) at (unknown file) _jl_invoke at /source/src/gf.c:3692 [inlined] ijl_apply_generic at /source/src/gf.c:3892 map at ./abstractarray.jl:3374 unknown function (ip: 0x778976084346) at (unknown file) _jl_invoke at /source/src/gf.c:3692 [inlined] ijl_apply_generic at /source/src/gf.c:3892 #Quadrature#1 at /home/pkgeval/.julia/packages/Gridap/20Ftr/src/Geometry/Grids.jl:221 Quadrature at /home/pkgeval/.julia/packages/Gridap/20Ftr/src/Geometry/Grids.jl:218 [inlined] #CellQuadrature#44 at /home/pkgeval/.julia/packages/Gridap/20Ftr/src/CellData/CellQuadratures.jl:36 CellQuadrature at /home/pkgeval/.julia/packages/Gridap/20Ftr/src/CellData/CellQuadratures.jl:33 unknown function (ip: 0x778975d5a8d0) at (unknown file) _jl_invoke at /source/src/gf.c:3692 [inlined] ijl_apply_generic at /source/src/gf.c:3892 #Measure#58 at /home/pkgeval/.julia/packages/Gridap/20Ftr/src/CellData/DomainContributions.jl:143 _jl_invoke at /source/src/gf.c:3692 [inlined] ijl_apply_generic at /source/src/gf.c:3892 jl_apply at /source/src/julia.h:2340 [inlined] jl_f__apply_iterate at /source/src/builtins.c:864 Measure at /home/pkgeval/.julia/packages/Gridap/20Ftr/src/CellData/DomainContributions.jl:143 _jl_invoke at /source/src/gf.c:3692 [inlined] ijl_apply_generic at /source/src/gf.c:3892 #2 at /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/sequential/PoissonDriver.jl:42 #with#2 at /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/src/Environment.jl:38 unknown function (ip: 0x778976cab432) at (unknown file) _jl_invoke at /source/src/gf.c:3692 [inlined] ijl_apply_generic at /source/src/gf.c:3892 with at /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/src/Environment.jl:36 unknown function (ip: 0x778976ca35e6) at (unknown file) _jl_invoke at /source/src/gf.c:3692 [inlined] ijl_apply_generic at /source/src/gf.c:3892 jl_apply at /source/src/julia.h:2340 [inlined] do_call at /source/src/interpreter.c:123 eval_value at /source/src/interpreter.c:243 eval_stmt_value at /source/src/interpreter.c:194 [inlined] eval_body at /source/src/interpreter.c:708 jl_interpret_toplevel_thunk at /source/src/interpreter.c:899 top-level scope at /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/sequential/PoissonDriver.jl:31 jl_toplevel_eval_flex at /source/src/toplevel.c:773 jl_eval_module_expr at /source/src/toplevel.c:196 [inlined] jl_toplevel_eval_flex at /source/src/toplevel.c:658 jl_toplevel_eval_flex at /source/src/toplevel.c:713 ijl_toplevel_eval at /source/src/toplevel.c:785 ijl_toplevel_eval_in at /source/src/toplevel.c:830 eval at ./boot.jl:489 include_string at ./loading.jl:2846 _jl_invoke at /source/src/gf.c:3692 [inlined] ijl_apply_generic at /source/src/gf.c:3892 _include at ./loading.jl:2906 include at ./Base.jl:310 IncludeInto at ./Base.jl:311 [inlined] macro expansion at /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/sequential/runtests.jl:18 [inlined] macro expansion at /source/usr/share/julia/stdlib/v1.13/Test/src/Test.jl:1837 [inlined] macro expansion at /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/sequential/runtests.jl:18 [inlined] macro expansion at ./timing.jl:645 [inlined] top-level scope at /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/sequential/runtests.jl:353 _jl_invoke at /source/src/gf.c:3692 [inlined] ijl_invoke at /source/src/gf.c:3699 jl_toplevel_eval_flex at /source/src/toplevel.c:762 jl_eval_module_expr at /source/src/toplevel.c:196 [inlined] jl_toplevel_eval_flex at /source/src/toplevel.c:658 jl_toplevel_eval_flex at /source/src/toplevel.c:713 ijl_toplevel_eval at /source/src/toplevel.c:785 ijl_toplevel_eval_in at /source/src/toplevel.c:830 eval at ./boot.jl:489 include_string at ./loading.jl:2846 _jl_invoke at /source/src/gf.c:3692 [inlined] ijl_apply_generic at /source/src/gf.c:3892 _include at ./loading.jl:2906 include at ./Base.jl:310 IncludeInto at ./Base.jl:311 [inlined] macro expansion at /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/runtests.jl:5 [inlined] macro expansion at /source/usr/share/julia/stdlib/v1.13/Test/src/Test.jl:1837 [inlined] macro expansion at /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/runtests.jl:5 [inlined] macro expansion at ./timing.jl:645 [inlined] top-level scope at /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/runtests.jl:353 _jl_invoke at /source/src/gf.c:3692 [inlined] ijl_invoke at /source/src/gf.c:3699 jl_toplevel_eval_flex at /source/src/toplevel.c:762 jl_eval_module_expr at /source/src/toplevel.c:196 [inlined] jl_toplevel_eval_flex at /source/src/toplevel.c:658 jl_toplevel_eval_flex at /source/src/toplevel.c:713 ijl_toplevel_eval at /source/src/toplevel.c:785 ijl_toplevel_eval_in at /source/src/toplevel.c:830 eval at ./boot.jl:489 include_string at ./loading.jl:2846 _jl_invoke at /source/src/gf.c:3692 [inlined] ijl_apply_generic at /source/src/gf.c:3892 _include at ./loading.jl:2906 include at ./Base.jl:310 IncludeInto at ./Base.jl:311 jfptr_IncludeInto_63572.1 at /opt/julia/lib/julia/sys.so (unknown line) _jl_invoke at /source/src/gf.c:3692 [inlined] ijl_apply_generic at /source/src/gf.c:3892 jl_apply at /source/src/julia.h:2340 [inlined] do_call at /source/src/interpreter.c:123 eval_value at /source/src/interpreter.c:243 eval_stmt_value at /source/src/interpreter.c:194 [inlined] eval_body at /source/src/interpreter.c:708 jl_interpret_toplevel_thunk at /source/src/interpreter.c:899 top-level scope at none:6 jl_toplevel_eval_flex at /source/src/toplevel.c:773 jl_toplevel_eval_flex at /source/src/toplevel.c:713 ijl_toplevel_eval at /source/src/toplevel.c:785 ijl_toplevel_eval_in at /source/src/toplevel.c:830 eval at ./boot.jl:489 exec_options at ./client.jl:290 _start at ./client.jl:557 jfptr__start_70077.1 at /opt/julia/lib/julia/sys.so (unknown line) _jl_invoke at /source/src/gf.c:3692 [inlined] ijl_apply_generic at /source/src/gf.c:3892 jl_apply at /source/src/julia.h:2340 [inlined] true_main at /source/src/jlapi.c:964 jl_repl_entrypoint at /source/src/jlapi.c:1132 main at /source/cli/loader_exe.c:58 unknown function (ip: 0x7789961c7249) at /lib/x86_64-linux-gnu/libc.so.6 __libc_start_main at /lib/x86_64-linux-gnu/libc.so.6 (unknown line) unknown function (ip: 0x4010b8) at /workspace/srcdir/glibc-2.17/csu/../sysdeps/x86_64/start.S 0 KSP Residual norm 8.550000000000e-03 1 KSP Residual norm 2.039013341310e-04 2 KSP Residual norm 6.195852905698e-06 3 KSP Residual norm 1.300754719285e-07 4 KSP Residual norm 3.121504432740e-09 5 KSP Residual norm 7.977302448122e-11 6 KSP Residual norm 1.598148195898e-12 7 KSP Residual norm 3.741486163257e-14 Linear solve converged due to CONVERGED_RTOL iterations 7 KSP Object: 1 MPI processes type: cg maximum iterations=1000, initial guess is zero tolerances: relative=1e-10, absolute=1e-50, divergence=10000. left preconditioning using UNPRECONDITIONED norm type for convergence test PC Object: 1 MPI processes type: gamg type is MULTIPLICATIVE, levels=3 cycles=v Cycles per PCApply=1 Using externally compute Galerkin coarse grid matrices GAMG specific options Threshold for dropping small values in graph on each level = 0. 0. 0. Threshold scaling factor for each level not specified = 1. AGG specific options Symmetric graph false Number of levels to square graph 0 Number smoothing steps 1 Complexity: grid = 1.15878 Coarse grid solver -- level ------------------------------- KSP Object: (mg_coarse_) 1 MPI processes type: preonly maximum iterations=10000, initial guess is zero tolerances: relative=1e-05, absolute=1e-50, divergence=10000. left preconditioning using NONE norm type for convergence test PC Object: (mg_coarse_) 1 MPI processes type: bjacobi number of blocks = 1 Local solver information for first block is in the following KSP and PC objects on rank 0: Use -mg_coarse_ksp_view ::ascii_info_detail to display information for all blocks KSP Object: (mg_coarse_sub_) 1 MPI processes type: preonly maximum iterations=1, initial guess is zero tolerances: relative=1e-05, absolute=1e-50, divergence=10000. left preconditioning using NONE norm type for convergence test PC Object: (mg_coarse_sub_) 1 MPI processes type: cholesky out-of-place factorization tolerance for zero pivot 2.22045e-14 matrix ordering: nd factor fill ratio given 5., needed 1. Factored matrix follows: Mat Object: 1 MPI processes type: seqsbaij rows=7, cols=7 package used to perform factorization: petsc total: nonzeros=28, allocated nonzeros=28 block size is 1 linear system matrix = precond matrix: Mat Object: (mg_coarse_sub_) 1 MPI processes type: seqaij rows=7, cols=7 total: nonzeros=49, allocated nonzeros=49 total number of mallocs used during MatSetValues calls=0 using I-node routines: found 2 nodes, limit used is 5 linear system matrix = precond matrix: Mat Object: (mg_coarse_sub_) 1 MPI processes type: seqaij rows=7, cols=7 total: nonzeros=49, allocated nonzeros=49 total number of mallocs used during MatSetValues calls=0 using I-node routines: found 2 nodes, limit used is 5 Down solver (pre-smoother) on level 1 ------------------------------- KSP Object: (mg_levels_1_) 1 MPI processes type: chebyshev eigenvalue estimates used: min = 0.0998343, max = 1.09818 eigenvalues estimate via cg min 0.516584, max 0.998343 eigenvalues estimated using cg with translations [0. 0.1; 0. 1.1] KSP Object: (mg_levels_1_esteig_) 1 MPI processes type: cg maximum iterations=10, initial guess is zero tolerances: relative=1e-12, absolute=1e-50, divergence=10000. left preconditioning using PRECONDITIONED norm type for convergence test estimating eigenvalues using noisy right hand side maximum iterations=2, nonzero initial guess tolerances: relative=1e-05, absolute=1e-50, divergence=10000. left preconditioning using NONE norm type for convergence test PC Object: (mg_levels_1_) 1 MPI processes type: sor type = local_symmetric, iterations = 1, local iterations = 1, omega = 1. linear system matrix = precond matrix: Mat Object: 1 MPI processes type: seqaij rows=76, cols=76 total: nonzeros=2432, allocated nonzeros=2432 total number of mallocs used during MatSetValues calls=0 not using I-node routines Up solver (post-smoother) same as down solver (pre-smoother) Down solver (pre-smoother) on level 2 ------------------------------- KSP Object: (mg_levels_2_) 1 MPI processes type: chebyshev eigenvalue estimates used: min = 0.0998116, max = 1.09793 eigenvalues estimate via cg min 0.336175, max 0.998116 eigenvalues estimated using cg with translations [0. 0.1; 0. 1.1] KSP Object: (mg_levels_2_esteig_) 1 MPI processes type: cg maximum iterations=10, initial guess is zero tolerances: relative=1e-12, absolute=1e-50, divergence=10000. left preconditioning using PRECONDITIONED norm type for convergence test estimating eigenvalues using noisy right hand side maximum iterations=2, nonzero initial guess tolerances: relative=1e-05, absolute=1e-50, divergence=10000. left preconditioning using NONE norm type for convergence test PC Object: (mg_levels_2_) 1 MPI processes type: sor type = local_symmetric, iterations = 1, local iterations = 1, omega = 1. linear system matrix = precond matrix: Mat Object: 1 MPI processes type: seqaij rows=729, cols=729 total: nonzeros=15625, allocated nonzeros=0 total number of mallocs used during MatSetValues calls=0 not using I-node routines Up solver (post-smoother) same as down solver (pre-smoother) linear system matrix = precond matrix: Mat Object: 1 MPI processes type: seqaij rows=729, cols=729 total: nonzeros=15625, allocated nonzeros=0 total number of mallocs used during MatSetValues calls=0 not using I-node routines 0 KSP Residual norm 8.550000000000e-03 1 KSP Residual norm 2.039013341310e-04 2 KSP Residual norm 6.195852905698e-06 3 KSP Residual norm 1.300754719285e-07 4 KSP Residual norm 3.121504432740e-09 5 KSP Residual norm 7.977302448122e-11 6 KSP Residual norm 1.598148195898e-12 7 KSP Residual norm 3.741486163257e-14 Linear solve converged due to CONVERGED_RTOL iterations 7 KSP Object: 1 MPI processes type: cg maximum iterations=1000, initial guess is zero tolerances: relative=1e-10, absolute=1e-50, divergence=10000. left preconditioning using UNPRECONDITIONED norm type for convergence test PC Object: 1 MPI processes type: gamg type is MULTIPLICATIVE, levels=3 cycles=v Cycles per PCApply=1 Using externally compute Galerkin coarse grid matrices GAMG specific options Threshold for dropping small values in graph on each level = 0. 0. 0. Threshold scaling factor for each level not specified = 1. AGG specific options Symmetric graph false Number of levels to square graph 0 Number smoothing steps 1 Complexity: grid = 1.15878 Coarse grid solver -- level ------------------------------- KSP Object: (mg_coarse_) 1 MPI processes type: preonly maximum iterations=10000, initial guess is zero tolerances: relative=1e-05, absolute=1e-50, divergence=10000. left preconditioning using NONE norm type for convergence test PC Object: (mg_coarse_) 1 MPI processes type: bjacobi number of blocks = 1 Local solver information for first block is in the following KSP and PC objects on rank 0: Use -mg_coarse_ksp_view ::ascii_info_detail to display information for all blocks KSP Object: (mg_coarse_sub_) 1 MPI processes type: preonly maximum iterations=1, initial guess is zero tolerances: relative=1e-05, absolute=1e-50, divergence=10000. left preconditioning using NONE norm type for convergence test PC Object: (mg_coarse_sub_) 1 MPI processes type: cholesky out-of-place factorization tolerance for zero pivot 2.22045e-14 matrix ordering: nd factor fill ratio given 5., needed 1. Factored matrix follows: Mat Object: 1 MPI processes type: seqsbaij rows=7, cols=7 package used to perform factorization: petsc total: nonzeros=28, allocated nonzeros=28 block size is 1 linear system matrix = precond matrix: Mat Object: (mg_coarse_sub_) 1 MPI processes type: seqaij rows=7, cols=7 total: nonzeros=49, allocated nonzeros=49 total number of mallocs used during MatSetValues calls=0 using I-node routines: found 2 nodes, limit used is 5 linear system matrix = precond matrix: Mat Object: (mg_coarse_sub_) 1 MPI processes type: seqaij rows=7, cols=7 total: nonzeros=49, allocated nonzeros=49 total number of mallocs used during MatSetValues calls=0 using I-node routines: found 2 nodes, limit used is 5 Down solver (pre-smoother) on level 1 ------------------------------- KSP Object: (mg_levels_1_) 1 MPI processes type: chebyshev eigenvalue estimates used: min = 0.0998343, max = 1.09818 eigenvalues estimate via cg min 0.516584, max 0.998343 eigenvalues estimated using cg with translations [0. 0.1; 0. 1.1] KSP Object: (mg_levels_1_esteig_) 1 MPI processes type: cg maximum iterations=10, initial guess is zero tolerances: relative=1e-12, absolute=1e-50, divergence=10000. left preconditioning using PRECONDITIONED norm type for convergence test estimating eigenvalues using noisy right hand side maximum iterations=2, nonzero initial guess tolerances: relative=1e-05, absolute=1e-50, divergence=10000. left preconditioning using NONE norm type for convergence test PC Object: (mg_levels_1_) 1 MPI processes type: sor type = local_symmetric, iterations = 1, local iterations = 1, omega = 1. linear system matrix = precond matrix: Mat Object: 1 MPI processes type: seqaij rows=76, cols=76 total: nonzeros=2432, allocated nonzeros=2432 total number of mallocs used during MatSetValues calls=0 not using I-node routines Up solver (post-smoother) same as down solver (pre-smoother) Down solver (pre-smoother) on level 2 ------------------------------- KSP Object: (mg_levels_2_) 1 MPI processes type: chebyshev eigenvalue estimates used: min = 0.0998116, max = 1.09793 eigenvalues estimate via cg min 0.336175, max 0.998116 eigenvalues estimated using cg with translations [0. 0.1; 0. 1.1] KSP Object: (mg_levels_2_esteig_) 1 MPI processes type: cg maximum iterations=10, initial guess is zero tolerances: relative=1e-12, absolute=1e-50, divergence=10000. left preconditioning using PRECONDITIONED norm type for convergence test estimating eigenvalues using noisy right hand side maximum iterations=2, nonzero initial guess tolerances: relative=1e-05, absolute=1e-50, divergence=10000. left preconditioning using NONE norm type for convergence test PC Object: (mg_levels_2_) 1 MPI processes type: sor type = local_symmetric, iterations = 1, local iterations = 1, omega = 1. linear system matrix = precond matrix: Mat Object: 1 MPI processes type: seqaij rows=729, cols=729 total: nonzeros=15625, allocated nonzeros=0 total number of mallocs used during MatSetValues calls=0 not using I-node routines Up solver (post-smoother) same as down solver (pre-smoother) linear system matrix = precond matrix: Mat Object: 1 MPI processes type: seqaij rows=729, cols=729 total: nonzeros=15625, allocated nonzeros=0 total number of mallocs used during MatSetValues calls=0 not using I-node routines 0 KSP Residual norm 8.550000000000e-03 1 KSP Residual norm 2.039013341310e-04 2 KSP Residual norm 6.195852905698e-06 3 KSP Residual norm 1.300754719285e-07 4 KSP Residual norm 3.121504432740e-09 5 KSP Residual norm 7.977302448122e-11 6 KSP Residual norm 1.598148195898e-12 7 KSP Residual norm 3.741486163257e-14 Linear solve converged due to CONVERGED_RTOL iterations 7 KSP Object: 1 MPI processes type: cg maximum iterations=1000, initial guess is zero tolerances: relative=1e-10, absolute=1e-50, divergence=10000. left preconditioning using UNPRECONDITIONED norm type for convergence test PC Object: 1 MPI processes type: gamg type is MULTIPLICATIVE, levels=3 cycles=v Cycles per PCApply=1 Using externally compute Galerkin coarse grid matrices GAMG specific options Threshold for dropping small values in graph on each level = 0. 0. 0. Threshold scaling factor for each level not specified = 1. AGG specific options Symmetric graph false Number of levels to square graph 0 Number smoothing steps 1 Complexity: grid = 1.15878 Coarse grid solver -- level ------------------------------- KSP Object: (mg_coarse_) 1 MPI processes type: preonly maximum iterations=10000, initial guess is zero tolerances: relative=1e-05, absolute=1e-50, divergence=10000. left preconditioning using NONE norm type for convergence test PC Object: (mg_coarse_) 1 MPI processes type: bjacobi number of blocks = 1 Local solver information for first block is in the following KSP and PC objects on rank 0: Use -mg_coarse_ksp_view ::ascii_info_detail to display information for all blocks KSP Object: (mg_coarse_sub_) 1 MPI processes type: preonly maximum iterations=1, initial guess is zero tolerances: relative=1e-05, absolute=1e-50, divergence=10000. left preconditioning using NONE norm type for convergence test PC Object: (mg_coarse_sub_) 1 MPI processes type: cholesky out-of-place factorization tolerance for zero pivot 2.22045e-14 matrix ordering: nd factor fill ratio given 5., needed 1. Factored matrix follows: Mat Object: 1 MPI processes type: seqsbaij rows=7, cols=7 package used to perform factorization: petsc total: nonzeros=28, allocated nonzeros=28 block size is 1 linear system matrix = precond matrix: Mat Object: (mg_coarse_sub_) 1 MPI processes type: seqaij rows=7, cols=7 total: nonzeros=49, allocated nonzeros=49 total number of mallocs used during MatSetValues calls=0 using I-node routines: found 2 nodes, limit used is 5 linear system matrix = precond matrix: Mat Object: (mg_coarse_sub_) 1 MPI processes type: seqaij rows=7, cols=7 total: nonzeros=49, allocated nonzeros=49 total number of mallocs used during MatSetValues calls=0 using I-node routines: found 2 nodes, limit used is 5 Down solver (pre-smoother) on level 1 ------------------------------- KSP Object: (mg_levels_1_) 1 MPI processes type: chebyshev eigenvalue estimates used: min = 0.0998343, max = 1.09818 eigenvalues estimate via cg min 0.516584, max 0.998343 eigenvalues estimated using cg with translations [0. 0.1; 0. 1.1] KSP Object: (mg_levels_1_esteig_) 1 MPI processes type: cg maximum iterations=10, initial guess is zero tolerances: relative=1e-12, absolute=1e-50, divergence=10000. left preconditioning using PRECONDITIONED norm type for convergence test estimating eigenvalues using noisy right hand side maximum iterations=2, nonzero initial guess tolerances: relative=1e-05, absolute=1e-50, divergence=10000. left preconditioning using NONE norm type for convergence test PC Object: (mg_levels_1_) 1 MPI processes type: sor type = local_symmetric, iterations = 1, local iterations = 1, omega = 1. linear system matrix = precond matrix: Mat Object: 1 MPI processes type: seqaij rows=76, cols=76 total: nonzeros=2432, allocated nonzeros=2432 total number of mallocs used during MatSetValues calls=0 not using I-node routines Up solver (post-smoother) same as down solver (pre-smoother) Down solver (pre-smoother) on level 2 ------------------------------- KSP Object: (mg_levels_2_) 1 MPI processes type: chebyshev eigenvalue estimates used: min = 0.0998116, max = 1.09793 eigenvalues estimate via cg min 0.336175, max 0.998116 eigenvalues estimated using cg with translations [0. 0.1; 0. 1.1] KSP Object: (mg_levels_2_esteig_) 1 MPI processes type: cg maximum iterations=10, initial guess is zero tolerances: relative=1e-12, absolute=1e-50, divergence=10000. left preconditioning using PRECONDITIONED norm type for convergence test estimating eigenvalues using noisy right hand side maximum iterations=2, nonzero initial guess tolerances: relative=1e-05, absolute=1e-50, divergence=10000. left preconditioning using NONE norm type for convergence test PC Object: (mg_levels_2_) 1 MPI processes type: sor type = local_symmetric, iterations = 1, local iterations = 1, omega = 1. linear system matrix = precond matrix: Mat Object: 1 MPI processes type: seqaij rows=729, cols=729 total: nonzeros=15625, allocated nonzeros=39304 total number of mallocs used during MatSetValues calls=0 not using I-node routines Up solver (post-smoother) same as down solver (pre-smoother) linear system matrix = precond matrix: Mat Object: 1 MPI processes type: seqaij rows=729, cols=729 total: nonzeros=15625, allocated nonzeros=39304 total number of mallocs used during MatSetValues calls=0 not using I-node routines 0 KSP Residual norm 8.550000000000e-03 1 KSP Residual norm 2.039013341310e-04 2 KSP Residual norm 6.195852905698e-06 3 KSP Residual norm 1.300754719285e-07 4 KSP Residual norm 3.121504432740e-09 5 KSP Residual norm 7.977302448122e-11 6 KSP Residual norm 1.598148195898e-12 7 KSP Residual norm 3.741486163257e-14 Linear solve converged due to CONVERGED_RTOL iterations 7 KSP Object: 1 MPI processes type: cg maximum iterations=1000, initial guess is zero tolerances: relative=1e-10, absolute=1e-50, divergence=10000. left preconditioning using UNPRECONDITIONED norm type for convergence test PC Object: 1 MPI processes type: gamg type is MULTIPLICATIVE, levels=3 cycles=v Cycles per PCApply=1 Using externally compute Galerkin coarse grid matrices GAMG specific options Threshold for dropping small values in graph on each level = 0. 0. 0. Threshold scaling factor for each level not specified = 1. AGG specific options Symmetric graph false Number of levels to square graph 0 Number smoothing steps 1 Complexity: grid = 1.15878 Coarse grid solver -- level ------------------------------- KSP Object: (mg_coarse_) 1 MPI processes type: preonly maximum iterations=10000, initial guess is zero tolerances: relative=1e-05, absolute=1e-50, divergence=10000. left preconditioning using NONE norm type for convergence test PC Object: (mg_coarse_) 1 MPI processes type: bjacobi number of blocks = 1 Local solver information for first block is in the following KSP and PC objects on rank 0: Use -mg_coarse_ksp_view ::ascii_info_detail to display information for all blocks KSP Object: (mg_coarse_sub_) 1 MPI processes type: preonly maximum iterations=1, initial guess is zero tolerances: relative=1e-05, absolute=1e-50, divergence=10000. left preconditioning using NONE norm type for convergence test PC Object: (mg_coarse_sub_) 1 MPI processes type: cholesky out-of-place factorization tolerance for zero pivot 2.22045e-14 matrix ordering: nd factor fill ratio given 5., needed 1. Factored matrix follows: Mat Object: 1 MPI processes type: seqsbaij rows=7, cols=7 package used to perform factorization: petsc total: nonzeros=28, allocated nonzeros=28 block size is 1 linear system matrix = precond matrix: Mat Object: (mg_coarse_sub_) 1 MPI processes type: seqaij rows=7, cols=7 total: nonzeros=49, allocated nonzeros=49 total number of mallocs used during MatSetValues calls=0 using I-node routines: found 2 nodes, limit used is 5 linear system matrix = precond matrix: Mat Object: (mg_coarse_sub_) 1 MPI processes type: seqaij rows=7, cols=7 total: nonzeros=49, allocated nonzeros=49 total number of mallocs used during MatSetValues calls=0 using I-node routines: found 2 nodes, limit used is 5 Down solver (pre-smoother) on level 1 ------------------------------- KSP Object: (mg_levels_1_) 1 MPI processes type: chebyshev eigenvalue estimates used: min = 0.0998343, max = 1.09818 eigenvalues estimate via cg min 0.516584, max 0.998343 eigenvalues estimated using cg with translations [0. 0.1; 0. 1.1] KSP Object: (mg_levels_1_esteig_) 1 MPI processes type: cg maximum iterations=10, initial guess is zero tolerances: relative=1e-12, absolute=1e-50, divergence=10000. left preconditioning using PRECONDITIONED norm type for convergence test estimating eigenvalues using noisy right hand side maximum iterations=2, nonzero initial guess tolerances: relative=1e-05, absolute=1e-50, divergence=10000. left preconditioning using NONE norm type for convergence test PC Object: (mg_levels_1_) 1 MPI processes type: sor type = local_symmetric, iterations = 1, local iterations = 1, omega = 1. linear system matrix = precond matrix: Mat Object: 1 MPI processes type: seqaij rows=76, cols=76 total: nonzeros=2432, allocated nonzeros=2432 total number of mallocs used during MatSetValues calls=0 not using I-node routines Up solver (post-smoother) same as down solver (pre-smoother) Down solver (pre-smoother) on level 2 ------------------------------- KSP Object: (mg_levels_2_) 1 MPI processes type: chebyshev eigenvalue estimates used: min = 0.0998116, max = 1.09793 eigenvalues estimate via cg min 0.336175, max 0.998116 eigenvalues estimated using cg with translations [0. 0.1; 0. 1.1] KSP Object: (mg_levels_2_esteig_) 1 MPI processes type: cg maximum iterations=10, initial guess is zero tolerances: relative=1e-12, absolute=1e-50, divergence=10000. left preconditioning using PRECONDITIONED norm type for convergence test estimating eigenvalues using noisy right hand side maximum iterations=2, nonzero initial guess tolerances: relative=1e-05, absolute=1e-50, divergence=10000. left preconditioning using NONE norm type for convergence test PC Object: (mg_levels_2_) 1 MPI processes type: sor type = local_symmetric, iterations = 1, local iterations = 1, omega = 1. linear system matrix = precond matrix: Mat Object: 1 MPI processes type: seqaij rows=729, cols=729 total: nonzeros=15625, allocated nonzeros=39304 total number of mallocs used during MatSetValues calls=0 not using I-node routines Up solver (post-smoother) same as down solver (pre-smoother) linear system matrix = precond matrix: Mat Object: 1 MPI processes type: seqaij rows=729, cols=729 total: nonzeros=15625, allocated nonzeros=39304 total number of mallocs used during MatSetValues calls=0 not using I-node routines 138.437880 seconds (122.37 M allocations: 6.282 GiB, 3.05% gc time, 99.17% compilation time) 0 KSP Residual norm 0.0467654 1 KSP Residual norm 0.0051782 2 KSP Residual norm 0.000607254 3 KSP Residual norm 4.67692e-05 4 KSP Residual norm 3.52562e-06 5 KSP Residual norm 3.76699e-06 6 KSP Residual norm 1.13505e-06 7 KSP Residual norm 8.08586e-07 8 KSP Residual norm 1.87761e-07 9 KSP Residual norm 1.5687e-07 10 KSP Residual norm 3.40583e-08 11 KSP Residual norm 1.62582e-08 12 KSP Residual norm 8.72662e-09 13 KSP Residual norm 4.20939e-09 14 KSP Residual norm 1.82524e-09 15 KSP Residual norm 4.163e-10 Linear solve converged due to CONVERGED_RTOL iterations 15 32.350599 seconds (22.22 M allocations: 1.128 GiB, 3.34% gc time, 96.37% compilation time) [ Info: Skipping DarcyDriver since petsc is not configured with mumps. 6.652448 seconds (9.25 M allocations: 478.248 MiB, 14.19% gc time, 85.53% compilation time: 5% of which was recompilation) 0 SNES Function norm 2.409822369424e+06 Linear solve converged due to CONVERGED_RTOL iterations 54 1 SNES Function norm 8.075649416004e+00 Linear solve converged due to CONVERGED_RTOL iterations 39 2 SNES Function norm 4.662170753064e-01 Linear solve converged due to CONVERGED_RTOL iterations 37 3 SNES Function norm 8.399998666395e-03 Linear solve converged due to CONVERGED_RTOL iterations 37 4 SNES Function norm 4.665302402881e-06 Linear solve converged due to CONVERGED_RTOL iterations 44 5 SNES Function norm 1.573803618773e-11 Nonlinear solve converged due to CONVERGED_FNORM_RELATIVE iterations 5 0 SNES Function norm 2.409822369424e+06 Linear solve converged due to CONVERGED_RTOL iterations 54 1 SNES Function norm 8.075649416004e+00 Linear solve converged due to CONVERGED_RTOL iterations 39 2 SNES Function norm 4.662170753064e-01 Linear solve converged due to CONVERGED_RTOL iterations 37 3 SNES Function norm 8.399998666395e-03 Linear solve converged due to CONVERGED_RTOL iterations 37 4 SNES Function norm 4.665302402881e-06 Linear solve converged due to CONVERGED_RTOL iterations 44 5 SNES Function norm 1.573803618773e-11 Nonlinear solve converged due to CONVERGED_FNORM_RELATIVE iterations 5 53.674729 seconds (33.88 M allocations: 1.750 GiB, 2.67% gc time, 96.13% compilation time: <1% of which was recompilation) Linear solve converged due to CONVERGED_RTOL iterations 28 184.456649 seconds (101.25 M allocations: 5.242 GiB, 1.30% gc time, 94.79% compilation time: 2% of which was recompilation) 0 SNES Function norm 2.409822369424e+06 Linear solve converged due to CONVERGED_RTOL iterations 258 1 SNES Function norm 8.077651079465e+00 Linear solve converged due to CONVERGED_RTOL iterations 225 2 SNES Function norm 4.661493250862e-01 Linear solve converged due to CONVERGED_RTOL iterations 215 3 SNES Function norm 8.397548297111e-03 Linear solve converged due to CONVERGED_RTOL iterations 246 4 SNES Function norm 4.669108166700e-06 Linear solve converged due to CONVERGED_RTOL iterations 313 5 SNES Function norm 4.635673871892e-11 Nonlinear solve converged due to CONVERGED_FNORM_RELATIVE iterations 5 0 SNES Function norm 4.635673871892e-11 Linear solve converged due to CONVERGED_RTOL iterations 287 1 SNES Function norm 5.072214559542e-13 Nonlinear solve converged due to CONVERGED_SNORM_RELATIVE iterations 1 0 SNES Function norm 2.409822369424e+06 Linear solve converged due to CONVERGED_RTOL iterations 258 1 SNES Function norm 8.077651079465e+00 Linear solve converged due to CONVERGED_RTOL iterations 225 2 SNES Function norm 4.661493250862e-01 Linear solve converged due to CONVERGED_RTOL iterations 215 3 SNES Function norm 8.397548297115e-03 Linear solve converged due to CONVERGED_RTOL iterations 246 4 SNES Function norm 4.669108167910e-06 Linear solve converged due to CONVERGED_RTOL iterations 313 5 SNES Function norm 4.635742176199e-11 Nonlinear solve converged due to CONVERGED_FNORM_RELATIVE iterations 5 0 SNES Function norm 4.635742176199e-11 Linear solve converged due to CONVERGED_RTOL iterations 287 1 SNES Function norm 5.163441011326e-13 Nonlinear solve converged due to CONVERGED_SNORM_RELATIVE iterations 1 70.887592 seconds (24.04 M allocations: 1.333 GiB, 2.35% gc time, 61.74% compilation time: <1% of which was recompilation) Test Summary: | Pass Total Time SEQUENTIAL | 172 172 10m11.8s 611.895222 seconds (376.49 M allocations: 19.562 GiB, 2.88% gc time, 88.48% compilation time: 2% of which was recompilation) ┌ Warning: mpiexec() is deprecated, use the non-do-block form │ caller = ip:0x0 └ @ Core :-1 ERROR: ERROR: ERROR: LoadError: LoadError: LoadError: ArgumentError: Package Gridap [56d4f2e9-7ea1-5844-9cf6-b9c51ca7ce8e] is required but does not seem to be installed: - Run `Pkg.instantiate()` to install all recorded dependencies. ArgumentError: Package Gridap [56d4f2e9-7ea1-5844-9cf6-b9c51ca7ce8e] is required but does not seem to be installed: - Run `Pkg.instantiate()` to install all recorded dependencies. Stacktrace: Stacktrace: ArgumentError: Package Gridap [56d4f2e9-7ea1-5844-9cf6-b9c51ca7ce8e] is required but does not seem to be installed: - Run `Pkg.instantiate()` to install all recorded dependencies. Stacktrace: [1] [1] __require_prelocked(pkg::Base.PkgId, env::String)__require_prelocked(pkg::Base.PkgId, env::String) @ Base ./loading.jl:2588 [2] _require_prelocked(uuidkey::Base.PkgId, env::String) @ Base ./loading.jl:2466 [3] macro expansion @ ./loading.jl:2394 [inlined] [4] macro expansion @ ./lock.jl:376 [inlined] [5] __require(into::Module, mod::Symbol) @ Base ./loading.jl:2359 [6] require @ ./loading.jl:2335 [inlined] [7] eval_import_path @ ./module.jl:36 [inlined] [8] eval_import_path_all(at::Module, path::Expr, keyword::String) @ Base ./module.jl:60 [9] _eval_using(to::Module, path::Expr) @ Base ./module.jl:137 [10] top-level scope @ @ Base ./loading.jl:2588 [2] _require_prelocked(uuidkey::Base.PkgId, env::String) @ Base ./loading.jl:2466 [3] macro expansion @ ./loading.jl:2394 [inlined] [4] macro expansion @ ./lock.jl:376 [inlined] [5] __require(into::Module, mod::Symbol) @ Base ./loading.jl:2359 [6] require @ ./loading.jl:2335 [inlined] [7] eval_import_path @ ./module.jl:36 [inlined] [8] eval_import_path_all(at::Module, path::Expr, keyword::String) @ Base ./module.jl:60 [9] _eval_using(to::Module, path::Expr) @ Base ./module.jl:137 [10] top-level scope @ ~/.julia/packages/GridapPETSc/X0jXh/test/PartitionedArraysTests.jl:1 [11] include(mapexpr::Function, mod::Module, _path::String) @ Base ./Base.jl:310 [12] top-level scope @ ~/.julia/packages/GridapPETSc/X0jXh/test/mpi/PartitionedArraysTests.jl:1 [13] include(mod::Module, _path::String) @ Base ./Base.jl:309 [14] exec_options(opts::Base.JLOptions) @ Base ./client.jl:324 [15] _start() @ Base ./client.jl:557 ~/.julia/packages/GridapPETSc/X0jXh/test/PartitionedArraysTests.jl:1 [11] include(mapexpr::Function, mod::Module, _path::String) @ Base ./Base.jl:310 [12] top-level scope @ ~/.julia/packages/GridapPETSc/X0jXh/test/mpi/PartitionedArraysTests.jl:1 [13] include(mod::Module, _path::String) @ Base ./Base.jl:309 [14] exec_options(opts::Base.JLOptions) @ Base ./client.jl:324 [15] _start() @ Base ./client.jl:557 [1] in expression starting at /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/PartitionedArraysTests.jl:1 in expression starting at /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/PartitionedArraysTests.jl:1 in expression starting at /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/mpi/PartitionedArraysTests.jl:1 in expression starting at /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/mpi/PartitionedArraysTests.jl:1 __require_prelocked(pkg::Base.PkgId, env::String) @ Base ./loading.jl:2588 [2] _require_prelocked(uuidkey::Base.PkgId, env::String) @ Base ./loading.jl:2466 [3] macro expansion @ ./loading.jl:2394 [inlined] [4] macro expansion @ ./lock.jl:376 [inlined] [5] __require(into::Module, mod::Symbol) @ Base ./loading.jl:2359 [6] require @ ./loading.jl:2335 [inlined] [7] eval_import_path @ ./module.jl:36 [inlined] [8] eval_import_path_all(at::Module, path::Expr, keyword::String) @ Base ./module.jl:60 [9] _eval_using(to::Module, path::Expr) @ Base ./module.jl:137 [10] top-level scope @ ~/.julia/packages/GridapPETSc/X0jXh/test/PartitionedArraysTests.jl:1 [11] include(mapexpr::Function, mod::Module, _path::String) @ Base ./Base.jl:310 [12] top-level scope @ ~/.julia/packages/GridapPETSc/X0jXh/test/mpi/PartitionedArraysTests.jl:1 [13] include(mod::Module, _path::String) @ Base ./Base.jl:309 [14] exec_options(opts::Base.JLOptions) @ Base ./client.jl:324 [15] _start() @ Base ./client.jl:557 in expression starting at /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/PartitionedArraysTests.jl:1 in expression starting at /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/mpi/PartitionedArraysTests.jl:1 PartitionedArrays: Error During Test at /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/mpi/runtests.jl:7 Got exception outside of a @test LoadError: failed process: Process(`/home/pkgeval/.julia/artifacts/05d8c79b270470018e9de8dd24ddb6d7954aff9d/bin/mpiexec -n 3 /opt/julia/bin/julia -C native -J/opt/julia/lib/julia/sys.so --depwarn=yes --check-bounds=yes --pkgimages=existing -g1 --startup-file=no --project=/home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/mpi/../.. /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/mpi/PartitionedArraysTests.jl`, ProcessExited(1)) [1] Stacktrace: [1] pipeline_error @ ./process.jl:602 [inlined] [2] run(::Cmd; wait::Bool) @ Base ./process.jl:517 [3] run @ ./process.jl:514 [inlined] [4] (::Main.GridapPETScTests.GridapPETScMPITests.PartitionedArraysTestsRun.var"#3#4"{Int64, String, String, String})(cmd::String) @ Main.GridapPETScTests.GridapPETScMPITests.PartitionedArraysTestsRun ~/.julia/packages/GridapPETSc/X0jXh/test/mpi/mpiexec.jl:11 [5] top-level scope @ ~/.julia/packages/JLLWrappers/GfYNv/src/runtime.jl:49 [6] withenv(::JLLWrappers.var"#withenv_executable_wrapper##0#withenv_executable_wrapper##1"{Main.GridapPETScTests.GridapPETScMPITests.PartitionedArraysTestsRun.var"#3#4"{Int64, String, String, String}, String}, ::Pair{String, String}, ::Vararg{Pair{String, String}}) @ Base ./env.jl:265 [7] top-level scope @ ~/.julia/packages/JLLWrappers/GfYNv/src/runtime.jl:48 [8] #mpiexec#1 @ ~/.julia/packages/JLLWrappers/GfYNv/src/products/executable_generators.jl:28 [inlined] [9] mpiexec @ ~/.julia/packages/JLLWrappers/GfYNv/src/products/executable_generators.jl:25 [inlined] [10] run_mpi_driver(; procs::Int64, file::String) @ Main.GridapPETScTests.GridapPETScMPITests.PartitionedArraysTestsRun ~/.julia/packages/GridapPETSc/X0jXh/test/mpi/mpiexec.jl:7 [11] top-level scope @ ~/.julia/packages/GridapPETSc/X0jXh/test/mpi/PartitionedArraysTestsRun.jl:3 [12] include(mapexpr::Function, mod::Module, _path::String) @ Base ./Base.jl:310 [13] IncludeInto @ ./Base.jl:311 [inlined] [14] macro expansion @ ~/.julia/packages/GridapPETSc/X0jXh/test/mpi/runtests.jl:7 [inlined] [15] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1837 [inlined] [16] macro expansion @ ~/.julia/packages/GridapPETSc/X0jXh/test/mpi/runtests.jl:7 [inlined] [17] macro expansion @ ./timing.jl:645 [inlined] [18] top-level scope @ ~/.julia/packages/GridapPETSc/X0jXh/test/mpi/runtests.jl:353 [19] include(mapexpr::Function, mod::Module, _path::String) @ Base ./Base.jl:310 [20] IncludeInto @ ./Base.jl:311 [inlined] [21] macro expansion @ ~/.julia/packages/GridapPETSc/X0jXh/test/runtests.jl:6 [inlined] [22] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1837 [inlined] [23] macro expansion @ ~/.julia/packages/GridapPETSc/X0jXh/test/runtests.jl:6 [inlined] [24] macro expansion @ ./timing.jl:645 [inlined] [25] top-level scope @ ~/.julia/packages/GridapPETSc/X0jXh/test/runtests.jl:353 [26] include(mapexpr::Function, mod::Module, _path::String) @ Base ./Base.jl:310 [27] top-level scope @ none:6 [28] eval(m::Module, e::Any) @ Core ./boot.jl:489 [29] exec_options(opts::Base.JLOptions) @ Base ./client.jl:290 [30] _start() @ Base ./client.jl:557 in expression starting at /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/mpi/PartitionedArraysTestsRun.jl:1 27.601332 seconds (2.16 M allocations: 118.072 MiB, 19.07% compilation time: <1% of which was recompilation) ERROR: ERROR: LoadError: LoadError: ArgumentError: Package Gridap [56d4f2e9-7ea1-5844-9cf6-b9c51ca7ce8e] is required but does not seem to be installed: - Run `Pkg.instantiate()` to install all recorded dependencies. ArgumentError: Package Gridap [56d4f2e9-7ea1-5844-9cf6-b9c51ca7ce8e] is required but does not seem to be installed: - Run `Pkg.instantiate()` to install all recorded dependencies. Stacktrace: Stacktrace: [1] [1] __require_prelocked(pkg::Base.PkgId, env::String) @ Base ./loading.jl:2588 [2] _require_prelocked(uuidkey::Base.PkgId, env::String) @ Base ./loading.jl:2466 [3] macro expansion @ ./loading.jl:2394 [inlined] [4] macro expansion @ ./lock.jl:376 [inlined] [5] __require(into::Module, mod::Symbol) @ Base ./loading.jl:2359 [6] require @ ./loading.jl:2335 [inlined] [7] eval_import_path @ ./module.jl:36 [inlined] [8] eval_import_path_all(at::Module, path::Expr, keyword::String) @ Base ./module.jl:60 [9] _eval_using(to::Module, path::Expr) @ Base ./module.jl:137 [10] top-level scope @ ~/.julia/packages/GridapPETSc/X0jXh/test/PLaplacianTests.jl:1 [11] include(mapexpr::Function, mod::Module, _path::String) @ Base ./Base.jl:310 [12]__require_prelocked(pkg::Base.PkgId, env::String) @ Base ./loading.jl:2588 [2] _require_prelocked(uuidkey::Base.PkgId, env::String) @ Base ./loading.jl:2466 [3] macro expansion @ ./loading.jl:2394 [inlined] [4] macro expansion @ ./lock.jl:376 [inlined] [5] __require(into::Module, mod::Symbol) @ Base ./loading.jl:2359 [6] require @ ./loading.jl:2335 [inlined] [7] top-level scope @ ~/.julia/packages/GridapPETSc/X0jXh/test/mpi/PLaplacianTests.jl:1 [13] include(mod::Module, _path::String) @ Base ./Base.jl:309 [14] exec_options(opts::Base.JLOptions) @ Base ./client.jl:324 [15] _start() @ Base ./client.jl:557 eval_import_path @ ./module.jl:36 [inlined] [8] eval_import_path_all(at::Module, path::Expr, keyword::String) @ Base ./module.jl:60 [9] _eval_using(to::Module, path::Expr) @ Base ./module.jl:137 [10] top-level scope @ ~/.julia/packages/GridapPETSc/X0jXh/test/PLaplacianTests.jl:1 [11] include(mapexpr::Function, mod::Module, _path::String) @ Base ./Base.jl:310 [12] top-level scope @ ~/.julia/packages/GridapPETSc/X0jXh/test/mpi/PLaplacianTests.jl:1 [13] include(mod::Module, _path::String) @ Base ./Base.jl:309 [14] exec_options(opts::Base.JLOptions) @ Base ./client.jl:324 [15] _start() @ Base ./client.jl:557 in expression starting at /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/PLaplacianTests.jl:1 in expression starting at /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/mpi/PLaplacianTests.jl:1 in expression starting at /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/PLaplacianTests.jl:1 in expression starting at /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/mpi/PLaplacianTests.jl:1 PLaplacianTests: Error During Test at /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/mpi/runtests.jl:8 Got exception outside of a @test LoadError: failed process: Process(`/home/pkgeval/.julia/artifacts/05d8c79b270470018e9de8dd24ddb6d7954aff9d/bin/mpiexec -n 2 /opt/julia/bin/julia -C native -J/opt/julia/lib/julia/sys.so --depwarn=yes --check-bounds=yes --pkgimages=existing -g1 --startup-file=no --project=/home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/mpi/../.. /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/mpi/PLaplacianTests.jl`, ProcessExited(1)) [1] Stacktrace: [1] pipeline_error @ ./process.jl:602 [inlined] [2] run(::Cmd; wait::Bool) @ Base ./process.jl:517 [3] run @ ./process.jl:514 [inlined] [4] (::Main.GridapPETScTests.GridapPETScMPITests.PLaplacianTestsRun.var"#3#4"{Int64, String, String, String})(cmd::String) @ Main.GridapPETScTests.GridapPETScMPITests.PLaplacianTestsRun ~/.julia/packages/GridapPETSc/X0jXh/test/mpi/mpiexec.jl:11 [5] top-level scope @ ~/.julia/packages/JLLWrappers/GfYNv/src/runtime.jl:49 [6] withenv(::JLLWrappers.var"#withenv_executable_wrapper##0#withenv_executable_wrapper##1"{Main.GridapPETScTests.GridapPETScMPITests.PLaplacianTestsRun.var"#3#4"{Int64, String, String, String}, String}, ::Pair{String, String}, ::Vararg{Pair{String, String}}) @ Base ./env.jl:265 [7] top-level scope @ ~/.julia/packages/JLLWrappers/GfYNv/src/runtime.jl:48 [8] #mpiexec#1 @ ~/.julia/packages/JLLWrappers/GfYNv/src/products/executable_generators.jl:28 [inlined] [9] mpiexec @ ~/.julia/packages/JLLWrappers/GfYNv/src/products/executable_generators.jl:25 [inlined] [10] run_mpi_driver(; procs::Int64, file::String) @ Main.GridapPETScTests.GridapPETScMPITests.PLaplacianTestsRun ~/.julia/packages/GridapPETSc/X0jXh/test/mpi/mpiexec.jl:7 [11] top-level scope @ ~/.julia/packages/GridapPETSc/X0jXh/test/mpi/PLaplacianTestsRun.jl:3 [12] include(mapexpr::Function, mod::Module, _path::String) @ Base ./Base.jl:310 [13] IncludeInto @ ./Base.jl:311 [inlined] [14] macro expansion @ ~/.julia/packages/GridapPETSc/X0jXh/test/mpi/runtests.jl:8 [inlined] [15] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1837 [inlined] [16] macro expansion @ ~/.julia/packages/GridapPETSc/X0jXh/test/mpi/runtests.jl:8 [inlined] [17] macro expansion @ ./timing.jl:645 [inlined] [18] top-level scope @ ~/.julia/packages/GridapPETSc/X0jXh/test/mpi/runtests.jl:353 [19] include(mapexpr::Function, mod::Module, _path::String) @ Base ./Base.jl:310 [20] IncludeInto @ ./Base.jl:311 [inlined] [21] macro expansion @ ~/.julia/packages/GridapPETSc/X0jXh/test/runtests.jl:6 [inlined] [22] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1837 [inlined] [23] macro expansion @ ~/.julia/packages/GridapPETSc/X0jXh/test/runtests.jl:6 [inlined] [24] macro expansion @ ./timing.jl:645 [inlined] [25] top-level scope @ ~/.julia/packages/GridapPETSc/X0jXh/test/runtests.jl:353 [26] include(mapexpr::Function, mod::Module, _path::String) @ Base ./Base.jl:310 [27] top-level scope @ none:6 [28] eval(m::Module, e::Any) @ Core ./boot.jl:489 [29] exec_options(opts::Base.JLOptions) @ Base ./client.jl:290 [30] _start() @ Base ./client.jl:557 in expression starting at /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/mpi/PLaplacianTestsRun.jl:1 15.889980 seconds (80.98 k allocations: 4.446 MiB, 1.63% compilation time) ERROR: ERROR: ERROR: ERROR: LoadError: LoadError: LoadError: LoadError: ArgumentError: Package Gridap [56d4f2e9-7ea1-5844-9cf6-b9c51ca7ce8e] is required but does not seem to be installed: - Run `Pkg.instantiate()` to install all recorded dependencies. ArgumentError: Package Gridap [56d4f2e9-7ea1-5844-9cf6-b9c51ca7ce8e] is required but does not seem to be installed: - Run `Pkg.instantiate()` to install all recorded dependencies. Stacktrace: Stacktrace: ArgumentError: Package Gridap [56d4f2e9-7ea1-5844-9cf6-b9c51ca7ce8e] is required but does not seem to be installed: - Run `Pkg.instantiate()` to install all recorded dependencies. ArgumentError: Stacktrace: Package Gridap [56d4f2e9-7ea1-5844-9cf6-b9c51ca7ce8e] is required but does not seem to be installed: - Run `Pkg.instantiate()` to install all recorded dependencies. Stacktrace: [1] [1] __require_prelocked(pkg::Base.PkgId, env::String) @ Base ./loading.jl:2588 [2] _require_prelocked(uuidkey::Base.PkgId, env::String) @ Base ./loading.jl:2466 [3] macro expansion @ ./loading.jl:2394 [inlined] [4] macro expansion @ ./lock.jl:376 [inlined] [5] __require(into::Module, mod::Symbol) @ Base ./loading.jl:2359 [6] require @ ./loading.jl:2335 [inlined] [7] eval_import_path @ ./module.jl:36 [inlined] [8] __require_prelocked(pkg::Base.PkgId, env::String) @ Base ./loading.jl:2588 [2] _require_prelocked(uuidkey::Base.PkgId, env::String) @ Base ./loading.jl:2466 [3] macro expansion @ ./loading.jl:2394 [inlined] [4] macro expansion @ ./lock.jl:376 [inlined] [5] __require(into::Module, mod::Symbol) @ Base ./loading.jl:2359 [6] require @ ./loading.jl:2335 [inlined] [7] eval_import_path @ ./module.jl:36 [inlined] [8] eval_import_path_all(at::Module, path::Expr, keyword::String) @ Base ./module.jl:60 [9] _eval_using(to::Module, path::Expr) @ Base ./module.jl:137 [10] top-level scope @ ~/.julia/packages/GridapPETSc/X0jXh/test/PLaplacianTests.jl:1 [11] include(mapexpr::Function, mod::Module, _path::String) @ Base ./Base.jl:310 [12] top-level scope @ ~/.julia/packages/GridapPETSc/X0jXh/test/mpi/GCTests.jl:1 [13] include(mod::Module, _path::String) @ Base ./Base.jl:309 [14] exec_options(opts::Base.JLOptions) @ Base ./client.jl:324 [15] _start() @ Base ./client.jl:557eval_import_path_all(at::Module, path::Expr, keyword::String) @ Base ./module.jl:60 [9] _eval_using(to::Module, path::Expr) @ Base ./module.jl:137 [10] top-level scope @ ~/.julia/packages/GridapPETSc/X0jXh/test/PLaplacianTests.jl:1 [11] include(mapexpr::Function, mod::Module, _path::String) @ Base ./Base.jl:310 [12] top-level scope @ ~/.julia/packages/GridapPETSc/X0jXh/test/mpi/GCTests.jl:1 [13] include(mod::Module, _path::String) @ Base ./Base.jl:309 [14] exec_options(opts::Base.JLOptions) @ Base ./client.jl:324 [15] _start() @ Base ./client.jl:557 [1] in expression starting at /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/PLaplacianTests.jl:1 in expression starting at /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/mpi/GCTests.jl:1 [1] __require_prelocked(pkg::Base.PkgId, env::String) @ in expression starting at /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/PLaplacianTests.jl:1 in expression starting at /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/mpi/GCTests.jl:1 Base ./loading.jl:2588 [2] _require_prelocked(uuidkey::Base.PkgId, env::String) @ Base ./loading.jl:2466 [3] macro expansion @ ./loading.jl:2394 [inlined] [4] macro expansion @ ./lock.jl:376 [inlined] [5] __require(into::Module, mod::Symbol) @ Base ./loading.jl:2359 [6] require @ ./loading.jl:2335 [inlined] [7] eval_import_path @ ./module.jl:36 [inlined] [8] eval_import_path_all(at::Module, path::Expr, keyword::String) @ Base ./module.jl:60 [9] _eval_using(to::Module, path::Expr) @ Base ./module.jl:137 [10] top-level scope @ ~/.julia/packages/GridapPETSc/X0jXh/test/PLaplacianTests.jl:1 [11] include(mapexpr::Function, mod::Module, _path::String) @ Base ./Base.jl:310 [12] top-level scope @ ~/.julia/packages/GridapPETSc/X0jXh/test/mpi/GCTests.jl:1 [13] include(mod::Module, _path::String) @ Base ./Base.jl:309 [14] exec_options(opts::Base.JLOptions) @ Base ./client.jl:324 [15] _start() @ Base ./client.jl:557__require_prelocked(pkg::Base.PkgId, env::String) @ Base ./loading.jl:2588 [2] _require_prelocked(uuidkey::Base.PkgId, env::String) @ Base ./loading.jl:2466 [3] macro expansion @ ./loading.jl:2394 [inlined] [4] macro expansion @ ./lock.jl:376 [inlined] [5] __require(into::Module, mod::Symbol) @ Base ./loading.jl:2359 [6] require @ ./loading.jl:2335 [inlined] [7] eval_import_path @ ./module.jl:36 [inlined] [8] eval_import_path_all(at::Module, path::Expr, keyword::String) @ Base ./module.jl:60 [9] _eval_using(to::Module, path::Expr) @ Base ./module.jl:137 [10] top-level scope @ ~/.julia/packages/GridapPETSc/X0jXh/test/PLaplacianTests.jl:1 [11] include(mapexpr::Function, mod::Module, _path::String) @ Base ./Base.jl:310 [12] top-level scope @ ~/.julia/packages/GridapPETSc/X0jXh/test/mpi/GCTests.jl:1 [13] include(mod::Module, _path::String) @ Base ./Base.jl:309 [14] exec_options(opts::Base.JLOptions) @ Base ./client.jl:324 [15] _start() @ Base ./client.jl:557 in expression starting at /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/PLaplacianTests.jl:1 in expression starting at /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/mpi/GCTests.jl:1 in expression starting at /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/PLaplacianTests.jl:1 in expression starting at /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/mpi/GCTests.jl:1 GCTests: Error During Test at /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/mpi/runtests.jl:9 Got exception outside of a @test LoadError: failed process: Process(`/home/pkgeval/.julia/artifacts/05d8c79b270470018e9de8dd24ddb6d7954aff9d/bin/mpiexec -n 4 /opt/julia/bin/julia -C native -J/opt/julia/lib/julia/sys.so --depwarn=yes --check-bounds=yes --pkgimages=existing -g1 --startup-file=no --project=/home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/mpi/../.. /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/mpi/GCTests.jl`, ProcessExited(1)) [1] Stacktrace: [1] pipeline_error @ ./process.jl:602 [inlined] [2] run(::Cmd; wait::Bool) @ Base ./process.jl:517 [3] run @ ./process.jl:514 [inlined] [4] (::Main.GridapPETScTests.GridapPETScMPITests.GCTestsRun.var"#3#4"{Int64, String, String, String})(cmd::String) @ Main.GridapPETScTests.GridapPETScMPITests.GCTestsRun ~/.julia/packages/GridapPETSc/X0jXh/test/mpi/mpiexec.jl:11 [5] top-level scope @ ~/.julia/packages/JLLWrappers/GfYNv/src/runtime.jl:49 [6] withenv(::JLLWrappers.var"#withenv_executable_wrapper##0#withenv_executable_wrapper##1"{Main.GridapPETScTests.GridapPETScMPITests.GCTestsRun.var"#3#4"{Int64, String, String, String}, String}, ::Pair{String, String}, ::Vararg{Pair{String, String}}) @ Base ./env.jl:265 [7] top-level scope @ ~/.julia/packages/JLLWrappers/GfYNv/src/runtime.jl:48 [8] #mpiexec#1 @ ~/.julia/packages/JLLWrappers/GfYNv/src/products/executable_generators.jl:28 [inlined] [9] mpiexec @ ~/.julia/packages/JLLWrappers/GfYNv/src/products/executable_generators.jl:25 [inlined] [10] run_mpi_driver(; procs::Int64, file::String) @ Main.GridapPETScTests.GridapPETScMPITests.GCTestsRun ~/.julia/packages/GridapPETSc/X0jXh/test/mpi/mpiexec.jl:7 [11] top-level scope @ ~/.julia/packages/GridapPETSc/X0jXh/test/mpi/GCTestsRun.jl:3 [12] include(mapexpr::Function, mod::Module, _path::String) @ Base ./Base.jl:310 [13] IncludeInto @ ./Base.jl:311 [inlined] [14] macro expansion @ ~/.julia/packages/GridapPETSc/X0jXh/test/mpi/runtests.jl:9 [inlined] [15] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1837 [inlined] [16] macro expansion @ ~/.julia/packages/GridapPETSc/X0jXh/test/mpi/runtests.jl:9 [inlined] [17] macro expansion @ ./timing.jl:645 [inlined] [18] top-level scope @ ~/.julia/packages/GridapPETSc/X0jXh/test/mpi/runtests.jl:353 [19] include(mapexpr::Function, mod::Module, _path::String) @ Base ./Base.jl:310 [20] IncludeInto @ ./Base.jl:311 [inlined] [21] macro expansion @ ~/.julia/packages/GridapPETSc/X0jXh/test/runtests.jl:6 [inlined] [22] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1837 [inlined] [23] macro expansion @ ~/.julia/packages/GridapPETSc/X0jXh/test/runtests.jl:6 [inlined] [24] macro expansion @ ./timing.jl:645 [inlined] [25] top-level scope @ ~/.julia/packages/GridapPETSc/X0jXh/test/runtests.jl:353 [26] include(mapexpr::Function, mod::Module, _path::String) @ Base ./Base.jl:310 [27] top-level scope @ none:6 [28] eval(m::Module, e::Any) @ Core ./boot.jl:489 [29] exec_options(opts::Base.JLOptions) @ Base ./client.jl:290 [30] _start() @ Base ./client.jl:557 in expression starting at /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/mpi/GCTestsRun.jl:1 30.809978 seconds (80.99 k allocations: 4.448 MiB, 0.84% compilation time) ERROR: ERROR: ERROR: ERROR: LoadError: LoadError: LoadError: LoadError: ArgumentError: Package SparseMatricesCSR [a0a7dd2c-ebf4-11e9-1f05-cf50bc540ca1] is required but does not seem to be installed: - Run `Pkg.instantiate()` to install all recorded dependencies. ArgumentError: Package SparseMatricesCSR [a0a7dd2c-ebf4-11e9-1f05-cf50bc540ca1] is required but does not seem to be installed: - Run `Pkg.instantiate()` to install all recorded dependencies. ArgumentError: Package SparseMatricesCSR [a0a7dd2c-ebf4-11e9-1f05-cf50bc540ca1] is required but does not seem to be installed: - Run `Pkg.instantiate()` to install all recorded dependencies. ArgumentError: Package SparseMatricesCSR [a0a7dd2c-ebf4-11e9-1f05-cf50bc540ca1] is required but does not seem to be installed: - Run `Pkg.instantiate()` to install all recorded dependencies. Stacktrace: Stacktrace: Stacktrace: Stacktrace: [1] [1] [1] [1] __require_prelocked(pkg::Base.PkgId, env::String) @ Base ./loading.jl:2588 [2] _require_prelocked(uuidkey::Base.PkgId, env::String) @ Base ./loading.jl:2466 [3] macro expansion @ ./loading.jl:2394 [inlined] [4] macro expansion @ ./lock.jl:376 [inlined] [5] __require(into::Module, mod::Symbol) @ Base ./loading.jl:2359 [6] require @ ./loading.jl:2335 [inlined] [7] eval_import_path @ ./module.jl:36 [inlined] [8] eval_import_path_all(at::Module, path::Expr, keyword::String) @ Base ./module.jl:60 [9] _eval_using(to::Module, path::Expr) @ Base ./module.jl:137 [10] top-level scope @ ~/.julia/packages/GridapPETSc/X0jXh/test/PoissonTests.jl:1 [11] include(mapexpr::Function, mod::Module, _path::String) @ Base ./Base.jl:310 [12] top-level scope @ ~/.julia/packages/GridapPETSc/X0jXh/test/mpi/PoissonTests.jl:1 [13] include(mod::Module, _path::String) @ Base ./Base.jl:309 [14] exec_options(opts::Base.JLOptions) @ Base ./client.jl:324 [15] _start() @ Base ./client.jl:557__require_prelocked(pkg::Base.PkgId, env::String) @ Base ./loading.jl:2588 [2] _require_prelocked(uuidkey::Base.PkgId, env::String) @ Base ./loading.jl:2466 [3] macro expansion @ ./loading.jl:2394 [inlined] [4] macro expansion @ ./lock.jl:376 [inlined] [5] __require(into::Module, mod::Symbol) @ Base ./loading.jl:2359 [6] require @ ./loading.jl:2335 [inlined] [7] eval_import_path @ ./module.jl:36 [inlined] [8] eval_import_path_all(at::Module, path::Expr, keyword::String) @ Base ./module.jl:60 [9] _eval_using(to::Module, path::Expr) @ Base ./module.jl:137 [10] top-level scope @ ~/.julia/packages/GridapPETSc/X0jXh/test/PoissonTests.jl:1 [11] include(mapexpr::Function, mod::Module, _path::String) @ Base ./Base.jl:310 [12] top-level scope @ ~/.julia/packages/GridapPETSc/X0jXh/test/mpi/PoissonTests.jl:1 [13] include(mod::Module, _path::String) @ Base ./Base.jl:309 [14] exec_options(opts::Base.JLOptions) @ Base ./client.jl:324 [15] _start() @ Base ./client.jl:557__require_prelocked(pkg::Base.PkgId, env::String) @ Base ./loading.jl:2588 [2] _require_prelocked(uuidkey::Base.PkgId, env::String) @ Base ./loading.jl:2466 [3] macro expansion @ ./loading.jl:2394 [inlined] [4] macro expansion @ ./lock.jl:376 [inlined] [5] __require(into::Module, mod::Symbol) @ Base ./loading.jl:2359 [6] require @ ./loading.jl:2335 [inlined] [7] eval_import_path @ ./module.jl:36 [inlined] [8] eval_import_path_all(at::Module, path::Expr, keyword::String) @ Base ./module.jl:60 [9] _eval_using(to::Module, path::Expr) @ Base ./module.jl:137 [10] top-level scope @ ~/.julia/packages/GridapPETSc/X0jXh/test/PoissonTests.jl:1 [11] include(mapexpr::Function, mod::Module, _path::String) @ Base ./Base.jl:310 [12] top-level scope @ ~/.julia/packages/GridapPETSc/X0jXh/test/mpi/PoissonTests.jl:1 [13] include(mod::Module, _path::String) @ Base ./Base.jl:309 [14] exec_options(opts::Base.JLOptions) @ Base ./client.jl:324 [15] _start() @ Base ./client.jl:557__require_prelocked(pkg::Base.PkgId, env::String) @ Base ./loading.jl:2588 [2] _require_prelocked(uuidkey::Base.PkgId, env::String) @ Base ./loading.jl:2466 [3] macro expansion @ ./loading.jl:2394 [inlined] [4] macro expansion @ ./lock.jl:376 [inlined] [5] __require(into::Module, mod::Symbol) @ Base ./loading.jl:2359 [6] require @ ./loading.jl:2335 [inlined] [7] eval_import_path @ ./module.jl:36 [inlined] [8] eval_import_path_all(at::Module, path::Expr, keyword::String) @ Base ./module.jl:60 [9] _eval_using(to::Module, path::Expr) @ Base ./module.jl:137 [10] top-level scope @ ~/.julia/packages/GridapPETSc/X0jXh/test/PoissonTests.jl:1 [11] include(mapexpr::Function, mod::Module, _path::String) @ Base ./Base.jl:310 [12] top-level scope @ ~/.julia/packages/GridapPETSc/X0jXh/test/mpi/PoissonTests.jl:1 [13] include(mod::Module, _path::String) @ Base ./Base.jl:309 [14] exec_options(opts::Base.JLOptions) @ Base ./client.jl:324 [15] _start() @ Base ./client.jl:557 in expression starting at /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/PoissonTests.jl:1 in expression starting at /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/PoissonTests.jl:1 in expression starting at /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/mpi/PoissonTests.jl:1 in expression starting at /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/mpi/PoissonTests.jl:1 in expression starting at /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/PoissonTests.jl:1 in expression starting at /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/mpi/PoissonTests.jl:1 in expression starting at /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/PoissonTests.jl:1 in expression starting at /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/mpi/PoissonTests.jl:1 PoissonTests: Error During Test at /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/mpi/runtests.jl:10 Got exception outside of a @test LoadError: failed process: Process(`/home/pkgeval/.julia/artifacts/05d8c79b270470018e9de8dd24ddb6d7954aff9d/bin/mpiexec -n 4 /opt/julia/bin/julia -C native -J/opt/julia/lib/julia/sys.so --depwarn=yes --check-bounds=yes --pkgimages=existing -g1 --startup-file=no --project=/home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/mpi/../.. /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/mpi/PoissonTests.jl`, ProcessExited(1)) [1] Stacktrace: [1] pipeline_error @ ./process.jl:602 [inlined] [2] run(::Cmd; wait::Bool) @ Base ./process.jl:517 [3] run @ ./process.jl:514 [inlined] [4] (::Main.GridapPETScTests.GridapPETScMPITests.PoissonTestsRun.var"#3#4"{Int64, String, String, String})(cmd::String) @ Main.GridapPETScTests.GridapPETScMPITests.PoissonTestsRun ~/.julia/packages/GridapPETSc/X0jXh/test/mpi/mpiexec.jl:11 [5] top-level scope @ ~/.julia/packages/JLLWrappers/GfYNv/src/runtime.jl:49 [6] withenv(::JLLWrappers.var"#withenv_executable_wrapper##0#withenv_executable_wrapper##1"{Main.GridapPETScTests.GridapPETScMPITests.PoissonTestsRun.var"#3#4"{Int64, String, String, String}, String}, ::Pair{String, String}, ::Vararg{Pair{String, String}}) @ Base ./env.jl:265 [7] top-level scope @ ~/.julia/packages/JLLWrappers/GfYNv/src/runtime.jl:48 [8] #mpiexec#1 @ ~/.julia/packages/JLLWrappers/GfYNv/src/products/executable_generators.jl:28 [inlined] [9] mpiexec @ ~/.julia/packages/JLLWrappers/GfYNv/src/products/executable_generators.jl:25 [inlined] [10] run_mpi_driver(; procs::Int64, file::String) @ Main.GridapPETScTests.GridapPETScMPITests.PoissonTestsRun ~/.julia/packages/GridapPETSc/X0jXh/test/mpi/mpiexec.jl:7 [11] top-level scope @ ~/.julia/packages/GridapPETSc/X0jXh/test/mpi/PoissonTestsRun.jl:3 [12] include(mapexpr::Function, mod::Module, _path::String) @ Base ./Base.jl:310 [13] IncludeInto @ ./Base.jl:311 [inlined] [14] macro expansion @ ~/.julia/packages/GridapPETSc/X0jXh/test/mpi/runtests.jl:10 [inlined] [15] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1837 [inlined] [16] macro expansion @ ~/.julia/packages/GridapPETSc/X0jXh/test/mpi/runtests.jl:10 [inlined] [17] macro expansion @ ./timing.jl:645 [inlined] [18] top-level scope @ ~/.julia/packages/GridapPETSc/X0jXh/test/mpi/runtests.jl:353 [19] include(mapexpr::Function, mod::Module, _path::String) @ Base ./Base.jl:310 [20] IncludeInto @ ./Base.jl:311 [inlined] [21] macro expansion @ ~/.julia/packages/GridapPETSc/X0jXh/test/runtests.jl:6 [inlined] [22] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1837 [inlined] [23] macro expansion @ ~/.julia/packages/GridapPETSc/X0jXh/test/runtests.jl:6 [inlined] [24] macro expansion @ ./timing.jl:645 [inlined] [25] top-level scope @ ~/.julia/packages/GridapPETSc/X0jXh/test/runtests.jl:353 [26] include(mapexpr::Function, mod::Module, _path::String) @ Base ./Base.jl:310 [27] top-level scope @ none:6 [28] eval(m::Module, e::Any) @ Core ./boot.jl:489 [29] exec_options(opts::Base.JLOptions) @ Base ./client.jl:290 [30] _start() @ Base ./client.jl:557 in expression starting at /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/mpi/PoissonTestsRun.jl:1 30.260811 seconds (80.99 k allocations: 4.447 MiB, 0.82% compilation time) ERROR: ERROR: ERROR: ERROR: LoadError: LoadError: LoadError: LoadError: ArgumentError: Package Gridap [56d4f2e9-7ea1-5844-9cf6-b9c51ca7ce8e] is required but does not seem to be installed: - Run `Pkg.instantiate()` to install all recorded dependencies. ArgumentError: Package Gridap [56d4f2e9-7ea1-5844-9cf6-b9c51ca7ce8e] is required but does not seem to be installed: - Run `Pkg.instantiate()` to install all recorded dependencies. ArgumentError: Package Gridap [56d4f2e9-7ea1-5844-9cf6-b9c51ca7ce8e] is required but does not seem to be installed: - Run `Pkg.instantiate()` to install all recorded dependencies. Stacktrace: Stacktrace: Stacktrace: ArgumentError: Package Gridap [56d4f2e9-7ea1-5844-9cf6-b9c51ca7ce8e] is required but does not seem to be installed: - Run `Pkg.instantiate()` to install all recorded dependencies. Stacktrace: [1] [1] [1] [1] __require_prelocked(pkg::Base.PkgId, env::String) @ Base ./loading.jl:2588 [2] _require_prelocked(uuidkey::Base.PkgId, env::String) @ Base ./loading.jl:2466 [3] macro expansion @ ./loading.jl:2394 [inlined] [4] macro expansion @ ./lock.jl:376 [inlined] [5] __require(into::Module, mod::Symbol) @ Base ./loading.jl:2359 [6] require @ ./loading.jl:2335 [inlined] [7] eval_import_path @ ./module.jl:36 [inlined] [8] eval_import_path_all(at::Module, path::Expr, keyword::String) @ Base ./module.jl:60 [9] _eval_using(to::Module, path::Expr) @ Base ./module.jl:137 [10] top-level scope @ ~/.julia/packages/GridapPETSc/X0jXh/test/DarcyTests.jl:1 [11] include(mapexpr::Function, mod::Module, _path::String) @ Base ./Base.jl:310 [12] top-level scope @ ~/.julia/packages/GridapPETSc/X0jXh/test/mpi/DarcyTests.jl:1 [13] include(mod::Module, _path::String) @ Base ./Base.jl:309 [14] exec_options(opts::Base.JLOptions) @ Base ./client.jl:324 [15] _start() @ Base ./client.jl:557__require_prelocked(pkg::Base.PkgId, env::String) @ Base ./loading.jl:2588 [2] _require_prelocked(uuidkey::Base.PkgId, env::String) @ Base ./loading.jl:2466 [3] macro expansion @ ./loading.jl:2394 [inlined] [4] macro expansion @ ./lock.jl:376 [inlined] [5] __require(into::Module, mod::Symbol) @ Base ./loading.jl:2359 [6] require @ ./loading.jl:2335 [inlined] [7] eval_import_path @ ./module.jl:36 [inlined] [8] eval_import_path_all(at::Module, path::Expr, keyword::String) @ Base ./module.jl:60 [9] _eval_using(to::Module, path::Expr) @ Base ./module.jl:137 [10] top-level scope @ ~/.julia/packages/GridapPETSc/X0jXh/test/DarcyTests.jl:1 [11] include(mapexpr::Function, mod::Module, _path::String) @ Base ./Base.jl:310 [12] top-level scope @ ~/.julia/packages/GridapPETSc/X0jXh/test/mpi/DarcyTests.jl:1 [13] include(mod::Module, _path::String) @ Base ./Base.jl:309 [14] exec_options(opts::Base.JLOptions) @ Base ./client.jl:324 [15] _start() @ Base ./client.jl:557__require_prelocked(pkg::Base.PkgId, env::String) @ Base ./loading.jl:2588 [2] _require_prelocked(uuidkey::Base.PkgId, env::String) @ Base ./loading.jl:2466 [3] macro expansion @ ./loading.jl:2394 [inlined] [4] macro expansion @ ./lock.jl:376 [inlined] [5] __require(into::Module, mod::Symbol) @ Base ./loading.jl:2359 [6] require @ ./loading.jl:2335 [inlined] [7] eval_import_path @ ./module.jl:36 [inlined] [8] eval_import_path_all(at::Module, path::Expr, keyword::String) @ Base ./module.jl:60 [9] _eval_using(to::Module, path::Expr) @ Base ./module.jl:137 [10] top-level scope @ ~/.julia/packages/GridapPETSc/X0jXh/test/DarcyTests.jl:1 [11] include(mapexpr::Function, mod::Module, _path::String) @ Base ./Base.jl:310 [12] top-level scope @ ~/.julia/packages/GridapPETSc/X0jXh/test/mpi/DarcyTests.jl:1 [13] include(mod::Module, _path::String) @ Base ./Base.jl:309 [14] exec_options(opts::Base.JLOptions) @ Base ./client.jl:324 [15] _start() @ Base ./client.jl:557 in expression starting at /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/DarcyTests.jl:1 in expression starting at /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/mpi/DarcyTests.jl:1 in expression starting at /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/DarcyTests.jl:1 in expression starting at /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/mpi/DarcyTests.jl:1 in expression starting at /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/DarcyTests.jl:1 in expression starting at /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/mpi/DarcyTests.jl:1 __require_prelocked(pkg::Base.PkgId, env::String) @ Base ./loading.jl:2588 [2] _require_prelocked(uuidkey::Base.PkgId, env::String) @ Base ./loading.jl:2466 [3] macro expansion @ ./loading.jl:2394 [inlined] [4] macro expansion @ ./lock.jl:376 [inlined] [5] __require(into::Module, mod::Symbol) @ Base ./loading.jl:2359 [6] require @ ./loading.jl:2335 [inlined] [7] eval_import_path @ ./module.jl:36 [inlined] [8] eval_import_path_all(at::Module, path::Expr, keyword::String) @ Base ./module.jl:60 [9] _eval_using(to::Module, path::Expr) @ Base ./module.jl:137 [10] top-level scope @ ~/.julia/packages/GridapPETSc/X0jXh/test/DarcyTests.jl:1 [11] include(mapexpr::Function, mod::Module, _path::String) @ Base ./Base.jl:310 [12] top-level scope @ ~/.julia/packages/GridapPETSc/X0jXh/test/mpi/DarcyTests.jl:1 [13] include(mod::Module, _path::String) @ Base ./Base.jl:309 [14] exec_options(opts::Base.JLOptions) @ Base ./client.jl:324 [15] _start() @ Base ./client.jl:557 in expression starting at /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/DarcyTests.jl:1 in expression starting at /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/mpi/DarcyTests.jl:1 DarcyTests: Error During Test at /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/mpi/runtests.jl:11 Got exception outside of a @test LoadError: failed process: Process(`/home/pkgeval/.julia/artifacts/05d8c79b270470018e9de8dd24ddb6d7954aff9d/bin/mpiexec -n 4 /opt/julia/bin/julia -C native -J/opt/julia/lib/julia/sys.so --depwarn=yes --check-bounds=yes --pkgimages=existing -g1 --startup-file=no --project=/home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/mpi/../.. /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/mpi/DarcyTests.jl`, ProcessExited(1)) [1] Stacktrace: [1] pipeline_error @ ./process.jl:602 [inlined] [2] run(::Cmd; wait::Bool) @ Base ./process.jl:517 [3] run @ ./process.jl:514 [inlined] [4] (::Main.GridapPETScTests.GridapPETScMPITests.DarcyTestsRun.var"#3#4"{Int64, String, String, String})(cmd::String) @ Main.GridapPETScTests.GridapPETScMPITests.DarcyTestsRun ~/.julia/packages/GridapPETSc/X0jXh/test/mpi/mpiexec.jl:11 [5] top-level scope @ ~/.julia/packages/JLLWrappers/GfYNv/src/runtime.jl:49 [6] withenv(::JLLWrappers.var"#withenv_executable_wrapper##0#withenv_executable_wrapper##1"{Main.GridapPETScTests.GridapPETScMPITests.DarcyTestsRun.var"#3#4"{Int64, String, String, String}, String}, ::Pair{String, String}, ::Vararg{Pair{String, String}}) @ Base ./env.jl:265 [7] top-level scope @ ~/.julia/packages/JLLWrappers/GfYNv/src/runtime.jl:48 [8] #mpiexec#1 @ ~/.julia/packages/JLLWrappers/GfYNv/src/products/executable_generators.jl:28 [inlined] [9] mpiexec @ ~/.julia/packages/JLLWrappers/GfYNv/src/products/executable_generators.jl:25 [inlined] [10] run_mpi_driver(; procs::Int64, file::String) @ Main.GridapPETScTests.GridapPETScMPITests.DarcyTestsRun ~/.julia/packages/GridapPETSc/X0jXh/test/mpi/mpiexec.jl:7 [11] top-level scope @ ~/.julia/packages/GridapPETSc/X0jXh/test/mpi/DarcyTestsRun.jl:3 [12] include(mapexpr::Function, mod::Module, _path::String) @ Base ./Base.jl:310 [13] IncludeInto @ ./Base.jl:311 [inlined] [14] macro expansion @ ~/.julia/packages/GridapPETSc/X0jXh/test/mpi/runtests.jl:11 [inlined] [15] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1837 [inlined] [16] macro expansion @ ~/.julia/packages/GridapPETSc/X0jXh/test/mpi/runtests.jl:11 [inlined] [17] macro expansion @ ./timing.jl:645 [inlined] [18] top-level scope @ ~/.julia/packages/GridapPETSc/X0jXh/test/mpi/runtests.jl:353 [19] include(mapexpr::Function, mod::Module, _path::String) @ Base ./Base.jl:310 [20] IncludeInto @ ./Base.jl:311 [inlined] [21] macro expansion @ ~/.julia/packages/GridapPETSc/X0jXh/test/runtests.jl:6 [inlined] [22] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1837 [inlined] [23] macro expansion @ ~/.julia/packages/GridapPETSc/X0jXh/test/runtests.jl:6 [inlined] [24] macro expansion @ ./timing.jl:645 [inlined] [25] top-level scope @ ~/.julia/packages/GridapPETSc/X0jXh/test/runtests.jl:353 [26] include(mapexpr::Function, mod::Module, _path::String) @ Base ./Base.jl:310 [27] top-level scope @ none:6 [28] eval(m::Module, e::Any) @ Core ./boot.jl:489 [29] exec_options(opts::Base.JLOptions) @ Base ./client.jl:290 [30] _start() @ Base ./client.jl:557 in expression starting at /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/mpi/DarcyTestsRun.jl:1 31.016833 seconds (81.00 k allocations: 4.450 MiB, 0.82% compilation time) Test Summary: | Error Total Time MPI | 5 5 2m16.6s PartitionedArrays | 1 1 27.6s PLaplacianTests | 1 1 15.9s GCTests | 1 1 30.8s PoissonTests | 1 1 30.3s DarcyTests | 1 1 31.0s RNG of the outermost testset: Random.Xoshiro(0x76fbac2410bca3d8, 0x3df51e98b36531d3, 0xd984740bc4212a1d, 0x4ec1414c87dbf8de, 0x97ba1540b695d9f4) ERROR: LoadError: Some tests did not pass: 0 passed, 0 failed, 5 errored, 0 broken. in expression starting at /home/pkgeval/.julia/packages/GridapPETSc/X0jXh/test/runtests.jl:1 Testing failed after 719.98s ERROR: LoadError: Package GridapPETSc errored during testing Stacktrace: [1] pkgerror(msg::String) @ Pkg.Types /opt/julia/share/julia/stdlib/v1.13/Pkg/src/Types.jl:68 [2] test(ctx::Pkg.Types.Context, pkgs::Vector{PackageSpec}; coverage::Bool, julia_args::Cmd, test_args::Cmd, test_fn::Nothing, force_latest_compatible_version::Bool, allow_earlier_backwards_compatible_versions::Bool, allow_reresolve::Bool) @ Pkg.Operations /opt/julia/share/julia/stdlib/v1.13/Pkg/src/Operations.jl:2421 [3] test @ /opt/julia/share/julia/stdlib/v1.13/Pkg/src/Operations.jl:2276 [inlined] [4] test(ctx::Pkg.Types.Context, pkgs::Vector{PackageSpec}; coverage::Bool, test_fn::Nothing, julia_args::Cmd, test_args::Cmd, force_latest_compatible_version::Bool, allow_earlier_backwards_compatible_versions::Bool, allow_reresolve::Bool, kwargs::@Kwargs{io::IOContext{IO}}) @ Pkg.API /opt/julia/share/julia/stdlib/v1.13/Pkg/src/API.jl:498 [5] test(pkgs::Vector{PackageSpec}; io::IOContext{IO}, kwargs::@Kwargs{julia_args::Cmd}) @ Pkg.API /opt/julia/share/julia/stdlib/v1.13/Pkg/src/API.jl:164 [6] test(pkgs::Vector{String}; kwargs::@Kwargs{julia_args::Cmd}) @ Pkg.API /opt/julia/share/julia/stdlib/v1.13/Pkg/src/API.jl:152 [7] test @ /opt/julia/share/julia/stdlib/v1.13/Pkg/src/API.jl:152 [inlined] [8] #test#81 @ /opt/julia/share/julia/stdlib/v1.13/Pkg/src/API.jl:151 [inlined] [9] top-level scope @ /PkgEval.jl/scripts/evaluate.jl:219 [10] include(mod::Module, _path::String) @ Base ./Base.jl:309 [11] exec_options(opts::Base.JLOptions) @ Base ./client.jl:324 [12] _start() @ Base ./client.jl:557 in expression starting at /PkgEval.jl/scripts/evaluate.jl:210 PkgEval failed after 917.97s: package tests unexpectedly errored