Package evaluation to test GridapPETSc on Julia 1.14.0-DEV.1711 (41ad7d9eeb*) started at 2026-02-12T22:17:53.407 ################################################################################ # Set-up # Installing PkgEval dependencies (TestEnv)... Activating project at `~/.julia/environments/v1.14` Set-up completed after 12.57s ################################################################################ # Installation # Installing GridapPETSc... Resolving package versions... Installed ChunkCodecLibZstd ──────── v1.0.0 Installed MacroTools ─────────────── v0.5.16 Installed LightXML ───────────────── v0.9.3 Installed DataStructures ─────────── v0.19.3 Installed DiffResults ────────────── v1.1.0 Installed PARMETIS_jll ───────────── v4.0.6+2 Installed JLD2 ───────────────────── v0.6.3 Installed ForwardDiff ────────────── v1.3.2 Installed FillArrays ─────────────── v1.16.0 Installed AbstractTrees ──────────── v0.4.5 Installed Bzip2_jll ──────────────── v1.0.9+0 Installed PkgVersion ─────────────── v0.3.3 Installed SimpleUnPack ───────────── v1.1.0 Installed Adapt ──────────────────── v4.4.0 Installed ConstructionBase ───────── v1.6.0 Installed PolynomialBases ────────── v0.4.25 Installed MKL_jll ────────────────── v2025.2.0+0 Installed oneTBB_jll ─────────────── v2022.0.0+1 Installed MPICH_jll ──────────────── v5.0.0+0 Installed MPI ────────────────────── v0.20.16 Installed CodecZlib ──────────────── v0.7.8 Installed OrderedCollections ─────── v1.8.1 Installed TranscodingStreams ─────── v0.11.3 Installed ArrayInterface ─────────── v7.22.0 Installed MicrosoftMPI_jll ───────── v10.1.4+3 Installed SCOTCH_jll ─────────────── v7.0.4+3 Installed HashArrayMappedTries ───── v0.2.0 Installed NearestNeighbors ───────── v0.4.27 Installed RecipesBase ────────────── v1.3.4 Installed Combinatorics ──────────── v1.1.0 Installed Statistics ─────────────── v1.11.1 Installed PrecompileTools ────────── v1.3.3 Installed Xorg_libpciaccess_jll ──── v0.18.1+0 Installed IntelOpenMP_jll ────────── v2025.2.0+0 Installed StatsAPI ───────────────── v1.8.0 Installed CommonSubexpressions ───── v0.3.1 Installed OffsetArrays ───────────── v1.17.0 Installed PETSc_jll ──────────────── v3.18.8+0 Installed FileIO ─────────────────── v1.18.0 Installed PartitionedArrays ──────── v0.3.5 Installed ArgCheck ───────────────── v2.5.0 Installed FastGaussQuadrature ────── v1.1.0 Installed StaticArraysCore ───────── v1.4.4 Installed StaticArrays ───────────── v1.9.16 Installed StructUtils ────────────── v2.6.2 Installed IrrationalConstants ────── v0.2.6 Installed Distances ──────────────── v0.10.12 Installed AbstractFFTs ───────────── v1.5.0 Installed QuadGK ─────────────────── v2.11.2 Installed BSON ───────────────────── v0.3.9 Installed SparseMatricesCSR ──────── v0.6.9 Installed SCALAPACK32_jll ────────── v2.2.1+1 Installed NaNMath ────────────────── v1.1.3 Installed Requires ───────────────── v1.3.1 Installed OpenSpecFun_jll ────────── v0.5.6+0 Installed ArrayLayouts ───────────── v1.12.2 Installed ChunkCodecLibZlib ──────── v1.0.0 Installed NLsolve ────────────────── v4.5.1 Installed LogExpFunctions ────────── v0.3.29 Installed OpenMPI_jll ────────────── v5.0.9+0 Installed DiffRules ──────────────── v1.15.1 Installed Gridap ─────────────────── v0.19.7 Installed Hwloc_jll ──────────────── v2.12.2+0 Installed Parsers ────────────────── v2.8.3 Installed ScopedValues ───────────── v1.5.0 Installed OpenBLAS32_jll ─────────── v0.3.30+0 Installed FiniteDiff ─────────────── v2.29.0 Installed AutoHashEquals ─────────── v2.2.0 Installed PTSCOTCH_jll ───────────── v7.0.4+0 Installed JSON ───────────────────── v1.4.0 Installed XML2_jll ───────────────── v2.13.9+0 Installed Reexport ───────────────── v1.2.2 Installed Setfield ───────────────── v1.1.2 Installed SpecialFunctions ───────── v2.7.1 Installed NLSolversBase ──────────── v7.10.0 Installed Preferences ────────────── v1.5.1 Installed BlockArrays ────────────── v1.9.3 Installed JLLWrappers ────────────── v1.7.1 Installed MPIPreferences ─────────── v0.1.11 Installed Libiconv_jll ───────────── v1.18.0+0 Installed FFTW_jll ───────────────── v3.3.11+0 Installed FFTW ───────────────────── v1.10.0 Installed IterativeSolvers ───────── v0.9.4 Installed WriteVTK ───────────────── v1.21.2 Installed GridapDistributed ──────── v0.4.10 Installed LLVMOpenMP_jll ─────────── v18.1.8+0 Installed XZ_jll ─────────────────── v5.8.2+0 Installed ChunkCodecCore ─────────── v1.0.1 Installed VTKBase ────────────────── v1.0.1 Installed LineSearches ───────────── v7.5.1 Installed SuperLU_DIST_jll ───────── v8.2.1+0 Installed ADTypes ────────────────── v1.21.0 Installed GridapPETSc ────────────── v0.5.6 Installed MPItrampoline_jll ──────── v5.5.4+0 Installed DocStringExtensions ────── v0.9.5 Installed METIS_jll ──────────────── v5.1.3+0 Installed CircularArrays ─────────── v1.4.0 Installed DifferentiationInterface ─ v0.7.16 Installed MUMPS_jll ──────────────── v5.7.3+0 Installing 19 artifacts Installed artifact LLVMOpenMP 661.6 KiB Installed artifact SCOTCH 343.5 KiB Installed artifact METIS 1.1 MiB Installed artifact PARMETIS 573.1 KiB Installed artifact FFTW 2.2 MiB Installed artifact Libiconv 1.9 MiB Installed artifact OpenSpecFun 194.9 KiB Installed artifact Xorg_libpciaccess 24.8 KiB Installed artifact PTSCOTCH 175.3 KiB Installed artifact XML2 2.5 MiB Installed artifact Bzip2 503.5 KiB Installed artifact XZ 1.5 MiB Installed artifact SCALAPACK32 2.3 MiB Installed artifact MPICH 8.8 MiB Installed artifact MUMPS 4.7 MiB Installed artifact Hwloc 3.5 MiB Installed artifact OpenBLAS32 10.0 MiB Installed artifact SuperLU_DIST 1.6 MiB Installed artifact PETSc 144.6 MiB Updating `~/.julia/environments/v1.14/Project.toml` [bcdc36c2] + GridapPETSc v0.5.6 Updating `~/.julia/environments/v1.14/Manifest.toml` [47edcb42] + ADTypes v1.21.0 [621f4979] + AbstractFFTs v1.5.0 [1520ce14] + AbstractTrees v0.4.5 [79e6a3ab] + Adapt v4.4.0 [dce04be8] + ArgCheck v2.5.0 [4fba245c] + ArrayInterface v7.22.0 [4c555306] + ArrayLayouts v1.12.2 [15f4f7f2] + AutoHashEquals v2.2.0 [fbb218c0] + BSON v0.3.9 [8e7c35d0] + BlockArrays v1.9.3 [0b6fb165] + ChunkCodecCore v1.0.1 [4c0bbee4] + ChunkCodecLibZlib v1.0.0 [55437552] + ChunkCodecLibZstd v1.0.0 [7a955b69] + CircularArrays v1.4.0 [944b1d66] + CodecZlib v0.7.8 [861a8166] + Combinatorics v1.1.0 [bbf7d656] + CommonSubexpressions v0.3.1 [187b0558] + ConstructionBase v1.6.0 [864edb3b] + DataStructures v0.19.3 [163ba53b] + DiffResults v1.1.0 [b552c78f] + DiffRules v1.15.1 [a0c0ee7d] + DifferentiationInterface v0.7.16 [b4f34e82] + Distances v0.10.12 [ffbed154] + DocStringExtensions v0.9.5 [7a1cc6ca] + FFTW v1.10.0 [442a2c76] + FastGaussQuadrature v1.1.0 [5789e2e9] + FileIO v1.18.0 [1a297f60] + FillArrays v1.16.0 [6a86dc24] + FiniteDiff v2.29.0 [f6369f11] + ForwardDiff v1.3.2 [56d4f2e9] + Gridap v0.19.7 [f9701e48] + GridapDistributed v0.4.10 [bcdc36c2] + GridapPETSc v0.5.6 [076d061b] + HashArrayMappedTries v0.2.0 [92d709cd] + IrrationalConstants v0.2.6 [42fd0dbc] + IterativeSolvers v0.9.4 [033835bb] + JLD2 v0.6.3 [692b3bcd] + JLLWrappers v1.7.1 [682c06a0] + JSON v1.4.0 [9c8b4983] + LightXML v0.9.3 ⌃ [d3d80556] + LineSearches v7.5.1 [2ab3a3ac] + LogExpFunctions v0.3.29 ⌃ [da04e1cc] + MPI v0.20.16 [3da0fdf6] + MPIPreferences v0.1.11 [1914dd2f] + MacroTools v0.5.16 ⌅ [d41bc354] + NLSolversBase v7.10.0 [2774e3e8] + NLsolve v4.5.1 [77ba4419] + NaNMath v1.1.3 [b8a86587] + NearestNeighbors v0.4.27 [6fe1bfb0] + OffsetArrays v1.17.0 [bac558e1] + OrderedCollections v1.8.1 [69de0a69] + Parsers v2.8.3 ⌅ [5a9dfac6] + PartitionedArrays v0.3.5 [eebad327] + PkgVersion v0.3.3 [c74db56a] + PolynomialBases v0.4.25 [aea7be01] + PrecompileTools v1.3.3 [21216c6a] + Preferences v1.5.1 [1fd47b50] + QuadGK v2.11.2 [3cdcf5f2] + RecipesBase v1.3.4 [189a3867] + Reexport v1.2.2 [ae029012] + Requires v1.3.1 [7e506255] + ScopedValues v1.5.0 [efcf1570] + Setfield v1.1.2 [ce78b400] + SimpleUnPack v1.1.0 [a0a7dd2c] + SparseMatricesCSR v0.6.9 [276daf66] + SpecialFunctions v2.7.1 [90137ffa] + StaticArrays v1.9.16 [1e83bf80] + StaticArraysCore v1.4.4 [10745b16] + Statistics v1.11.1 [82ae8749] + StatsAPI v1.8.0 [ec057cc2] + StructUtils v2.6.2 [3bb67fe8] + TranscodingStreams v0.11.3 [4004b06d] + VTKBase v1.0.1 [64499a7a] + WriteVTK v1.21.2 [6e34b625] + Bzip2_jll v1.0.9+0 [f5851436] + FFTW_jll v3.3.11+0 [e33a78d0] + Hwloc_jll v2.12.2+0 [1d5cc7b8] + IntelOpenMP_jll v2025.2.0+0 [1d63c593] + LLVMOpenMP_jll v18.1.8+0 [94ce4f54] + Libiconv_jll v1.18.0+0 [d00139f3] + METIS_jll v5.1.3+0 [856f044c] + MKL_jll v2025.2.0+0 [7cb0a576] + MPICH_jll v5.0.0+0 [f1f71cc9] + MPItrampoline_jll v5.5.4+0 ⌃ [ca64183c] + MUMPS_jll v5.7.3+0 [9237b28f] + MicrosoftMPI_jll v10.1.4+3 [656ef2d0] + OpenBLAS32_jll v0.3.30+0 [fe0851c0] + OpenMPI_jll v5.0.9+0 [efe28fd5] + OpenSpecFun_jll v0.5.6+0 [b247a4be] + PARMETIS_jll v4.0.6+2 ⌃ [8fa3689e] + PETSc_jll v3.18.8+0 [b3ec0f5a] + PTSCOTCH_jll v7.0.4+0 ⌅ [aabda75e] + SCALAPACK32_jll v2.2.1+1 ⌅ [a8d0f55d] + SCOTCH_jll v7.0.4+3 [9a1356b0] + SuperLU_DIST_jll v8.2.1+0 ⌅ [02c8fc9c] + XML2_jll v2.13.9+0 [ffd25f8a] + XZ_jll v5.8.2+0 [a65dc6b1] + Xorg_libpciaccess_jll v0.18.1+0 [1317d2d5] + oneTBB_jll v2022.0.0+1 [0dad84c5] + ArgTools v1.1.2 [56f22d72] + Artifacts v1.11.0 [2a0f44e3] + Base64 v1.11.0 [ade2ca70] + Dates v1.11.0 [8ba89e20] + Distributed v1.11.0 [f43a241f] + Downloads v1.7.0 [7b1f6079] + FileWatching v1.11.0 [9fa8497b] + Future v1.11.0 [b77e0a4c] + InteractiveUtils v1.11.0 [ac6e5ff7] + JuliaSyntaxHighlighting v1.13.0 [4af54fe1] + LazyArtifacts v1.11.0 [b27032c2] + LibCURL v1.0.0 [76f85450] + LibGit2 v1.11.0 [8f399da3] + Libdl v1.11.0 [37e2e46d] + LinearAlgebra v1.13.0 [56ddb016] + Logging v1.11.0 [d6f4376e] + Markdown v1.11.0 [a63ad114] + Mmap v1.11.0 [ca575930] + NetworkOptions v1.3.0 [44cfe95a] + Pkg v1.14.0 [de0858da] + Printf v1.11.0 [9a3f8284] + Random v1.11.0 [ea8e919c] + SHA v1.0.0 [9e88b42a] + Serialization v1.11.0 [6462fe0b] + Sockets v1.11.0 [2f01184e] + SparseArrays v1.13.0 [f489334b] + StyledStrings v1.13.0 [4607b0f0] + SuiteSparse [fa267f1f] + TOML v1.0.3 [a4e569a6] + Tar v1.10.0 [8dfed614] + Test v1.11.0 [cf7118a7] + UUIDs v1.11.0 [4ec0a83e] + Unicode v1.11.0 [e66e0078] + CompilerSupportLibraries_jll v1.3.0+1 [deac9b47] + LibCURL_jll v8.18.0+0 [e37daf67] + LibGit2_jll v1.9.2+0 [29816b5a] + LibSSH2_jll v1.11.3+1 [14a3606d] + MozillaCACerts_jll v2025.12.2 [4536629a] + OpenBLAS_jll v0.3.30+0 [05823500] + OpenLibm_jll v0.8.7+0 [458c3c95] + OpenSSL_jll v3.5.5+0 [efcefdf7] + PCRE2_jll v10.47.0+0 [bea87d4a] + SuiteSparse_jll v7.10.1+0 [83775a58] + Zlib_jll v1.3.1+2 [3161d3a3] + Zstd_jll v1.5.7+1 [8e850b90] + libblastrampoline_jll v5.15.0+0 [8e850ede] + nghttp2_jll v1.68.0+1 [3f19e933] + p7zip_jll v17.7.0+0 Info Packages marked with ⌃ and ⌅ have new versions available. Those with ⌃ may be upgradable, but those with ⌅ are restricted by compatibility constraints from upgrading. To see why use `status --outdated -m` Building GridapPETSc → `~/.julia/scratchspaces/44cfe95a-1eb2-52ea-b672-e2afdf69b78f/aeb0905e921f122e11669e3f10faa318687aa3e0/build.log` Installation completed after 106.86s ################################################################################ # Precompilation # Precompiling PkgEval dependencies... Precompiling packages... 3610.2 ms ✓ TestEnv 1 dependency successfully precompiled in 4 seconds. 27 already precompiled. Precompiling package dependencies... Precompiling packages... 4572.7 ms ✓ MacroTools 792.9 ms ✓ Reexport 1084.3 ms ✓ Statistics 1179.9 ms ✓ ConstructionBase 2195.5 ms ✓ IrrationalConstants 1000.4 ms ✓ StaticArraysCore 905.2 ms ✓ StatsAPI 1186.0 ms ✓ Requires 1012.3 ms ✓ VTKBase 1302.6 ms ✓ OrderedCollections 956.3 ms ✓ HashArrayMappedTries 1410.4 ms ✓ ADTypes 1149.0 ms ✓ DocStringExtensions 1116.1 ms ✓ AbstractTrees 815.9 ms ✓ SimpleUnPack 2066.6 ms ✓ BSON 1972.0 ms ✓ Combinatorics 974.9 ms ✓ AbstractFFTs 1683.3 ms ✓ OffsetArrays 920.9 ms ✓ ArgCheck 2217.4 ms ✓ FillArrays 1005.9 ms ✓ AutoHashEquals 1300.0 ms ✓ TranscodingStreams 998.7 ms ✓ ChunkCodecCore 1445.2 ms ✓ SparseMatricesCSR 1001.3 ms ✓ NaNMath 1415.1 ms ✓ StructUtils 1170.5 ms ✓ Preferences 4365.4 ms ✓ PkgVersion 1795.6 ms ✓ CommonSubexpressions 1364.1 ms ✓ Statistics → SparseArraysExt 813.6 ms ✓ ConstructionBase → ConstructionBaseLinearAlgebraExt 880.0 ms ✓ DiffResults 1507.3 ms ✓ Distances 975.3 ms ✓ Adapt 8917.2 ms ✓ FileIO 3526.1 ms ✓ DataStructures 913.0 ms ✓ ScopedValues 1894.2 ms ✓ DifferentiationInterface 793.8 ms ✓ ADTypes → ADTypesConstructionBaseExt 1364.4 ms ✓ LogExpFunctions 4268.7 ms ✓ AbstractFFTs → AbstractFFTsTestExt 863.2 ms ✓ CircularArrays 1514.7 ms ✓ FillArrays → FillArraysSparseArraysExt 983.8 ms ✓ FillArrays → FillArraysStatisticsExt 994.7 ms ✓ CodecZlib 1058.9 ms ✓ ChunkCodecLibZlib 1182.5 ms ✓ ChunkCodecLibZstd 1249.9 ms ✓ JLLWrappers 1506.3 ms ✓ MPIPreferences 950.9 ms ✓ PrecompileTools 3137.5 ms ✓ Setfield 1237.3 ms ✓ Distances → DistancesSparseArraysExt 1186.1 ms ✓ ArrayInterface 1299.2 ms ✓ Adapt → AdaptSparseArraysExt 878.7 ms ✓ OffsetArrays → OffsetArraysAdaptExt 2371.0 ms ✓ QuadGK 1342.8 ms ✓ DifferentiationInterface → DifferentiationInterfaceSparseArraysExt 1451.9 ms ✓ METIS_jll 5601.0 ms ✓ IntelOpenMP_jll 1289.8 ms ✓ OpenBLAS32_jll 1469.7 ms ✓ Xorg_libpciaccess_jll 1445.5 ms ✓ Bzip2_jll 1359.2 ms ✓ LLVMOpenMP_jll 1357.3 ms ✓ Libiconv_jll 4315.9 ms ✓ oneTBB_jll 1414.0 ms ✓ XZ_jll 1420.5 ms ✓ FFTW_jll 1210.2 ms ✓ MicrosoftMPI_jll 1409.0 ms ✓ OpenSpecFun_jll 4830.3 ms ✓ MPItrampoline_jll 2711.1 ms ✓ RecipesBase 13876.0 ms ✓ StaticArrays 17155.3 ms ✓ Parsers 46947.5 ms ✓ JLD2 799.1 ms ✓ ArrayInterface → ArrayInterfaceStaticArraysCoreExt 1308.5 ms ✓ ArrayInterface → ArrayInterfaceSparseArraysExt 1497.8 ms ✓ XML2_jll  Downloading artifact: IntelOpenMP  Downloading artifact: oneTBB 10332.4 ms ✓ MKL_jll 1550.0 ms ✓ SCOTCH_jll 4992.1 ms ✓ SpecialFunctions 3852.2 ms ✓ IterativeSolvers 1425.6 ms ✓ StaticArrays → StaticArraysStatisticsExt 1432.1 ms ✓ ConstructionBase → ConstructionBaseStaticArraysExt 1605.5 ms ✓ FillArrays → FillArraysStaticArraysExt 1435.0 ms ✓ Adapt → AdaptStaticArraysExt 1464.9 ms ✓ DifferentiationInterface → DifferentiationInterfaceStaticArraysExt 7384.1 ms ✓ JSON 1349.3 ms ✓ FiniteDiff 1144.4 ms ✓ LightXML 1649.6 ms ✓ Hwloc_jll 6888.9 ms ✓ FFTW 5234.6 ms ✓ FastGaussQuadrature 1347.8 ms ✓ DiffRules 8279.7 ms ✓ NearestNeighbors 23337.1 ms ✓ ArrayLayouts 1519.6 ms ✓ FiniteDiff → FiniteDiffSparseArraysExt 1493.6 ms ✓ FiniteDiff → FiniteDiffStaticArraysExt 1048.9 ms ✓ DifferentiationInterface → DifferentiationInterfaceFiniteDiffExt 1864.0 ms ✓ WriteVTK 1588.6 ms ✓ MPICH_jll 4486.7 ms ✓ OpenMPI_jll 4877.9 ms ✓ PolynomialBases 6929.0 ms ✓ ForwardDiff 3596.8 ms ✓ ArrayLayouts → ArrayLayoutsSparseArraysExt 4438.0 ms ✓ BlockArrays 10607.6 ms ✓ MPI 5233.9 ms ✓ SCALAPACK32_jll 4916.3 ms ✓ PARMETIS_jll 4558.2 ms ✓ PTSCOTCH_jll 1733.0 ms ✓ ForwardDiff → ForwardDiffStaticArraysExt 1459.2 ms ✓ DifferentiationInterface → DifferentiationInterfaceForwardDiffExt 3036.9 ms ✓ BlockArrays → BlockArraysAdaptExt 8053.8 ms ✓ PartitionedArrays 5386.3 ms ✓ SuperLU_DIST_jll 5711.3 ms ✓ MUMPS_jll 1813.4 ms ✓ NLSolversBase 5590.6 ms ✓ PETSc_jll 2427.8 ms ✓ LineSearches 2272.7 ms ✓ NLsolve 48208.2 ms ✓ Gridap WARNING: Constructor for type "FESpace" was extended in `GridapDistributed` without explicit qualification or import.  NOTE: Assumed "FESpace" refers to `.FESpace`. This behavior is deprecated and may differ in future versions.  NOTE: This behavior may have differed in Julia versions prior to 1.12.  Hint: If you intended to create a new generic function of the same name, use `function FESpace end`.  Hint: To silence the warning, qualify `FESpace` as `.FESpace` in the method signature or explicitly `import : FESpace`. 17362.4 ms ✓ GridapDistributed 15254.6 ms ✓ GridapPETSc 123 dependencies successfully precompiled in 467 seconds. 38 already precompiled. 2 dependencies had output during precompilation: ┌ GridapDistributed │ WARNING: Constructor for type "FESpace" was extended in `GridapDistributed` without explicit qualification or import. │ NOTE: Assumed "FESpace" refers to `.FESpace`. This behavior is deprecated and may differ in future versions. │ NOTE: This behavior may have differed in Julia versions prior to 1.12. │ Hint: If you intended to create a new generic function of the same name, use `function FESpace end`. │ Hint: To silence the warning, qualify `FESpace` as `.FESpace` in the method signature or explicitly `import : FESpace`. └ ┌ MKL_jll │ Downloading artifact: IntelOpenMP │ Downloading artifact: oneTBB └ Precompilation completed after 496.47s ################################################################################ # Testing # Testing GridapPETSc Status `/tmp/jl_NebtQJ/Project.toml` [56d4f2e9] Gridap v0.19.7 [f9701e48] GridapDistributed v0.4.10 [bcdc36c2] GridapPETSc v0.5.6 ⌃ [da04e1cc] MPI v0.20.16 ⌅ [5a9dfac6] PartitionedArrays v0.3.5 [a0a7dd2c] SparseMatricesCSR v0.6.9 ⌃ [8fa3689e] PETSc_jll v3.18.8+0 [8f399da3] Libdl v1.11.0 [37e2e46d] LinearAlgebra v1.13.0 [9a3f8284] Random v1.11.0 [2f01184e] SparseArrays v1.13.0 [8dfed614] Test v1.11.0 Status `/tmp/jl_NebtQJ/Manifest.toml` [47edcb42] ADTypes v1.21.0 [621f4979] AbstractFFTs v1.5.0 [1520ce14] AbstractTrees v0.4.5 [79e6a3ab] Adapt v4.4.0 [dce04be8] ArgCheck v2.5.0 [4fba245c] ArrayInterface v7.22.0 [4c555306] ArrayLayouts v1.12.2 [15f4f7f2] AutoHashEquals v2.2.0 [fbb218c0] BSON v0.3.9 [8e7c35d0] BlockArrays v1.9.3 [0b6fb165] ChunkCodecCore v1.0.1 [4c0bbee4] ChunkCodecLibZlib v1.0.0 [55437552] ChunkCodecLibZstd v1.0.0 [7a955b69] CircularArrays v1.4.0 [944b1d66] CodecZlib v0.7.8 [861a8166] Combinatorics v1.1.0 [bbf7d656] CommonSubexpressions v0.3.1 [187b0558] ConstructionBase v1.6.0 [864edb3b] DataStructures v0.19.3 [163ba53b] DiffResults v1.1.0 [b552c78f] DiffRules v1.15.1 [a0c0ee7d] DifferentiationInterface v0.7.16 [b4f34e82] Distances v0.10.12 [ffbed154] DocStringExtensions v0.9.5 [7a1cc6ca] FFTW v1.10.0 [442a2c76] FastGaussQuadrature v1.1.0 [5789e2e9] FileIO v1.18.0 [1a297f60] FillArrays v1.16.0 [6a86dc24] FiniteDiff v2.29.0 [f6369f11] ForwardDiff v1.3.2 [56d4f2e9] Gridap v0.19.7 [f9701e48] GridapDistributed v0.4.10 [bcdc36c2] GridapPETSc v0.5.6 [076d061b] HashArrayMappedTries v0.2.0 [92d709cd] IrrationalConstants v0.2.6 [42fd0dbc] IterativeSolvers v0.9.4 [033835bb] JLD2 v0.6.3 [692b3bcd] JLLWrappers v1.7.1 [682c06a0] JSON v1.4.0 [9c8b4983] LightXML v0.9.3 ⌃ [d3d80556] LineSearches v7.5.1 [2ab3a3ac] LogExpFunctions v0.3.29 ⌃ [da04e1cc] MPI v0.20.16 [3da0fdf6] MPIPreferences v0.1.11 [1914dd2f] MacroTools v0.5.16 ⌅ [d41bc354] NLSolversBase v7.10.0 [2774e3e8] NLsolve v4.5.1 [77ba4419] NaNMath v1.1.3 [b8a86587] NearestNeighbors v0.4.27 [6fe1bfb0] OffsetArrays v1.17.0 [bac558e1] OrderedCollections v1.8.1 [69de0a69] Parsers v2.8.3 ⌅ [5a9dfac6] PartitionedArrays v0.3.5 [eebad327] PkgVersion v0.3.3 [c74db56a] PolynomialBases v0.4.25 [aea7be01] PrecompileTools v1.3.3 [21216c6a] Preferences v1.5.1 [1fd47b50] QuadGK v2.11.2 [3cdcf5f2] RecipesBase v1.3.4 [189a3867] Reexport v1.2.2 [ae029012] Requires v1.3.1 [7e506255] ScopedValues v1.5.0 [efcf1570] Setfield v1.1.2 [ce78b400] SimpleUnPack v1.1.0 [a0a7dd2c] SparseMatricesCSR v0.6.9 [276daf66] SpecialFunctions v2.7.1 [90137ffa] StaticArrays v1.9.16 [1e83bf80] StaticArraysCore v1.4.4 [10745b16] Statistics v1.11.1 [82ae8749] StatsAPI v1.8.0 [ec057cc2] StructUtils v2.6.2 [3bb67fe8] TranscodingStreams v0.11.3 [4004b06d] VTKBase v1.0.1 [64499a7a] WriteVTK v1.21.2 [6e34b625] Bzip2_jll v1.0.9+0 [f5851436] FFTW_jll v3.3.11+0 [e33a78d0] Hwloc_jll v2.12.2+0 [1d5cc7b8] IntelOpenMP_jll v2025.2.0+0 [1d63c593] LLVMOpenMP_jll v18.1.8+0 [94ce4f54] Libiconv_jll v1.18.0+0 [d00139f3] METIS_jll v5.1.3+0 [856f044c] MKL_jll v2025.2.0+0 [7cb0a576] MPICH_jll v5.0.0+0 [f1f71cc9] MPItrampoline_jll v5.5.4+0 ⌃ [ca64183c] MUMPS_jll v5.7.3+0 [9237b28f] MicrosoftMPI_jll v10.1.4+3 [656ef2d0] OpenBLAS32_jll v0.3.30+0 [fe0851c0] OpenMPI_jll v5.0.9+0 [efe28fd5] OpenSpecFun_jll v0.5.6+0 [b247a4be] PARMETIS_jll v4.0.6+2 ⌃ [8fa3689e] PETSc_jll v3.18.8+0 [b3ec0f5a] PTSCOTCH_jll v7.0.4+0 ⌅ [aabda75e] SCALAPACK32_jll v2.2.1+1 ⌅ [a8d0f55d] SCOTCH_jll v7.0.4+3 [9a1356b0] SuperLU_DIST_jll v8.2.1+0 ⌅ [02c8fc9c] XML2_jll v2.13.9+0 [ffd25f8a] XZ_jll v5.8.2+0 [a65dc6b1] Xorg_libpciaccess_jll v0.18.1+0 [1317d2d5] oneTBB_jll v2022.0.0+1 [0dad84c5] ArgTools v1.1.2 [56f22d72] Artifacts v1.11.0 [2a0f44e3] Base64 v1.11.0 [ade2ca70] Dates v1.11.0 [8ba89e20] Distributed v1.11.0 [f43a241f] Downloads v1.7.0 [7b1f6079] FileWatching v1.11.0 [9fa8497b] Future v1.11.0 [b77e0a4c] InteractiveUtils v1.11.0 [ac6e5ff7] JuliaSyntaxHighlighting v1.13.0 [4af54fe1] LazyArtifacts v1.11.0 [b27032c2] LibCURL v1.0.0 [76f85450] LibGit2 v1.11.0 [8f399da3] Libdl v1.11.0 [37e2e46d] LinearAlgebra v1.13.0 [56ddb016] Logging v1.11.0 [d6f4376e] Markdown v1.11.0 [a63ad114] Mmap v1.11.0 [ca575930] NetworkOptions v1.3.0 [44cfe95a] Pkg v1.14.0 [de0858da] Printf v1.11.0 [9a3f8284] Random v1.11.0 [ea8e919c] SHA v1.0.0 [9e88b42a] Serialization v1.11.0 [6462fe0b] Sockets v1.11.0 [2f01184e] SparseArrays v1.13.0 [f489334b] StyledStrings v1.13.0 [4607b0f0] SuiteSparse [fa267f1f] TOML v1.0.3 [a4e569a6] Tar v1.10.0 [8dfed614] Test v1.11.0 [cf7118a7] UUIDs v1.11.0 [4ec0a83e] Unicode v1.11.0 [e66e0078] CompilerSupportLibraries_jll v1.3.0+1 [deac9b47] LibCURL_jll v8.18.0+0 [e37daf67] LibGit2_jll v1.9.2+0 [29816b5a] LibSSH2_jll v1.11.3+1 [14a3606d] MozillaCACerts_jll v2025.12.2 [4536629a] OpenBLAS_jll v0.3.30+0 [05823500] OpenLibm_jll v0.8.7+0 [458c3c95] OpenSSL_jll v3.5.5+0 [efcefdf7] PCRE2_jll v10.47.0+0 [bea87d4a] SuiteSparse_jll v7.10.1+0 [83775a58] Zlib_jll v1.3.1+2 [3161d3a3] Zstd_jll v1.5.7+1 [8e850b90] libblastrampoline_jll v5.15.0+0 [8e850ede] nghttp2_jll v1.68.0+1 [3f19e933] p7zip_jll v17.7.0+0 Info Packages marked with ⌃ and ⌅ have new versions available. Those with ⌃ may be upgradable, but those with ⌅ are restricted by compatibility constraints from upgrading. Testing Running tests... [0] PetscDetermineInitialFPTrap(): Floating point trapping is off by default 0 [0] PetscDeviceInitializeTypeFromOptions_Private(): PetscDeviceType host available, initializing [0] PetscDeviceInitializeTypeFromOptions_Private(): PetscDevice host initialized, default device id 0, view FALSE, init type lazy [0] PetscDeviceInitializeTypeFromOptions_Private(): PetscDeviceType cuda not available [0] PetscDeviceInitializeTypeFromOptions_Private(): PetscDeviceType hip not available [0] PetscDeviceInitializeTypeFromOptions_Private(): PetscDeviceType sycl not available [0] PetscInitialize_Common(): PETSc successfully started: number of processors = 1 [0] PetscGetHostName(): Rejecting domainname, likely is NIS GridapPETSc-primary-6aAH6AIb.(none) [0] PetscInitialize_Common(): Running on machine: GridapPETSc-primary-6aAH6AIb [0] PetscCommDuplicate(): Duplicating a communicator 1140850689 -2080374784 max tags = 1073741823 [0] PetscCommDuplicate(): Using internal PETSc communicator 1140850689 -2080374784 [0] PetscGetHostName(): Rejecting domainname, likely is NIS GridapPETSc-primary-6aAH6AIb.(none) Vec Object: 1 MPI process type: seq 1. 2. 4. 1. [0] PetscCommDuplicate(): Duplicating a communicator 1140850688 -2080374783 max tags = 1073741823 [0] PetscGetHostName(): Rejecting domainname, likely is NIS GridapPETSc-primary-6aAH6AIb.(none) Vec Object: 1 MPI process type: seq 1. 2. 4. 1. Vec Object: 1 MPI process type: seq 20. 40. 4. 60. [0] PetscCommDuplicate(): Using internal PETSc communicator 1140850689 -2080374784 [0] MatAssemblyEnd_SeqAIJ(): Matrix size: 4 X 5; storage space: 10 unneeded,10 used [0] MatAssemblyEnd_SeqAIJ(): Number of mallocs during MatSetValues() is 0 [0] MatAssemblyEnd_SeqAIJ(): Maximum nonzeros in any row is 3 [0] MatCheckCompressedRow(): Found the ratio (num_zerorows 0)/(num_localrows 4) < 0.6. Do not use CompressedRow routines. [0] MatSeqAIJCheckInode(): Found 3 nodes of 4. Limit used: 5. Using Inode routines Mat Object: 1 MPI process type: seqaij row 0: (1, 2.) (3, 3.) (4, 1.) row 1: (1, 6.) (3, 11.) (4, 5.) row 2: (1, 4.) (3, 3.) row 3: (3, 4.) (4, 3.) [0] PetscCommDuplicate(): Using internal PETSc communicator 1140850689 -2080374784 [0] MatAssemblyEnd_SeqAIJ(): Matrix size: 4 X 4; storage space: 0 unneeded,10 used [0] MatAssemblyEnd_SeqAIJ(): Number of mallocs during MatSetValues() is 0 [0] MatAssemblyEnd_SeqAIJ(): Maximum nonzeros in any row is 3 [0] MatCheckCompressedRow(): Found the ratio (num_zerorows 0)/(num_localrows 4) < 0.6. Do not use CompressedRow routines. [0] MatSeqAIJCheckInode(): Found 4 nodes out of 4 rows. Not using Inode routines Mat Object: 1 MPI process type: seqaij row 0: (0, 4.) (1, -2.) row 1: (0, -1.) (1, 6.) (2, -2.) row 2: (1, -1.) (2, 6.) (3, -2.) row 3: (2, -1.) (3, 4.) [0] PetscCommDuplicate(): Using internal PETSc communicator 1140850689 -2080374784 [0] PetscCommDuplicate(): Using internal PETSc communicator 1140850689 -2080374784 [0] PetscCommDuplicate(): Using internal PETSc communicator 1140850689 -2080374784 [0] PCSetUp(): Setting up PC for first time [0] PCSetUp(): Leaving PC with identical preconditioner since operator is unchanged [0] PetscCommDuplicate(): Using internal PETSc communicator 1140850689 -2080374784 [0] PetscCommDuplicate(): Using internal PETSc communicator 1140850689 -2080374784 [0] PCSetUp(): Leaving PC with identical preconditioner since operator is unchanged [0] PCSetUp(): Leaving PC with identical preconditioner since operator is unchanged [0] PCSetUp(): Leaving PC with identical preconditioner since operator is unchanged [0] PCSetUp(): Leaving PC with identical preconditioner since operator is unchanged [0] PCSetUp(): Leaving PC with identical preconditioner since operator is unchanged [0] PCSetUp(): Leaving PC with identical preconditioner since operator is unchanged [0] PCSetUp(): Leaving PC with identical preconditioner since operator is unchanged [0] PCSetUp(): Leaving PC with identical preconditioner since operator is unchanged [0] KSPConvergedDefault(): Linear solver has converged. Residual norm 1.260351938337e-16 is less than relative tolerance 1.000000000000e-05 times initial right hand side norm 1.145643923739e+00 at iteration 4 [0] PCSetUp(): Leaving PC with identical preconditioner since operator is unchanged [0] PCSetUp(): Leaving PC with identical preconditioner since operator is unchanged [0] PCSetUp(): Leaving PC with identical preconditioner since operator is unchanged [0] PCSetUp(): Leaving PC with identical preconditioner since operator is unchanged [0] PCSetUp(): Leaving PC with identical preconditioner since operator is unchanged [0] PCSetUp(): Leaving PC with identical preconditioner since operator is unchanged [0] PCSetUp(): Leaving PC with identical preconditioner since operator is unchanged [0] PCSetUp(): Leaving PC with identical preconditioner since operator is unchanged [0] PCSetUp(): Leaving PC with identical preconditioner since operator is unchanged [0] PCSetUp(): Leaving PC with identical preconditioner since operator is unchanged [0] KSPConvergedDefault(): Linear solver has converged. Residual norm 1.260351938337e-16 is less than relative tolerance 1.000000000000e-05 times initial right hand side norm 1.145643923739e+00 at iteration 4 [0] PCSetUp(): Leaving PC with identical preconditioner since operator is unchanged [0] PCSetUp(): Leaving PC with identical preconditioner since operator is unchanged [0] PCSetUp(): Leaving PC with identical preconditioner since operator is unchanged [0] PCSetUp(): Leaving PC with identical preconditioner since operator is unchanged [0] PCSetUp(): Leaving PC with identical preconditioner since operator is unchanged [0] PCSetUp(): Leaving PC with identical preconditioner since operator is unchanged [0] PCSetUp(): Leaving PC with identical preconditioner since operator is unchanged [0] PCSetUp(): Leaving PC with identical preconditioner since operator is unchanged [0] PCSetUp(): Leaving PC with identical preconditioner since operator is unchanged [0] KSPConvergedDefault(): Linear solver has converged. Residual norm 1.315348017250e-16 is less than relative tolerance 1.000000000000e-05 times initial right hand side norm 1.145643923739e+00 at iteration 4 KSP Object: (p_) 1 MPI process type: gmres restart=30, using Classical (unmodified) Gram-Schmidt Orthogonalization with no iterative refinement happy breakdown tolerance 1e-30 maximum iterations=10000, initial guess is zero tolerances: relative=1e-05, absolute=1e-50, divergence=10000. left preconditioning using PRECONDITIONED norm type for convergence test PC Object: (p_) 1 MPI process type: jacobi type DIAGONAL linear system matrix = precond matrix: Mat Object: 1 MPI process type: seqaij rows=4, cols=4 total: nonzeros=10, allocated nonzeros=0 total number of mallocs used during MatSetValues calls=0 not using I-node routines [0] PetscFinalize(): PetscFinalize() called [0] Petsc_DelViewer(): Removing viewer data attribute in an MPI_Comm -2080374784 [0] Petsc_OuterComm_Attr_Delete_Fn(): Removing reference to PETSc communicator embedded in a user MPI_Comm -2080374784 [0] Petsc_InnerComm_Attr_Delete_Fn(): User MPI_Comm 1140850689 is being unlinked from inner PETSc comm -2080374784 [0] PetscCommDestroy(): Deleting PETSc MPI_Comm -2080374784 [0] Petsc_Counter_Attr_Delete_Fn(): Deleting counter data in an MPI_Comm -2080374784 [0] Petsc_DelViewer(): Removing viewer data attribute in an MPI_Comm -2080374783 [0] Petsc_OuterComm_Attr_Delete_Fn(): Removing reference to PETSc communicator embedded in a user MPI_Comm -2080374783 [0] Petsc_InnerComm_Attr_Delete_Fn(): User MPI_Comm 1140850688 is being unlinked from inner PETSc comm -2080374783 [0] PetscCommDestroy(): Deleting PETSc MPI_Comm -2080374783 [0] Petsc_Counter_Attr_Delete_Fn(): Deleting counter data in an MPI_Comm -2080374783 20.199852 seconds (5.44 M allocations: 320.395 MiB, 6.81% gc time, 44.09% compilation time: 23% of which was recompilation) [0] PetscDetermineInitialFPTrap(): Floating point trapping is off by default 0 [0] PetscDeviceInitializeTypeFromOptions_Private(): PetscDeviceType host available, initializing [0] PetscDeviceInitializeTypeFromOptions_Private(): PetscDevice host initialized, default device id 0, view FALSE, init type lazy [0] PetscDeviceInitializeTypeFromOptions_Private(): PetscDeviceType cuda not available [0] PetscDeviceInitializeTypeFromOptions_Private(): PetscDeviceType hip not available [0] PetscDeviceInitializeTypeFromOptions_Private(): PetscDeviceType sycl not available [0] PetscInitialize_Common(): PETSc successfully started: number of processors = 1 [0] PetscGetHostName(): Rejecting domainname, likely is NIS GridapPETSc-primary-6aAH6AIb.(none) [0] PetscInitialize_Common(): Running on machine: GridapPETSc-primary-6aAH6AIb [0] PetscCommDuplicate(): Duplicating a communicator 1140850689 -2080374783 max tags = 1073741823 [0] PetscCommDuplicate(): Using internal PETSc communicator 1140850689 -2080374783 [0] PetscCommDuplicate(): Using internal PETSc communicator 1140850689 -2080374783 [0] PetscCommDuplicate(): Using internal PETSc communicator 1140850689 -2080374783 [0] PetscCommDuplicate(): Using internal PETSc communicator 1140850689 -2080374783 [0] PetscCommDuplicate(): Using internal PETSc communicator 1140850689 -2080374783 [0] MatAssemblyEnd_SeqAIJ(): Matrix size: 4 X 5; storage space: 20 unneeded,0 used [0] MatAssemblyEnd_SeqAIJ(): Number of mallocs during MatSetValues() is 0 [0] MatAssemblyEnd_SeqAIJ(): Maximum nonzeros in any row is 0 [0] MatCheckCompressedRow(): Found the ratio (num_zerorows 4)/(num_localrows 4) > 0.6. Use CompressedRow routines. [0] MatSeqAIJCheckInode(): Found 1 nodes of 4. Limit used: 5. Using Inode routines [0] MatAssemblyEnd_SeqAIJ(): Matrix size: 4 X 5; storage space: 14 unneeded,1 used [0] MatAssemblyEnd_SeqAIJ(): Number of mallocs during MatSetValues() is 1 [0] MatAssemblyEnd_SeqAIJ(): Maximum nonzeros in any row is 1 [0] MatCheckCompressedRow(): Found the ratio (num_zerorows 3)/(num_localrows 4) > 0.6. Use CompressedRow routines. [0] MatSeqAIJCheckInode(): Found 2 nodes of 4. Limit used: 5. Using Inode routines [0] MatAssemblyEnd_SeqAIJ(): Matrix size: 4 X 5; storage space: 14 unneeded,2 used [0] MatAssemblyEnd_SeqAIJ(): Number of mallocs during MatSetValues() is 1 [0] MatAssemblyEnd_SeqAIJ(): Maximum nonzeros in any row is 1 [0] MatCheckCompressedRow(): Found the ratio (num_zerorows 2)/(num_localrows 4) < 0.6. Do not use CompressedRow routines. [0] MatSeqAIJCheckInode(): Found 4 nodes out of 4 rows. Not using Inode routines 4×5 GridapPETSc.PETScMatrix: 0.0 0.0 5.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 7.0 0.0 0.0 0.0 0.0 0.0 [0] PetscCommDuplicate(): Using internal PETSc communicator 1140850689 -2080374783 [0] MatAssemblyEnd_SeqAIJ(): Matrix size: 3 X 2; storage space: 6 unneeded,0 used [0] MatAssemblyEnd_SeqAIJ(): Number of mallocs during MatSetValues() is 0 [0] MatAssemblyEnd_SeqAIJ(): Maximum nonzeros in any row is 0 [0] MatCheckCompressedRow(): Found the ratio (num_zerorows 3)/(num_localrows 3) > 0.6. Use CompressedRow routines. [0] MatSeqAIJCheckInode(): Found 1 nodes of 3. Limit used: 5. Using Inode routines [0] PetscCommDuplicate(): Using internal PETSc communicator 1140850689 -2080374783 [0] MatAssemblyEnd_SeqAIJ(): Matrix size: 4 X 4; storage space: 0 unneeded,10 used [0] MatAssemblyEnd_SeqAIJ(): Number of mallocs during MatSetValues() is 0 [0] MatAssemblyEnd_SeqAIJ(): Maximum nonzeros in any row is 3 [0] MatCheckCompressedRow(): Found the ratio (num_zerorows 0)/(num_localrows 4) < 0.6. Do not use CompressedRow routines. [0] MatSeqAIJCheckInode(): Found 4 nodes out of 4 rows. Not using Inode routines [0] PetscCommDuplicate(): Using internal PETSc communicator 1140850689 -2080374783 [0] PetscCommDuplicate(): Using internal PETSc communicator 1140850689 -2080374783 [0] MatConvert(): Calling duplicate for initial matrix seqaij 0 1 [0] MatConvert(): Calling duplicate for initial matrix seqaij 0 1 [0] PetscCommDuplicate(): Using internal PETSc communicator 1140850689 -2080374783 [0] PetscCommDuplicate(): Using internal PETSc communicator 1140850689 -2080374783 [0] PetscCommDuplicate(): Using internal PETSc communicator 1140850689 -2080374783 [0] PetscCommDuplicate(): Using internal PETSc communicator 1140850689 -2080374783 [0] PetscCommDuplicate(): Using internal PETSc communicator 1140850689 -2080374783 [0] PetscCommDuplicate(): Using internal PETSc communicator 1140850689 -2080374783 [0] PetscCommDuplicate(): Using internal PETSc communicator 1140850689 -2080374783 [0] PetscCommDuplicate(): Using internal PETSc communicator 1140850689 -2080374783 [0] Petsc_OuterComm_Attr_Delete_Fn(): Removing reference to PETSc communicator embedded in a user MPI_Comm -2080374783 [0] Petsc_InnerComm_Attr_Delete_Fn(): User MPI_Comm 1140850689 is being unlinked from inner PETSc comm -2080374783 [0] PetscCommDestroy(): Deleting PETSc MPI_Comm -2080374783 [0] Petsc_Counter_Attr_Delete_Fn(): Deleting counter data in an MPI_Comm -2080374783 [0] PetscFinalize(): PetscFinalize() called 8.203218 seconds (2.41 M allocations: 144.806 MiB, 7.50% gc time, 92.89% compilation time) 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 KSP Object: 1 MPI process type: gmres restart=30, using Classical (unmodified) Gram-Schmidt Orthogonalization with no iterative refinement happy breakdown tolerance 1e-30 maximum iterations=10000, initial guess is zero tolerances: relative=1e-05, absolute=1e-50, divergence=10000. left preconditioning using DEFAULT norm type for convergence test PC Object: 1 MPI process type: jacobi PC has not been set up so information may be incomplete type DIAGONAL linear system matrix = precond matrix: Mat Object: 1 MPI process type: seqaij rows=4, cols=4 total: nonzeros=10, allocated nonzeros=0 total number of mallocs used during MatSetValues calls=0 not using I-node routines ┌ Warning: 1 objects still not finalized before calling GridapPETSc.Finalize() └ @ GridapPETSc ~/.julia/packages/GridapPETSc/l6eIU/src/Environment.jl:45 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 0 KSP Residual norm 2.000000000000e+00 1 KSP Residual norm 4.002966042487e-16 9.766084 seconds (4.34 M allocations: 262.707 MiB, 11.50% gc time, 88.42% compilation time: 9% of which was recompilation) 0 SNES Function norm 3.605551275464e+00 1 SNES Function norm 4.444444444444e-01 2 SNES Function norm 7.111111111111e-02 3 SNES Function norm 3.936947327951e-03 4 SNES Function norm 1.525925473445e-05 5 SNES Function norm 2.328306437081e-10 6 SNES Function norm 0.000000000000e+00 Nonlinear solve converged due to CONVERGED_FNORM_RELATIVE iterations 6 0 SNES Function norm 0.000000000000e+00 1 SNES Function norm 0.000000000000e+00 Nonlinear solve converged due to CONVERGED_FNORM_RELATIVE iterations 1 0 SNES Function norm 0.000000000000e+00 1 SNES Function norm 0.000000000000e+00 Nonlinear solve converged due to CONVERGED_FNORM_RELATIVE iterations 1 0 SNES Function norm 1.486606874732e-01 1 SNES Function norm 8.402777777778e-03 2 SNES Function norm 6.831067663990e-05 3 SNES Function norm 4.665073682466e-09 4 SNES Function norm 0.000000000000e+00 Nonlinear solve converged due to CONVERGED_FNORM_RELATIVE iterations 4 0 SNES Function norm 0.000000000000e+00 1 SNES Function norm 0.000000000000e+00 Nonlinear solve converged due to CONVERGED_FNORM_RELATIVE iterations 1 0 SNES Function norm 0.000000000000e+00 1 SNES Function norm 0.000000000000e+00 Nonlinear solve converged due to CONVERGED_FNORM_RELATIVE iterations 1 0 SNES Function norm 3.605551275464e+00 1 SNES Function norm 4.444444444444e-01 2 SNES Function norm 7.111111111111e-02 3 SNES Function norm 3.936947327951e-03 4 SNES Function norm 1.525925473445e-05 5 SNES Function norm 2.328306437081e-10 6 SNES Function norm 0.000000000000e+00 Nonlinear solve converged due to CONVERGED_FNORM_RELATIVE iterations 6 2.718302 seconds (1.16 M allocations: 67.043 MiB, 21.82% gc time, 77.56% compilation time) [0] PetscDetermineInitialFPTrap(): Floating point trapping is off by default 0 [0] PetscDeviceInitializeTypeFromOptions_Private(): PetscDeviceType host available, initializing [0] PetscDeviceInitializeTypeFromOptions_Private(): PetscDevice host initialized, default device id 0, view FALSE, init type lazy [0] PetscDeviceInitializeTypeFromOptions_Private(): PetscDeviceType cuda not available [0] PetscDeviceInitializeTypeFromOptions_Private(): PetscDeviceType hip not available [0] PetscDeviceInitializeTypeFromOptions_Private(): PetscDeviceType sycl not available [0] PetscInitialize_Common(): PETSc successfully started: number of processors = 1 [0] PetscGetHostName(): Rejecting domainname, likely is NIS GridapPETSc-primary-6aAH6AIb.(none) [0] PetscInitialize_Common(): Running on machine: GridapPETSc-primary-6aAH6AIb [0] PetscCommDuplicate(): Duplicating a communicator 1140850689 -2080374783 max tags = 1073741823 [0] MatAssemblyEnd_SeqAIJ(): Matrix size: 4 X 3; storage space: 2 unneeded,1 used [0] MatAssemblyEnd_SeqAIJ(): Number of mallocs during MatSetValues() is 0 [0] MatAssemblyEnd_SeqAIJ(): Maximum nonzeros in any row is 1 [0] MatCheckCompressedRow(): Found the ratio (num_zerorows 3)/(num_localrows 4) > 0.6. Use CompressedRow routines. [0] MatSeqAIJCheckInode(): Found 2 nodes of 4. Limit used: 5. Using Inode routines 4×3 GridapPETSc.PETScMatrix: -4.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 [0] PetscCommDuplicate(): Using internal PETSc communicator 1140850689 -2080374783 [0] MatAssemblyEnd_SeqAIJ(): Matrix size: 4 X 3; storage space: 11 unneeded,1 used [0] MatAssemblyEnd_SeqAIJ(): Number of mallocs during MatSetValues() is 0 [0] MatAssemblyEnd_SeqAIJ(): Maximum nonzeros in any row is 1 [0] MatCheckCompressedRow(): Found the ratio (num_zerorows 3)/(num_localrows 4) > 0.6. Use CompressedRow routines. [0] MatSeqAIJCheckInode(): Found 2 nodes of 4. Limit used: 5. Using Inode routines 4×3 GridapPETSc.PETScMatrix: -4.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 [0] PetscCommDuplicate(): Using internal PETSc communicator 1140850689 -2080374783 [0] MatAssemblyEnd_SeqAIJ(): Matrix size: 4 X 3; storage space: 0 unneeded,1 used [0] MatAssemblyEnd_SeqAIJ(): Number of mallocs during MatSetValues() is 0 [0] MatAssemblyEnd_SeqAIJ(): Maximum nonzeros in any row is 1 [0] MatCheckCompressedRow(): Found the ratio (num_zerorows 3)/(num_localrows 4) > 0.6. Use CompressedRow routines. [0] MatSeqAIJCheckInode(): Found 2 nodes of 4. Limit used: 5. Using Inode routines 4×3 GridapPETSc.PETScMatrix: -2.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 [0] MatConvert(): Calling duplicate for initial matrix seqaij 0 1 [0] MatConvert(): Calling duplicate for initial matrix seqaij 0 1 [0] PetscCommDuplicate(): Using internal PETSc communicator 1140850689 -2080374783 4-element GridapPETSc.PETScVector: 6.0 0.0 0.0 1.0 ┌ Warning: 1 objects still not finalized before calling GridapPETSc.Finalize() └ @ GridapPETSc ~/.julia/packages/GridapPETSc/l6eIU/src/Environment.jl:45 [0] PetscFinalize(): PetscFinalize() called 6.868858 seconds (2.01 M allocations: 121.552 MiB, 7.85% gc time, 91.64% compilation time) 0 KSP Residual norm 0.0467654 1 KSP Residual norm 0.0156177 2 KSP Residual norm 0.00341436 3 KSP Residual norm 0.000676358 4 KSP Residual norm 0.000183444 5 KSP Residual norm 0.000117881 6 KSP Residual norm 6.05433e-05 7 KSP Residual norm 3.68918e-05 8 KSP Residual norm 2.7057e-05 9 KSP Residual norm 1.66452e-05 10 KSP Residual norm 1.32895e-05 11 KSP Residual norm 9.7877e-06 12 KSP Residual norm 7.2283e-06 13 KSP Residual norm 5.3931e-06 14 KSP Residual norm 5.57218e-06 15 KSP Residual norm 4.05519e-06 16 KSP Residual norm 3.3327e-06 17 KSP Residual norm 2.37606e-06 18 KSP Residual norm 1.95187e-06 19 KSP Residual norm 1.69663e-06 20 KSP Residual norm 1.55528e-06 21 KSP Residual norm 1.42183e-06 22 KSP Residual norm 1.31173e-06 23 KSP Residual norm 1.09312e-06 24 KSP Residual norm 7.84426e-07 25 KSP Residual norm 5.77822e-07 26 KSP Residual norm 4.9466e-07 27 KSP Residual norm 3.51927e-07 28 KSP Residual norm 3.12006e-07 29 KSP Residual norm 2.68e-07 30 KSP Residual norm 2.60337e-07 31 KSP Residual norm 2.32644e-07 32 KSP Residual norm 2.05178e-07 33 KSP Residual norm 1.9385e-07 34 KSP Residual norm 1.65993e-07 35 KSP Residual norm 1.24799e-07 36 KSP Residual norm 9.29335e-08 37 KSP Residual norm 6.59482e-08 38 KSP Residual norm 4.46886e-08 39 KSP Residual norm 3.31405e-08 40 KSP Residual norm 2.2325e-08 41 KSP Residual norm 1.63814e-08 42 KSP Residual norm 1.26832e-08 43 KSP Residual norm 8.92343e-09 44 KSP Residual norm 7.18674e-09 45 KSP Residual norm 4.92297e-09 46 KSP Residual norm 3.54159e-09 47 KSP Residual norm 2.57601e-09 48 KSP Residual norm 1.96237e-09 49 KSP Residual norm 1.43583e-09 50 KSP Residual norm 7.661e-10 51 KSP Residual norm 4.922e-10 52 KSP Residual norm 3.686e-10 Linear solve converged due to CONVERGED_RTOL iterations 52 139.930251 seconds (111.45 M allocations: 6.107 GiB, 3.73% gc time, 99.05% compilation time: <1% of which was recompilation) At line 5668 of file dana_driver.F (unit = 10) Fortran runtime error: Cannot open file '': No such file or directory Testing failed after 405.64s ERROR: LoadError: Package GridapPETSc errored during testing (exit code: 2) Stacktrace: [1] pkgerror(msg::String) @ Pkg.Types /opt/julia/share/julia/stdlib/v1.14/Pkg/src/Types.jl:68 [2] test(ctx::Pkg.Types.Context, pkgs::Vector{PackageSpec}; coverage::Bool, julia_args::Cmd, test_args::Cmd, test_fn::Nothing, force_latest_compatible_version::Bool, allow_earlier_backwards_compatible_versions::Bool, allow_reresolve::Bool) @ Pkg.Operations /opt/julia/share/julia/stdlib/v1.14/Pkg/src/Operations.jl:3138 [3] test @ /opt/julia/share/julia/stdlib/v1.14/Pkg/src/Operations.jl:3003 [inlined] [4] test(ctx::Pkg.Types.Context, pkgs::Vector{PackageSpec}; coverage::Bool, test_fn::Nothing, julia_args::Cmd, test_args::Cmd, force_latest_compatible_version::Bool, allow_earlier_backwards_compatible_versions::Bool, allow_reresolve::Bool, kwargs::@Kwargs{io::IOContext{IO}}) @ Pkg.API /opt/julia/share/julia/stdlib/v1.14/Pkg/src/API.jl:586 [5] kwcall(::@NamedTuple{julia_args::Cmd, io::IOContext{IO}}, ::typeof(Pkg.API.test), ctx::Pkg.Types.Context, pkgs::Vector{PackageSpec}) @ Pkg.API /opt/julia/share/julia/stdlib/v1.14/Pkg/src/API.jl:562 [6] test(pkgs::Vector{PackageSpec}; io::IOContext{IO}, kwargs::@Kwargs{julia_args::Cmd}) @ Pkg.API /opt/julia/share/julia/stdlib/v1.14/Pkg/src/API.jl:172 [7] kwcall(::@NamedTuple{julia_args::Cmd}, ::typeof(Pkg.API.test), pkgs::Vector{PackageSpec}) @ Pkg.API /opt/julia/share/julia/stdlib/v1.14/Pkg/src/API.jl:161 [8] test(pkgs::Vector{String}; kwargs::@Kwargs{julia_args::Cmd}) @ Pkg.API /opt/julia/share/julia/stdlib/v1.14/Pkg/src/API.jl:160 [9] test @ /opt/julia/share/julia/stdlib/v1.14/Pkg/src/API.jl:160 [inlined] [10] kwcall(::@NamedTuple{julia_args::Cmd}, ::typeof(Pkg.API.test), pkg::String) @ Pkg.API /opt/julia/share/julia/stdlib/v1.14/Pkg/src/API.jl:159 [11] top-level scope @ /PkgEval.jl/scripts/evaluate.jl:223 [12] include(mod::Module, _path::String) @ Base ./Base.jl:309 [13] exec_options(opts::Base.JLOptions) @ Base ./client.jl:344 [14] _start() @ Base ./client.jl:585 in expression starting at /PkgEval.jl/scripts/evaluate.jl:214 PkgEval failed after 1049.09s: package tests unexpectedly errored