Package evaluation of NDTensors on Julia 1.13.0-DEV.974 (7bbb213719*) started at 2025-08-12T22:38:27.752 ################################################################################ # Set-up # Installing PkgEval dependencies (TestEnv)... Set-up completed after 7.24s ################################################################################ # Installation # Installing NDTensors... Resolving package versions... Updating `~/.julia/environments/v1.13/Project.toml` [23ae76d9] + NDTensors v0.4.9 Updating `~/.julia/environments/v1.13/Manifest.toml` [7d9f7c33] + Accessors v0.1.42 [79e6a3ab] + Adapt v4.3.0 [dce04be8] + ArgCheck v2.5.0 [4fba245c] + ArrayInterface v7.19.0 [4c555306] + ArrayLayouts v1.11.2 [198e06fe] + BangBang v0.4.4 [9718e550] + Baselet v0.1.1 [8e7c35d0] + BlockArrays v1.7.0 [f70d9fcc] + CommonWorldInvalidations v1.0.0 [34da2185] + Compat v4.18.0 [a33af91c] + CompositionsBase v0.1.2 [187b0558] + ConstructionBase v1.6.0 [9a962f9c] + DataAPI v1.16.0 [e2d170a0] + DataValueInterfaces v1.0.0 [244e2a9f] + DefineSingletons v0.1.2 [85a47980] + Dictionaries v0.4.5 [da5c29d0] + EllipsisNotation v1.8.0 [e2ba6199] + ExprTools v0.1.10 [e189563c] + ExternalDocstrings v0.1.1 [1a297f60] + FillArrays v1.13.0 [41a02a25] + Folds v0.2.10 [d9f16b24] + Functors v0.5.2 [f0d1745a] + HalfIntegers v1.6.0 [615f187c] + IfElse v0.1.1 [313cdc1a] + Indexing v1.1.1 [22cec73e] + InitialValues v0.3.1 [842dd82b] + InlineStrings v1.4.4 [3587e190] + InverseFunctions v0.1.17 [82899510] + IteratorInterfaceExtensions v1.0.0 [1914dd2f] + MacroTools v0.5.16 [128add7d] + MicroCollections v0.2.0 [23ae76d9] + NDTensors v0.4.9 [bac558e1] + OrderedCollections v1.8.1 [65ce6f38] + PackageExtensionCompat v1.0.2 [aea7be01] + PrecompileTools v1.3.2 [21216c6a] + Preferences v1.5.0 [42d2dcc6] + Referenceables v0.1.3 [ae029012] + Requires v1.3.1 [efcf1570] + Setfield v1.1.2 [699a6c99] + SimpleTraits v0.9.5 [03a91e81] + SplitApplyCombine v1.2.3 [171d559e] + SplittablesBase v0.1.15 [aedffcd0] + Static v1.2.0 [0d7ed370] + StaticArrayInterface v1.8.0 [90137ffa] + StaticArrays v1.9.14 [1e83bf80] + StaticArraysCore v1.4.3 [5e0ebb24] + Strided v2.3.2 [4db3bf67] + StridedViews v0.4.1 [3783bdb8] + TableTraits v1.0.1 [bd369af6] + Tables v1.12.1 [24d252fe] + ThreadedScans v0.1.0 [a759f4b9] + TimerOutputs v0.5.29 [28d57a85] + Transducers v0.4.84 [9d95972d] + TupleTools v1.6.0 ⌅ [7e5a90cf] + TypeParameterAccessors v0.3.10 [409d34a3] + VectorInterface v0.5.0 [56f22d72] + Artifacts v1.11.0 [2a0f44e3] + Base64 v1.11.0 [ade2ca70] + Dates v1.11.0 [8ba89e20] + Distributed v1.11.0 [9fa8497b] + Future v1.11.0 [b77e0a4c] + InteractiveUtils v1.11.0 [ac6e5ff7] + JuliaSyntaxHighlighting v1.12.0 [8f399da3] + Libdl v1.11.0 [37e2e46d] + LinearAlgebra v1.13.0 [56ddb016] + Logging v1.11.0 [d6f4376e] + Markdown v1.11.0 [de0858da] + Printf v1.11.0 [9a3f8284] + Random v1.11.0 [ea8e919c] + SHA v0.7.0 [9e88b42a] + Serialization v1.11.0 [6462fe0b] + Sockets v1.11.0 [2f01184e] + SparseArrays v1.13.0 [f489334b] + StyledStrings v1.11.0 [fa267f1f] + TOML v1.0.3 [8dfed614] + Test v1.11.0 [cf7118a7] + UUIDs v1.11.0 [4ec0a83e] + Unicode v1.11.0 [e66e0078] + CompilerSupportLibraries_jll v1.3.0+1 [4536629a] + OpenBLAS_jll v0.3.29+0 [bea87d4a] + SuiteSparse_jll v7.10.1+0 [8e850b90] + libblastrampoline_jll v5.13.1+0 Info Packages marked with ⌅ have new versions available but compatibility constraints restrict them from upgrading. To see why use `status --outdated -m` Installation completed after 2.74s ################################################################################ # Precompilation # Precompiling PkgEval dependencies... Precompiling package dependencies... Precompilation completed after 69.73s ################################################################################ # Testing # Testing NDTensors Status `/tmp/jl_Cd37ne/Project.toml` [79e6a3ab] Adapt v4.3.0 [4c555306] ArrayLayouts v1.11.2 [8e7c35d0] BlockArrays v1.7.0 [861a8166] Combinatorics v1.0.3 [85a47980] Dictionaries v0.4.5 [da5c29d0] EllipsisNotation v1.8.0 [1a297f60] FillArrays v1.13.0 [46192b85] GPUArraysCore v0.2.0 [27aeb0d3] JLArrays v0.2.0 [dbb5928d] MappedArrays v0.4.2 [23ae76d9] NDTensors v0.4.9 [6fd5a793] Octavian v0.3.29 [1bc83da4] SafeTestsets v0.1.0 [860ef19b] StableRNGs v1.0.3 [4db3bf67] StridedViews v0.4.1 [6aa20fa7] TensorOperations v5.3.0 ⌅ [7e5a90cf] TypeParameterAccessors v0.3.10 [e88e6eb3] Zygote v0.7.10 [37e2e46d] LinearAlgebra v1.13.0 [44cfe95a] Pkg v1.13.0 [9a3f8284] Random v1.11.0 [2f01184e] SparseArrays v1.13.0 [8dfed614] Test v1.11.0 Status `/tmp/jl_Cd37ne/Manifest.toml` [621f4979] AbstractFFTs v1.5.0 [7d9f7c33] Accessors v0.1.42 [79e6a3ab] Adapt v4.3.0 [dce04be8] ArgCheck v2.5.0 [4fba245c] ArrayInterface v7.19.0 [4c555306] ArrayLayouts v1.11.2 [a9b6321e] Atomix v1.1.2 [198e06fe] BangBang v0.4.4 [9718e550] Baselet v0.1.1 [62783981] BitTwiddlingConvenienceFunctions v0.1.6 [8e7c35d0] BlockArrays v1.7.0 [fa961155] CEnum v0.5.0 [2a0fbf3d] CPUSummary v0.2.7 [082447d4] ChainRules v1.72.5 [d360d2e6] ChainRulesCore v1.26.0 [fb6a15b2] CloseOpenIntervals v0.1.13 [861a8166] Combinatorics v1.0.3 [bbf7d656] CommonSubexpressions v0.3.1 [f70d9fcc] CommonWorldInvalidations v1.0.0 [34da2185] Compat v4.18.0 [a33af91c] CompositionsBase v0.1.2 [187b0558] ConstructionBase v1.6.0 [adafc99b] CpuId v0.3.1 [9a962f9c] DataAPI v1.16.0 [e2d170a0] DataValueInterfaces v1.0.0 [244e2a9f] DefineSingletons v0.1.2 [85a47980] Dictionaries v0.4.5 [163ba53b] DiffResults v1.1.0 [b552c78f] DiffRules v1.15.1 [ffbed154] DocStringExtensions v0.9.5 [da5c29d0] EllipsisNotation v1.8.0 [e2ba6199] ExprTools v0.1.10 [e189563c] ExternalDocstrings v0.1.1 [1a297f60] FillArrays v1.13.0 [41a02a25] Folds v0.2.10 [f6369f11] ForwardDiff v1.0.1 [d9f16b24] Functors v0.5.2 [0c68f7d7] GPUArrays v11.2.3 [46192b85] GPUArraysCore v0.2.0 [f0d1745a] HalfIntegers v1.6.0 [076d061b] HashArrayMappedTries v0.2.0 [3e5b6fbb] HostCPUFeatures v0.1.17 [7869d1d1] IRTools v0.4.15 [615f187c] IfElse v0.1.1 [313cdc1a] Indexing v1.1.1 [22cec73e] InitialValues v0.3.1 [842dd82b] InlineStrings v1.4.4 [3587e190] InverseFunctions v0.1.17 [92d709cd] IrrationalConstants v0.2.4 [82899510] IteratorInterfaceExtensions v1.0.0 [27aeb0d3] JLArrays v0.2.0 [692b3bcd] JLLWrappers v1.7.1 [63c18a36] KernelAbstractions v0.9.38 [929cbde3] LLVM v9.4.2 [8ac3fa9e] LRUCache v1.6.2 [10f19ff3] LayoutPointers v0.1.17 [2ab3a3ac] LogExpFunctions v0.3.29 [bdcacae8] LoopVectorization v0.12.172 [1914dd2f] MacroTools v0.5.16 [d125e4d3] ManualMemory v0.1.8 [dbb5928d] MappedArrays v0.4.2 [128add7d] MicroCollections v0.2.0 [23ae76d9] NDTensors v0.4.9 [77ba4419] NaNMath v1.1.3 [6fd5a793] Octavian v0.3.29 [6fe1bfb0] OffsetArrays v1.17.0 [bac558e1] OrderedCollections v1.8.1 [65ce6f38] PackageExtensionCompat v1.0.2 [1d0040c9] PolyesterWeave v0.2.2 [aea7be01] PrecompileTools v1.3.2 [21216c6a] Preferences v1.5.0 [43287f4e] PtrArrays v1.3.0 [c1ae055f] RealDot v0.1.0 [189a3867] Reexport v1.2.2 [42d2dcc6] Referenceables v0.1.3 [ae029012] Requires v1.3.1 [94e857df] SIMDTypes v0.1.0 [476501e8] SLEEFPirates v0.6.43 [1bc83da4] SafeTestsets v0.1.0 [7e506255] ScopedValues v1.4.0 [efcf1570] Setfield v1.1.2 [699a6c99] SimpleTraits v0.9.5 [dc90abb0] SparseInverseSubset v0.1.2 [276daf66] SpecialFunctions v2.5.1 [03a91e81] SplitApplyCombine v1.2.3 [171d559e] SplittablesBase v0.1.15 [860ef19b] StableRNGs v1.0.3 [aedffcd0] Static v1.2.0 [0d7ed370] StaticArrayInterface v1.8.0 [90137ffa] StaticArrays v1.9.14 [1e83bf80] StaticArraysCore v1.4.3 [10745b16] Statistics v1.11.1 [5e0ebb24] Strided v2.3.2 [4db3bf67] StridedViews v0.4.1 [09ab397b] StructArrays v0.7.1 [3783bdb8] TableTraits v1.0.1 [bd369af6] Tables v1.12.1 [6aa20fa7] TensorOperations v5.3.0 [24d252fe] ThreadedScans v0.1.0 [8290d209] ThreadingUtilities v0.5.5 [a759f4b9] TimerOutputs v0.5.29 [28d57a85] Transducers v0.4.84 [9d95972d] TupleTools v1.6.0 ⌅ [7e5a90cf] TypeParameterAccessors v0.3.10 [3a884ed6] UnPack v1.0.2 [013be700] UnsafeAtomics v0.3.0 [409d34a3] VectorInterface v0.5.0 [3d5dd08c] VectorizationBase v0.21.71 [e88e6eb3] Zygote v0.7.10 [700de1a5] ZygoteRules v0.2.7 [dad2f222] LLVMExtra_jll v0.0.37+2 [efe28fd5] OpenSpecFun_jll v0.5.6+0 [0dad84c5] ArgTools v1.1.2 [56f22d72] Artifacts v1.11.0 [2a0f44e3] Base64 v1.11.0 [ade2ca70] Dates v1.11.0 [8ba89e20] Distributed v1.11.0 [f43a241f] Downloads v1.7.0 [7b1f6079] FileWatching v1.11.0 [9fa8497b] Future v1.11.0 [b77e0a4c] InteractiveUtils v1.11.0 [ac6e5ff7] JuliaSyntaxHighlighting v1.12.0 [4af54fe1] LazyArtifacts v1.11.0 [b27032c2] LibCURL v0.6.4 [76f85450] LibGit2 v1.11.0 [8f399da3] Libdl v1.11.0 [37e2e46d] LinearAlgebra v1.13.0 [56ddb016] Logging v1.11.0 [d6f4376e] Markdown v1.11.0 [ca575930] NetworkOptions v1.3.0 [44cfe95a] Pkg v1.13.0 [de0858da] Printf v1.11.0 [9a3f8284] Random v1.11.0 [ea8e919c] SHA v0.7.0 [9e88b42a] Serialization v1.11.0 [6462fe0b] Sockets v1.11.0 [2f01184e] SparseArrays v1.13.0 [f489334b] StyledStrings v1.11.0 [4607b0f0] SuiteSparse [fa267f1f] TOML v1.0.3 [a4e569a6] Tar v1.10.0 [8dfed614] Test v1.11.0 [cf7118a7] UUIDs v1.11.0 [4ec0a83e] Unicode v1.11.0 [e66e0078] CompilerSupportLibraries_jll v1.3.0+1 [deac9b47] LibCURL_jll v8.15.0+1 [e37daf67] LibGit2_jll v1.9.1+0 [29816b5a] LibSSH2_jll v1.11.3+1 [14a3606d] MozillaCACerts_jll v2025.7.15 [4536629a] OpenBLAS_jll v0.3.29+0 [05823500] OpenLibm_jll v0.8.5+0 [458c3c95] OpenSSL_jll v3.5.2+0 [efcefdf7] PCRE2_jll v10.45.0+0 [bea87d4a] SuiteSparse_jll v7.10.1+0 [83775a58] Zlib_jll v1.3.1+2 [3161d3a3] Zstd_jll v1.5.7+1 [8e850b90] libblastrampoline_jll v5.13.1+0 [8e850ede] nghttp2_jll v1.65.0+0 [3f19e933] p7zip_jll v17.5.0+2 Info Packages marked with ⌅ have new versions available but compatibility constraints restrict them from upgrading. Testing Running tests... Running /home/pkgeval/.julia/packages/NDTensors/Lb78J/test/test_blocksparse.jl svd example 1: Error During Test at /home/pkgeval/.julia/packages/NDTensors/Lb78J/test/test_blocksparse.jl:293 Got exception outside of a @test Scalar indexing is disallowed. Invocation of getindex resulted in scalar indexing of a GPU array. This is typically caused by calling an iterating implementation of a method. Such implementations *do not* execute on the GPU, but very slowly on the CPU, and therefore should be avoided. If you want to allow scalar iteration, use `allowscalar` or `@allowscalar` to enable scalar iteration globally or for the operations in question. Stacktrace: [1] error(s::String) @ Base ./error.jl:44 [2] errorscalar(op::String) @ GPUArraysCore ~/.julia/packages/GPUArraysCore/aNaXo/src/GPUArraysCore.jl:151 [3] _assertscalar(op::String, behavior::GPUArraysCore.ScalarIndexing) @ GPUArraysCore ~/.julia/packages/GPUArraysCore/aNaXo/src/GPUArraysCore.jl:124 [4] assertscalar(op::String) @ GPUArraysCore ~/.julia/packages/GPUArraysCore/aNaXo/src/GPUArraysCore.jl:112 [5] getindex @ ~/.julia/packages/GPUArrays/u6tui/src/host/indexing.jl:50 [inlined] [6] scalar_getindex @ ~/.julia/packages/GPUArrays/u6tui/src/host/indexing.jl:36 [inlined] [7] _getindex @ ~/.julia/packages/GPUArrays/u6tui/src/host/indexing.jl:19 [inlined] [8] getindex @ ~/.julia/packages/GPUArrays/u6tui/src/host/indexing.jl:17 [inlined] [9] __matmul2x2_elements(tA::Char, A::JLArrays.JLArray{Float32, 2}) @ LinearAlgebra /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:1205 [10] __matmul2x2_elements @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:1234 [inlined] [11] _matmul2x2_elements @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:1199 [inlined] [12] matmul2x2or3x3_nonzeroalpha!(C::JLArrays.JLArray{Float32, 2}, tA::Char, tB::Char, A::JLArrays.JLArray{Float32, 2}, B::JLArrays.JLArray{Float32, 2}, α::Bool, β::Bool) @ LinearAlgebra /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:490 [13] gemm_wrapper! @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:894 [inlined] [14] _syrk_herk_gemm_wrapper!(C::JLArrays.JLArray{Float32, 2}, tA::Char, tB::Char, A::JLArrays.JLArray{Float32, 2}, B::JLArrays.JLArray{Float32, 2}, α::Bool, β::Bool, ::Val{LinearAlgebra.BlasFlag.SYRK}) @ LinearAlgebra /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:519 [15] generic_matmatmul_wrapper!(C::JLArrays.JLArray{Float32, 2}, tA::Char, tB::Char, A::JLArrays.JLArray{Float32, 2}, B::JLArrays.JLArray{Float32, 2}, α::Bool, β::Bool, val::Val{LinearAlgebra.BlasFlag.SYRK}) @ LinearAlgebra /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:511 [16] _mul! @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:333 [inlined] [17] mul! @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:302 [inlined] [18] mul! @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:270 [inlined] [19] * @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:141 [inlined] [20] svd_recursive(M::JLArrays.JLArray{Float32, 2}; thresh::Float64, north_pass::Int64) @ NDTensors ~/.julia/packages/NDTensors/Lb78J/src/linearalgebra/svd.jl:39 [21] svd_recursive @ ~/.julia/packages/NDTensors/Lb78J/src/linearalgebra/svd.jl:29 [inlined] [22] svd(T::NDTensors.DenseTensor{Float32, 2, Tuple{Int64, Int64}, NDTensors.Dense{Float32, JLArrays.JLArray{Float32, 1}}}; mindim::Nothing, maxdim::Nothing, cutoff::Nothing, use_absolute_cutoff::Nothing, use_relative_cutoff::Nothing, alg::Nothing, min_blockdim::Nothing) @ NDTensors ~/.julia/packages/NDTensors/Lb78J/src/linearalgebra/linearalgebra.jl:108 [23] svd @ ~/.julia/packages/NDTensors/Lb78J/src/linearalgebra/linearalgebra.jl:87 [inlined] [24] svd(T::NDTensors.BlockSparseTensor{Float32, 2, Tuple{Vector{Int64}, Vector{Int64}}, NDTensors.BlockSparse{Float32, JLArrays.JLArray{Float32, 1}, 2}}; min_blockdim::Nothing, mindim::Nothing, maxdim::Nothing, cutoff::Nothing, alg::Nothing, use_absolute_cutoff::Nothing, use_relative_cutoff::Nothing) @ NDTensors ~/.julia/packages/NDTensors/Lb78J/src/blocksparse/linearalgebra.jl:58 [25] svd(T::NDTensors.BlockSparseTensor{Float32, 2, Tuple{Vector{Int64}, Vector{Int64}}, NDTensors.BlockSparse{Float32, JLArrays.JLArray{Float32, 1}, 2}}) @ NDTensors ~/.julia/packages/NDTensors/Lb78J/src/blocksparse/linearalgebra.jl:39 [26] top-level scope @ ~/.julia/packages/NDTensors/Lb78J/test/test_blocksparse.jl:30 [27] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1929 [inlined] [28] macro expansion @ ~/.julia/packages/NDTensors/Lb78J/test/test_blocksparse.jl:287 [inlined] [29] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:2018 [inlined] [30] macro expansion @ ~/.julia/packages/NDTensors/Lb78J/test/test_blocksparse.jl:294 [inlined] [31] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1929 [inlined] [32] macro expansion @ ~/.julia/packages/NDTensors/Lb78J/test/test_blocksparse.jl:296 [inlined] [33] eval(m::Module, e::Any) @ Core ./boot.jl:489 [34] top-level scope @ ~/.julia/packages/NDTensors/Lb78J/test/test_blocksparse.jl:1 [35] include(mapexpr::Function, mod::Module, _path::String) @ Base ./Base.jl:312 [36] top-level scope @ ~/.julia/packages/SafeTestsets/raUNr/src/SafeTestsets.jl:4 [37] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1929 [inlined] [38] macro expansion @ ~/.julia/packages/NDTensors/Lb78J/test/runtests.jl:7 [inlined] [39] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1929 [inlined] [40] macro expansion @ ~/.julia/packages/NDTensors/Lb78J/test/runtests.jl:13 [inlined] [41] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:2018 [inlined] [42] macro expansion @ ~/.julia/packages/NDTensors/Lb78J/test/runtests.jl:15 [inlined] [43] eval(m::Module, e::Any) @ Core ./boot.jl:489 [44] top-level scope @ ~/.julia/packages/NDTensors/Lb78J/test/runtests.jl:28 [45] include(mapexpr::Function, mod::Module, _path::String) @ Base ./Base.jl:312 [46] top-level scope @ none:6 [47] eval(m::Module, e::Any) @ Core ./boot.jl:489 [48] exec_options(opts::Base.JLOptions) @ Base ./client.jl:286 [49] _start() @ Base ./client.jl:553 svd example 2: Error During Test at /home/pkgeval/.julia/packages/NDTensors/Lb78J/test/test_blocksparse.jl:301 Got exception outside of a @test Scalar indexing is disallowed. Invocation of getindex resulted in scalar indexing of a GPU array. This is typically caused by calling an iterating implementation of a method. Such implementations *do not* execute on the GPU, but very slowly on the CPU, and therefore should be avoided. If you want to allow scalar iteration, use `allowscalar` or `@allowscalar` to enable scalar iteration globally or for the operations in question. Stacktrace: [1] error(s::String) @ Base ./error.jl:44 [2] errorscalar(op::String) @ GPUArraysCore ~/.julia/packages/GPUArraysCore/aNaXo/src/GPUArraysCore.jl:151 [3] _assertscalar(op::String, behavior::GPUArraysCore.ScalarIndexing) @ GPUArraysCore ~/.julia/packages/GPUArraysCore/aNaXo/src/GPUArraysCore.jl:124 [4] assertscalar(op::String) @ GPUArraysCore ~/.julia/packages/GPUArraysCore/aNaXo/src/GPUArraysCore.jl:112 [5] getindex @ ~/.julia/packages/GPUArrays/u6tui/src/host/indexing.jl:50 [inlined] [6] scalar_getindex @ ~/.julia/packages/GPUArrays/u6tui/src/host/indexing.jl:36 [inlined] [7] _getindex @ ~/.julia/packages/GPUArrays/u6tui/src/host/indexing.jl:19 [inlined] [8] getindex @ ~/.julia/packages/GPUArrays/u6tui/src/host/indexing.jl:17 [inlined] [9] __matmul2x2_elements(tA::Char, A::JLArrays.JLArray{Float32, 2}) @ LinearAlgebra /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:1205 [10] __matmul2x2_elements @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:1234 [inlined] [11] _matmul2x2_elements @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:1199 [inlined] [12] matmul2x2or3x3_nonzeroalpha!(C::JLArrays.JLArray{Float32, 2}, tA::Char, tB::Char, A::JLArrays.JLArray{Float32, 2}, B::JLArrays.JLArray{Float32, 2}, α::Bool, β::Bool) @ LinearAlgebra /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:490 [13] gemm_wrapper! @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:894 [inlined] [14] _syrk_herk_gemm_wrapper!(C::JLArrays.JLArray{Float32, 2}, tA::Char, tB::Char, A::JLArrays.JLArray{Float32, 2}, B::JLArrays.JLArray{Float32, 2}, α::Bool, β::Bool, ::Val{LinearAlgebra.BlasFlag.SYRK}) @ LinearAlgebra /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:519 [15] generic_matmatmul_wrapper!(C::JLArrays.JLArray{Float32, 2}, tA::Char, tB::Char, A::JLArrays.JLArray{Float32, 2}, B::JLArrays.JLArray{Float32, 2}, α::Bool, β::Bool, val::Val{LinearAlgebra.BlasFlag.SYRK}) @ LinearAlgebra /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:511 [16] _mul! @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:333 [inlined] [17] mul! @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:302 [inlined] [18] mul! @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:270 [inlined] [19] * @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:141 [inlined] [20] svd_recursive(M::JLArrays.JLArray{Float32, 2}; thresh::Float64, north_pass::Int64) @ NDTensors ~/.julia/packages/NDTensors/Lb78J/src/linearalgebra/svd.jl:39 [21] svd_recursive @ ~/.julia/packages/NDTensors/Lb78J/src/linearalgebra/svd.jl:29 [inlined] [22] svd(T::NDTensors.DenseTensor{Float32, 2, Tuple{Int64, Int64}, NDTensors.Dense{Float32, JLArrays.JLArray{Float32, 1}}}; mindim::Nothing, maxdim::Nothing, cutoff::Nothing, use_absolute_cutoff::Nothing, use_relative_cutoff::Nothing, alg::Nothing, min_blockdim::Nothing) @ NDTensors ~/.julia/packages/NDTensors/Lb78J/src/linearalgebra/linearalgebra.jl:108 [23] svd @ ~/.julia/packages/NDTensors/Lb78J/src/linearalgebra/linearalgebra.jl:87 [inlined] [24] svd(T::NDTensors.BlockSparseTensor{Float32, 2, Tuple{Vector{Int64}, Vector{Int64}}, NDTensors.BlockSparse{Float32, JLArrays.JLArray{Float32, 1}, 2}}; min_blockdim::Nothing, mindim::Nothing, maxdim::Nothing, cutoff::Nothing, alg::Nothing, use_absolute_cutoff::Nothing, use_relative_cutoff::Nothing) @ NDTensors ~/.julia/packages/NDTensors/Lb78J/src/blocksparse/linearalgebra.jl:58 [25] svd(T::NDTensors.BlockSparseTensor{Float32, 2, Tuple{Vector{Int64}, Vector{Int64}}, NDTensors.BlockSparse{Float32, JLArrays.JLArray{Float32, 1}, 2}}) @ NDTensors ~/.julia/packages/NDTensors/Lb78J/src/blocksparse/linearalgebra.jl:39 [26] top-level scope @ ~/.julia/packages/NDTensors/Lb78J/test/test_blocksparse.jl:30 [27] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1929 [inlined] [28] macro expansion @ ~/.julia/packages/NDTensors/Lb78J/test/test_blocksparse.jl:287 [inlined] [29] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:2018 [inlined] [30] macro expansion @ ~/.julia/packages/NDTensors/Lb78J/test/test_blocksparse.jl:302 [inlined] [31] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1929 [inlined] [32] macro expansion @ ~/.julia/packages/NDTensors/Lb78J/test/test_blocksparse.jl:304 [inlined] [33] eval(m::Module, e::Any) @ Core ./boot.jl:489 [34] top-level scope @ ~/.julia/packages/NDTensors/Lb78J/test/test_blocksparse.jl:1 [35] include(mapexpr::Function, mod::Module, _path::String) @ Base ./Base.jl:312 [36] top-level scope @ ~/.julia/packages/SafeTestsets/raUNr/src/SafeTestsets.jl:4 [37] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1929 [inlined] [38] macro expansion @ ~/.julia/packages/NDTensors/Lb78J/test/runtests.jl:7 [inlined] [39] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1929 [inlined] [40] macro expansion @ ~/.julia/packages/NDTensors/Lb78J/test/runtests.jl:13 [inlined] [41] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:2018 [inlined] [42] macro expansion @ ~/.julia/packages/NDTensors/Lb78J/test/runtests.jl:15 [inlined] [43] eval(m::Module, e::Any) @ Core ./boot.jl:489 [44] top-level scope @ ~/.julia/packages/NDTensors/Lb78J/test/runtests.jl:28 [45] include(mapexpr::Function, mod::Module, _path::String) @ Base ./Base.jl:312 [46] top-level scope @ none:6 [47] eval(m::Module, e::Any) @ Core ./boot.jl:489 [48] exec_options(opts::Base.JLOptions) @ Base ./client.jl:286 [49] _start() @ Base ./client.jl:553 svd example 3: Error During Test at /home/pkgeval/.julia/packages/NDTensors/Lb78J/test/test_blocksparse.jl:309 Got exception outside of a @test Scalar indexing is disallowed. Invocation of getindex resulted in scalar indexing of a GPU array. This is typically caused by calling an iterating implementation of a method. Such implementations *do not* execute on the GPU, but very slowly on the CPU, and therefore should be avoided. If you want to allow scalar iteration, use `allowscalar` or `@allowscalar` to enable scalar iteration globally or for the operations in question. Stacktrace: [1] error(s::String) @ Base ./error.jl:44 [2] errorscalar(op::String) @ GPUArraysCore ~/.julia/packages/GPUArraysCore/aNaXo/src/GPUArraysCore.jl:151 [3] _assertscalar(op::String, behavior::GPUArraysCore.ScalarIndexing) @ GPUArraysCore ~/.julia/packages/GPUArraysCore/aNaXo/src/GPUArraysCore.jl:124 [4] assertscalar(op::String) @ GPUArraysCore ~/.julia/packages/GPUArraysCore/aNaXo/src/GPUArraysCore.jl:112 [5] getindex @ ~/.julia/packages/GPUArrays/u6tui/src/host/indexing.jl:50 [inlined] [6] scalar_getindex @ ~/.julia/packages/GPUArrays/u6tui/src/host/indexing.jl:36 [inlined] [7] _getindex @ ~/.julia/packages/GPUArrays/u6tui/src/host/indexing.jl:19 [inlined] [8] getindex @ ~/.julia/packages/GPUArrays/u6tui/src/host/indexing.jl:17 [inlined] [9] __matmul2x2_elements(tA::Char, A::JLArrays.JLArray{Float32, 2}) @ LinearAlgebra /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:1205 [10] __matmul2x2_elements @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:1234 [inlined] [11] _matmul2x2_elements @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:1199 [inlined] [12] matmul2x2or3x3_nonzeroalpha!(C::JLArrays.JLArray{Float32, 2}, tA::Char, tB::Char, A::JLArrays.JLArray{Float32, 2}, B::JLArrays.JLArray{Float32, 2}, α::Bool, β::Bool) @ LinearAlgebra /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:490 [13] gemm_wrapper! @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:894 [inlined] [14] _syrk_herk_gemm_wrapper!(C::JLArrays.JLArray{Float32, 2}, tA::Char, tB::Char, A::JLArrays.JLArray{Float32, 2}, B::JLArrays.JLArray{Float32, 2}, α::Bool, β::Bool, ::Val{LinearAlgebra.BlasFlag.SYRK}) @ LinearAlgebra /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:519 [15] generic_matmatmul_wrapper!(C::JLArrays.JLArray{Float32, 2}, tA::Char, tB::Char, A::JLArrays.JLArray{Float32, 2}, B::JLArrays.JLArray{Float32, 2}, α::Bool, β::Bool, val::Val{LinearAlgebra.BlasFlag.SYRK}) @ LinearAlgebra /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:511 [16] _mul! @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:333 [inlined] [17] mul! @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:302 [inlined] [18] mul! @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:270 [inlined] [19] * @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:141 [inlined] [20] svd_recursive(M::JLArrays.JLArray{Float32, 2}; thresh::Float64, north_pass::Int64) @ NDTensors ~/.julia/packages/NDTensors/Lb78J/src/linearalgebra/svd.jl:39 [21] svd_recursive @ ~/.julia/packages/NDTensors/Lb78J/src/linearalgebra/svd.jl:29 [inlined] [22] svd(T::NDTensors.DenseTensor{Float32, 2, Tuple{Int64, Int64}, NDTensors.Dense{Float32, JLArrays.JLArray{Float32, 1}}}; mindim::Nothing, maxdim::Nothing, cutoff::Nothing, use_absolute_cutoff::Nothing, use_relative_cutoff::Nothing, alg::Nothing, min_blockdim::Nothing) @ NDTensors ~/.julia/packages/NDTensors/Lb78J/src/linearalgebra/linearalgebra.jl:108 [23] svd @ ~/.julia/packages/NDTensors/Lb78J/src/linearalgebra/linearalgebra.jl:87 [inlined] [24] svd(T::NDTensors.BlockSparseTensor{Float32, 2, Tuple{Vector{Int64}, Vector{Int64}}, NDTensors.BlockSparse{Float32, JLArrays.JLArray{Float32, 1}, 2}}; min_blockdim::Nothing, mindim::Nothing, maxdim::Nothing, cutoff::Nothing, alg::Nothing, use_absolute_cutoff::Nothing, use_relative_cutoff::Nothing) @ NDTensors ~/.julia/packages/NDTensors/Lb78J/src/blocksparse/linearalgebra.jl:58 [25] svd(T::NDTensors.BlockSparseTensor{Float32, 2, Tuple{Vector{Int64}, Vector{Int64}}, NDTensors.BlockSparse{Float32, JLArrays.JLArray{Float32, 1}, 2}}) @ NDTensors ~/.julia/packages/NDTensors/Lb78J/src/blocksparse/linearalgebra.jl:39 [26] top-level scope @ ~/.julia/packages/NDTensors/Lb78J/test/test_blocksparse.jl:30 [27] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1929 [inlined] [28] macro expansion @ ~/.julia/packages/NDTensors/Lb78J/test/test_blocksparse.jl:287 [inlined] [29] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:2018 [inlined] [30] macro expansion @ ~/.julia/packages/NDTensors/Lb78J/test/test_blocksparse.jl:310 [inlined] [31] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1929 [inlined] [32] macro expansion @ ~/.julia/packages/NDTensors/Lb78J/test/test_blocksparse.jl:312 [inlined] [33] eval(m::Module, e::Any) @ Core ./boot.jl:489 [34] top-level scope @ ~/.julia/packages/NDTensors/Lb78J/test/test_blocksparse.jl:1 [35] include(mapexpr::Function, mod::Module, _path::String) @ Base ./Base.jl:312 [36] top-level scope @ ~/.julia/packages/SafeTestsets/raUNr/src/SafeTestsets.jl:4 [37] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1929 [inlined] [38] macro expansion @ ~/.julia/packages/NDTensors/Lb78J/test/runtests.jl:7 [inlined] [39] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1929 [inlined] [40] macro expansion @ ~/.julia/packages/NDTensors/Lb78J/test/runtests.jl:13 [inlined] [41] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:2018 [inlined] [42] macro expansion @ ~/.julia/packages/NDTensors/Lb78J/test/runtests.jl:15 [inlined] [43] eval(m::Module, e::Any) @ Core ./boot.jl:489 [44] top-level scope @ ~/.julia/packages/NDTensors/Lb78J/test/runtests.jl:28 [45] include(mapexpr::Function, mod::Module, _path::String) @ Base ./Base.jl:312 [46] top-level scope @ none:6 [47] eval(m::Module, e::Any) @ Core ./boot.jl:489 [48] exec_options(opts::Base.JLOptions) @ Base ./client.jl:286 [49] _start() @ Base ./client.jl:553 svd example 1: Error During Test at /home/pkgeval/.julia/packages/NDTensors/Lb78J/test/test_blocksparse.jl:293 Got exception outside of a @test Scalar indexing is disallowed. Invocation of getindex resulted in scalar indexing of a GPU array. This is typically caused by calling an iterating implementation of a method. Such implementations *do not* execute on the GPU, but very slowly on the CPU, and therefore should be avoided. If you want to allow scalar iteration, use `allowscalar` or `@allowscalar` to enable scalar iteration globally or for the operations in question. Stacktrace: [1] error(s::String) @ Base ./error.jl:44 [2] errorscalar(op::String) @ GPUArraysCore ~/.julia/packages/GPUArraysCore/aNaXo/src/GPUArraysCore.jl:151 [3] _assertscalar(op::String, behavior::GPUArraysCore.ScalarIndexing) @ GPUArraysCore ~/.julia/packages/GPUArraysCore/aNaXo/src/GPUArraysCore.jl:124 [4] assertscalar(op::String) @ GPUArraysCore ~/.julia/packages/GPUArraysCore/aNaXo/src/GPUArraysCore.jl:112 [5] getindex @ ~/.julia/packages/GPUArrays/u6tui/src/host/indexing.jl:50 [inlined] [6] scalar_getindex @ ~/.julia/packages/GPUArrays/u6tui/src/host/indexing.jl:36 [inlined] [7] _getindex @ ~/.julia/packages/GPUArrays/u6tui/src/host/indexing.jl:19 [inlined] [8] getindex @ ~/.julia/packages/GPUArrays/u6tui/src/host/indexing.jl:17 [inlined] [9] __matmul2x2_elements(tA::Char, A::JLArrays.JLArray{Float64, 2}) @ LinearAlgebra /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:1205 [10] __matmul2x2_elements @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:1234 [inlined] [11] _matmul2x2_elements @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:1199 [inlined] [12] matmul2x2or3x3_nonzeroalpha!(C::JLArrays.JLArray{Float64, 2}, tA::Char, tB::Char, A::JLArrays.JLArray{Float64, 2}, B::JLArrays.JLArray{Float64, 2}, α::Bool, β::Bool) @ LinearAlgebra /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:490 [13] gemm_wrapper! @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:894 [inlined] [14] _syrk_herk_gemm_wrapper!(C::JLArrays.JLArray{Float64, 2}, tA::Char, tB::Char, A::JLArrays.JLArray{Float64, 2}, B::JLArrays.JLArray{Float64, 2}, α::Bool, β::Bool, ::Val{LinearAlgebra.BlasFlag.SYRK}) @ LinearAlgebra /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:519 [15] generic_matmatmul_wrapper!(C::JLArrays.JLArray{Float64, 2}, tA::Char, tB::Char, A::JLArrays.JLArray{Float64, 2}, B::JLArrays.JLArray{Float64, 2}, α::Bool, β::Bool, val::Val{LinearAlgebra.BlasFlag.SYRK}) @ LinearAlgebra /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:511 [16] _mul! @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:333 [inlined] [17] mul! @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:302 [inlined] [18] mul! @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:270 [inlined] [19] * @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:141 [inlined] [20] svd_recursive(M::JLArrays.JLArray{Float64, 2}; thresh::Float64, north_pass::Int64) @ NDTensors ~/.julia/packages/NDTensors/Lb78J/src/linearalgebra/svd.jl:39 [21] svd_recursive @ ~/.julia/packages/NDTensors/Lb78J/src/linearalgebra/svd.jl:29 [inlined] [22] svd(T::NDTensors.DenseTensor{Float64, 2, Tuple{Int64, Int64}, NDTensors.Dense{Float64, JLArrays.JLArray{Float64, 1}}}; mindim::Nothing, maxdim::Nothing, cutoff::Nothing, use_absolute_cutoff::Nothing, use_relative_cutoff::Nothing, alg::Nothing, min_blockdim::Nothing) @ NDTensors ~/.julia/packages/NDTensors/Lb78J/src/linearalgebra/linearalgebra.jl:108 [23] svd @ ~/.julia/packages/NDTensors/Lb78J/src/linearalgebra/linearalgebra.jl:87 [inlined] [24] svd(T::NDTensors.BlockSparseTensor{Float64, 2, Tuple{Vector{Int64}, Vector{Int64}}, NDTensors.BlockSparse{Float64, JLArrays.JLArray{Float64, 1}, 2}}; min_blockdim::Nothing, mindim::Nothing, maxdim::Nothing, cutoff::Nothing, alg::Nothing, use_absolute_cutoff::Nothing, use_relative_cutoff::Nothing) @ NDTensors ~/.julia/packages/NDTensors/Lb78J/src/blocksparse/linearalgebra.jl:58 [25] svd(T::NDTensors.BlockSparseTensor{Float64, 2, Tuple{Vector{Int64}, Vector{Int64}}, NDTensors.BlockSparse{Float64, JLArrays.JLArray{Float64, 1}, 2}}) @ NDTensors ~/.julia/packages/NDTensors/Lb78J/src/blocksparse/linearalgebra.jl:39 [26] top-level scope @ ~/.julia/packages/NDTensors/Lb78J/test/test_blocksparse.jl:30 [27] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1929 [inlined] [28] macro expansion @ ~/.julia/packages/NDTensors/Lb78J/test/test_blocksparse.jl:287 [inlined] [29] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:2018 [inlined] [30] macro expansion @ ~/.julia/packages/NDTensors/Lb78J/test/test_blocksparse.jl:294 [inlined] [31] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1929 [inlined] [32] macro expansion @ ~/.julia/packages/NDTensors/Lb78J/test/test_blocksparse.jl:296 [inlined] [33] eval(m::Module, e::Any) @ Core ./boot.jl:489 [34] top-level scope @ ~/.julia/packages/NDTensors/Lb78J/test/test_blocksparse.jl:1 [35] include(mapexpr::Function, mod::Module, _path::String) @ Base ./Base.jl:312 [36] top-level scope @ ~/.julia/packages/SafeTestsets/raUNr/src/SafeTestsets.jl:4 [37] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1929 [inlined] [38] macro expansion @ ~/.julia/packages/NDTensors/Lb78J/test/runtests.jl:7 [inlined] [39] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1929 [inlined] [40] macro expansion @ ~/.julia/packages/NDTensors/Lb78J/test/runtests.jl:13 [inlined] [41] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:2018 [inlined] [42] macro expansion @ ~/.julia/packages/NDTensors/Lb78J/test/runtests.jl:15 [inlined] [43] eval(m::Module, e::Any) @ Core ./boot.jl:489 [44] top-level scope @ ~/.julia/packages/NDTensors/Lb78J/test/runtests.jl:28 [45] include(mapexpr::Function, mod::Module, _path::String) @ Base ./Base.jl:312 [46] top-level scope @ none:6 [47] eval(m::Module, e::Any) @ Core ./boot.jl:489 [48] exec_options(opts::Base.JLOptions) @ Base ./client.jl:286 [49] _start() @ Base ./client.jl:553 svd example 2: Error During Test at /home/pkgeval/.julia/packages/NDTensors/Lb78J/test/test_blocksparse.jl:301 Got exception outside of a @test Scalar indexing is disallowed. Invocation of getindex resulted in scalar indexing of a GPU array. This is typically caused by calling an iterating implementation of a method. Such implementations *do not* execute on the GPU, but very slowly on the CPU, and therefore should be avoided. If you want to allow scalar iteration, use `allowscalar` or `@allowscalar` to enable scalar iteration globally or for the operations in question. Stacktrace: [1] error(s::String) @ Base ./error.jl:44 [2] errorscalar(op::String) @ GPUArraysCore ~/.julia/packages/GPUArraysCore/aNaXo/src/GPUArraysCore.jl:151 [3] _assertscalar(op::String, behavior::GPUArraysCore.ScalarIndexing) @ GPUArraysCore ~/.julia/packages/GPUArraysCore/aNaXo/src/GPUArraysCore.jl:124 [4] assertscalar(op::String) @ GPUArraysCore ~/.julia/packages/GPUArraysCore/aNaXo/src/GPUArraysCore.jl:112 [5] getindex @ ~/.julia/packages/GPUArrays/u6tui/src/host/indexing.jl:50 [inlined] [6] scalar_getindex @ ~/.julia/packages/GPUArrays/u6tui/src/host/indexing.jl:36 [inlined] [7] _getindex @ ~/.julia/packages/GPUArrays/u6tui/src/host/indexing.jl:19 [inlined] [8] getindex @ ~/.julia/packages/GPUArrays/u6tui/src/host/indexing.jl:17 [inlined] [9] __matmul2x2_elements(tA::Char, A::JLArrays.JLArray{Float64, 2}) @ LinearAlgebra /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:1205 [10] __matmul2x2_elements @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:1234 [inlined] [11] _matmul2x2_elements @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:1199 [inlined] [12] matmul2x2or3x3_nonzeroalpha!(C::JLArrays.JLArray{Float64, 2}, tA::Char, tB::Char, A::JLArrays.JLArray{Float64, 2}, B::JLArrays.JLArray{Float64, 2}, α::Bool, β::Bool) @ LinearAlgebra /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:490 [13] gemm_wrapper! @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:894 [inlined] [14] _syrk_herk_gemm_wrapper!(C::JLArrays.JLArray{Float64, 2}, tA::Char, tB::Char, A::JLArrays.JLArray{Float64, 2}, B::JLArrays.JLArray{Float64, 2}, α::Bool, β::Bool, ::Val{LinearAlgebra.BlasFlag.SYRK}) @ LinearAlgebra /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:519 [15] generic_matmatmul_wrapper!(C::JLArrays.JLArray{Float64, 2}, tA::Char, tB::Char, A::JLArrays.JLArray{Float64, 2}, B::JLArrays.JLArray{Float64, 2}, α::Bool, β::Bool, val::Val{LinearAlgebra.BlasFlag.SYRK}) @ LinearAlgebra /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:511 [16] _mul! @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:333 [inlined] [17] mul! @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:302 [inlined] [18] mul! @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:270 [inlined] [19] * @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:141 [inlined] [20] svd_recursive(M::JLArrays.JLArray{Float64, 2}; thresh::Float64, north_pass::Int64) @ NDTensors ~/.julia/packages/NDTensors/Lb78J/src/linearalgebra/svd.jl:39 [21] svd_recursive @ ~/.julia/packages/NDTensors/Lb78J/src/linearalgebra/svd.jl:29 [inlined] [22] svd(T::NDTensors.DenseTensor{Float64, 2, Tuple{Int64, Int64}, NDTensors.Dense{Float64, JLArrays.JLArray{Float64, 1}}}; mindim::Nothing, maxdim::Nothing, cutoff::Nothing, use_absolute_cutoff::Nothing, use_relative_cutoff::Nothing, alg::Nothing, min_blockdim::Nothing) @ NDTensors ~/.julia/packages/NDTensors/Lb78J/src/linearalgebra/linearalgebra.jl:108 [23] svd @ ~/.julia/packages/NDTensors/Lb78J/src/linearalgebra/linearalgebra.jl:87 [inlined] [24] svd(T::NDTensors.BlockSparseTensor{Float64, 2, Tuple{Vector{Int64}, Vector{Int64}}, NDTensors.BlockSparse{Float64, JLArrays.JLArray{Float64, 1}, 2}}; min_blockdim::Nothing, mindim::Nothing, maxdim::Nothing, cutoff::Nothing, alg::Nothing, use_absolute_cutoff::Nothing, use_relative_cutoff::Nothing) @ NDTensors ~/.julia/packages/NDTensors/Lb78J/src/blocksparse/linearalgebra.jl:58 [25] svd(T::NDTensors.BlockSparseTensor{Float64, 2, Tuple{Vector{Int64}, Vector{Int64}}, NDTensors.BlockSparse{Float64, JLArrays.JLArray{Float64, 1}, 2}}) @ NDTensors ~/.julia/packages/NDTensors/Lb78J/src/blocksparse/linearalgebra.jl:39 [26] top-level scope @ ~/.julia/packages/NDTensors/Lb78J/test/test_blocksparse.jl:30 [27] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1929 [inlined] [28] macro expansion @ ~/.julia/packages/NDTensors/Lb78J/test/test_blocksparse.jl:287 [inlined] [29] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:2018 [inlined] [30] macro expansion @ ~/.julia/packages/NDTensors/Lb78J/test/test_blocksparse.jl:302 [inlined] [31] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1929 [inlined] [32] macro expansion @ ~/.julia/packages/NDTensors/Lb78J/test/test_blocksparse.jl:304 [inlined] [33] eval(m::Module, e::Any) @ Core ./boot.jl:489 [34] top-level scope @ ~/.julia/packages/NDTensors/Lb78J/test/test_blocksparse.jl:1 [35] include(mapexpr::Function, mod::Module, _path::String) @ Base ./Base.jl:312 [36] top-level scope @ ~/.julia/packages/SafeTestsets/raUNr/src/SafeTestsets.jl:4 [37] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1929 [inlined] [38] macro expansion @ ~/.julia/packages/NDTensors/Lb78J/test/runtests.jl:7 [inlined] [39] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1929 [inlined] [40] macro expansion @ ~/.julia/packages/NDTensors/Lb78J/test/runtests.jl:13 [inlined] [41] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:2018 [inlined] [42] macro expansion @ ~/.julia/packages/NDTensors/Lb78J/test/runtests.jl:15 [inlined] [43] eval(m::Module, e::Any) @ Core ./boot.jl:489 [44] top-level scope @ ~/.julia/packages/NDTensors/Lb78J/test/runtests.jl:28 [45] include(mapexpr::Function, mod::Module, _path::String) @ Base ./Base.jl:312 [46] top-level scope @ none:6 [47] eval(m::Module, e::Any) @ Core ./boot.jl:489 [48] exec_options(opts::Base.JLOptions) @ Base ./client.jl:286 [49] _start() @ Base ./client.jl:553 svd example 3: Error During Test at /home/pkgeval/.julia/packages/NDTensors/Lb78J/test/test_blocksparse.jl:309 Got exception outside of a @test Scalar indexing is disallowed. Invocation of getindex resulted in scalar indexing of a GPU array. This is typically caused by calling an iterating implementation of a method. Such implementations *do not* execute on the GPU, but very slowly on the CPU, and therefore should be avoided. If you want to allow scalar iteration, use `allowscalar` or `@allowscalar` to enable scalar iteration globally or for the operations in question. Stacktrace: [1] error(s::String) @ Base ./error.jl:44 [2] errorscalar(op::String) @ GPUArraysCore ~/.julia/packages/GPUArraysCore/aNaXo/src/GPUArraysCore.jl:151 [3] _assertscalar(op::String, behavior::GPUArraysCore.ScalarIndexing) @ GPUArraysCore ~/.julia/packages/GPUArraysCore/aNaXo/src/GPUArraysCore.jl:124 [4] assertscalar(op::String) @ GPUArraysCore ~/.julia/packages/GPUArraysCore/aNaXo/src/GPUArraysCore.jl:112 [5] getindex @ ~/.julia/packages/GPUArrays/u6tui/src/host/indexing.jl:50 [inlined] [6] scalar_getindex @ ~/.julia/packages/GPUArrays/u6tui/src/host/indexing.jl:36 [inlined] [7] _getindex @ ~/.julia/packages/GPUArrays/u6tui/src/host/indexing.jl:19 [inlined] [8] getindex @ ~/.julia/packages/GPUArrays/u6tui/src/host/indexing.jl:17 [inlined] [9] __matmul2x2_elements(tA::Char, A::JLArrays.JLArray{Float64, 2}) @ LinearAlgebra /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:1205 [10] __matmul2x2_elements @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:1234 [inlined] [11] _matmul2x2_elements @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:1199 [inlined] [12] matmul2x2or3x3_nonzeroalpha!(C::JLArrays.JLArray{Float64, 2}, tA::Char, tB::Char, A::JLArrays.JLArray{Float64, 2}, B::JLArrays.JLArray{Float64, 2}, α::Bool, β::Bool) @ LinearAlgebra /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:490 [13] gemm_wrapper! @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:894 [inlined] [14] _syrk_herk_gemm_wrapper!(C::JLArrays.JLArray{Float64, 2}, tA::Char, tB::Char, A::JLArrays.JLArray{Float64, 2}, B::JLArrays.JLArray{Float64, 2}, α::Bool, β::Bool, ::Val{LinearAlgebra.BlasFlag.SYRK}) @ LinearAlgebra /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:519 [15] generic_matmatmul_wrapper!(C::JLArrays.JLArray{Float64, 2}, tA::Char, tB::Char, A::JLArrays.JLArray{Float64, 2}, B::JLArrays.JLArray{Float64, 2}, α::Bool, β::Bool, val::Val{LinearAlgebra.BlasFlag.SYRK}) @ LinearAlgebra /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:511 [16] _mul! @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:333 [inlined] [17] mul! @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:302 [inlined] [18] mul! @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:270 [inlined] [19] * @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:141 [inlined] [20] svd_recursive(M::JLArrays.JLArray{Float64, 2}; thresh::Float64, north_pass::Int64) @ NDTensors ~/.julia/packages/NDTensors/Lb78J/src/linearalgebra/svd.jl:39 [21] svd_recursive @ ~/.julia/packages/NDTensors/Lb78J/src/linearalgebra/svd.jl:29 [inlined] [22] svd(T::NDTensors.DenseTensor{Float64, 2, Tuple{Int64, Int64}, NDTensors.Dense{Float64, JLArrays.JLArray{Float64, 1}}}; mindim::Nothing, maxdim::Nothing, cutoff::Nothing, use_absolute_cutoff::Nothing, use_relative_cutoff::Nothing, alg::Nothing, min_blockdim::Nothing) @ NDTensors ~/.julia/packages/NDTensors/Lb78J/src/linearalgebra/linearalgebra.jl:108 [23] svd @ ~/.julia/packages/NDTensors/Lb78J/src/linearalgebra/linearalgebra.jl:87 [inlined] [24] svd(T::NDTensors.BlockSparseTensor{Float64, 2, Tuple{Vector{Int64}, Vector{Int64}}, NDTensors.BlockSparse{Float64, JLArrays.JLArray{Float64, 1}, 2}}; min_blockdim::Nothing, mindim::Nothing, maxdim::Nothing, cutoff::Nothing, alg::Nothing, use_absolute_cutoff::Nothing, use_relative_cutoff::Nothing) @ NDTensors ~/.julia/packages/NDTensors/Lb78J/src/blocksparse/linearalgebra.jl:58 [25] svd(T::NDTensors.BlockSparseTensor{Float64, 2, Tuple{Vector{Int64}, Vector{Int64}}, NDTensors.BlockSparse{Float64, JLArrays.JLArray{Float64, 1}, 2}}) @ NDTensors ~/.julia/packages/NDTensors/Lb78J/src/blocksparse/linearalgebra.jl:39 [26] top-level scope @ ~/.julia/packages/NDTensors/Lb78J/test/test_blocksparse.jl:30 [27] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1929 [inlined] [28] macro expansion @ ~/.julia/packages/NDTensors/Lb78J/test/test_blocksparse.jl:287 [inlined] [29] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:2018 [inlined] [30] macro expansion @ ~/.julia/packages/NDTensors/Lb78J/test/test_blocksparse.jl:310 [inlined] [31] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1929 [inlined] [32] macro expansion @ ~/.julia/packages/NDTensors/Lb78J/test/test_blocksparse.jl:312 [inlined] [33] eval(m::Module, e::Any) @ Core ./boot.jl:489 [34] top-level scope @ ~/.julia/packages/NDTensors/Lb78J/test/test_blocksparse.jl:1 [35] include(mapexpr::Function, mod::Module, _path::String) @ Base ./Base.jl:312 [36] top-level scope @ ~/.julia/packages/SafeTestsets/raUNr/src/SafeTestsets.jl:4 [37] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1929 [inlined] [38] macro expansion @ ~/.julia/packages/NDTensors/Lb78J/test/runtests.jl:7 [inlined] [39] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1929 [inlined] [40] macro expansion @ ~/.julia/packages/NDTensors/Lb78J/test/runtests.jl:13 [inlined] [41] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:2018 [inlined] [42] macro expansion @ ~/.julia/packages/NDTensors/Lb78J/test/runtests.jl:15 [inlined] [43] eval(m::Module, e::Any) @ Core ./boot.jl:489 [44] top-level scope @ ~/.julia/packages/NDTensors/Lb78J/test/runtests.jl:28 [45] include(mapexpr::Function, mod::Module, _path::String) @ Base ./Base.jl:312 [46] top-level scope @ none:6 [47] eval(m::Module, e::Any) @ Core ./boot.jl:489 [48] exec_options(opts::Base.JLOptions) @ Base ./client.jl:286 [49] _start() @ Base ./client.jl:553 Running /home/pkgeval/.julia/packages/NDTensors/Lb78J/test/test_combiner.jl Running /home/pkgeval/.julia/packages/NDTensors/Lb78J/test/test_dense.jl DenseTensor basic functionality: Error During Test at /home/pkgeval/.julia/packages/NDTensors/Lb78J/test/test_dense.jl:148 Test threw exception Expression: Array(J * K) ≈ Array(J) * Array(K) Scalar indexing is disallowed. Invocation of getindex resulted in scalar indexing of a GPU array. This is typically caused by calling an iterating implementation of a method. Such implementations *do not* execute on the GPU, but very slowly on the CPU, and therefore should be avoided. If you want to allow scalar iteration, use `allowscalar` or `@allowscalar` to enable scalar iteration globally or for the operations in question. Stacktrace: [1] error(s::String) @ Base ./error.jl:44 [2] errorscalar(op::String) @ GPUArraysCore ~/.julia/packages/GPUArraysCore/aNaXo/src/GPUArraysCore.jl:151 [3] _assertscalar(op::String, behavior::GPUArraysCore.ScalarIndexing) @ GPUArraysCore ~/.julia/packages/GPUArraysCore/aNaXo/src/GPUArraysCore.jl:124 [4] assertscalar(op::String) @ GPUArraysCore ~/.julia/packages/GPUArraysCore/aNaXo/src/GPUArraysCore.jl:112 [5] getindex @ ~/.julia/packages/GPUArrays/u6tui/src/host/indexing.jl:50 [inlined] [6] scalar_getindex @ ~/.julia/packages/GPUArrays/u6tui/src/host/indexing.jl:36 [inlined] [7] _getindex @ ~/.julia/packages/GPUArrays/u6tui/src/host/indexing.jl:19 [inlined] [8] getindex @ ~/.julia/packages/GPUArrays/u6tui/src/host/indexing.jl:17 [inlined] [9] __matmul2x2_elements(tA::Char, A::JLArrays.JLArray{Float64, 2}) @ LinearAlgebra /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:1205 [10] __matmul2x2_elements @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:1234 [inlined] [11] _matmul2x2_elements @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:1199 [inlined] [12] matmul2x2or3x3_nonzeroalpha!(C::JLArrays.JLArray{Float64, 2}, tA::Char, tB::Char, A::JLArrays.JLArray{Float64, 2}, B::JLArrays.JLArray{Float64, 2}, α::Bool, β::Bool) @ LinearAlgebra /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:490 [13] gemm_wrapper! @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:894 [inlined] [14] _syrk_herk_gemm_wrapper! @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:531 [inlined] [15] generic_matmatmul_wrapper!(C::JLArrays.JLArray{Float64, 2}, tA::Char, tB::Char, A::JLArrays.JLArray{Float64, 2}, B::JLArrays.JLArray{Float64, 2}, α::Bool, β::Bool, val::Val{LinearAlgebra.BlasFlag.GEMM}) @ LinearAlgebra /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:511 [16] _mul! @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:333 [inlined] [17] mul! @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:302 [inlined] [18] mul! @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:270 [inlined] [19] * @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:141 [inlined] [20] *(T1::NDTensors.DenseTensor{Float64, 2, Tuple{Int64, Int64}, NDTensors.Dense{Float64, JLArrays.JLArray{Float64, 1}}}, T2::NDTensors.DenseTensor{Float64, 2, Tuple{Int64, Int64}, NDTensors.Dense{Float64, JLArrays.JLArray{Float64, 1}}}) @ NDTensors ~/.julia/packages/NDTensors/Lb78J/src/linearalgebra/linearalgebra.jl:16 [21] top-level scope @ ~/.julia/packages/NDTensors/Lb78J/test/test_dense.jl:15 [22] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1929 [inlined] [23] macro expansion @ ~/.julia/packages/NDTensors/Lb78J/test/test_dense.jl:15 [inlined] [24] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:2018 [inlined] [25] macro expansion @ ~/.julia/packages/NDTensors/Lb78J/test/test_dense.jl:19 [inlined] [26] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1929 [inlined] [27] macro expansion @ ~/.julia/packages/NDTensors/Lb78J/test/test_dense.jl:148 [inlined] [28] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:742 [inlined] WARNING: llvmcall with integer pointers is deprecated. Use actual pointers instead, replacing i32 or i64 with i8* or ptr in _turbo_!(Base.Val{var"#UNROLL#"}, Base.Val{var"#OPS#"}, Base.Val{var"#ARF#"}, Base.Val{var"#AM#"}, Base.Val{var"#LPSYM#"}, Base.Val{Tuple{var"#LB#", var"#V#"}}, Vararg{Any, var"#num#vargs#"}) where {var"#UNROLL#", var"#OPS#", var"#ARF#", var"#AM#", var"#LPSYM#", var"#LB#", var"#V#", var"#num#vargs#"} at /home/pkgeval/.julia/packages/LoopVectorization/ImqiY/src/reconstruct_loopset.jl WARNING: llvmcall with integer pointers is deprecated. Use actual pointers instead, replacing i32 or i64 with i8* or ptr in matmul_st_only_pack_A!(LayoutPointers.AbstractStridedPointer{T, N, C, B, R, X, O} where O<:Tuple{Vararg{Union{Int16, Int32, Int64, Int8, UInt16, UInt32, UInt64, UInt8, Static.StaticInt{N} where N}, N}} where X<:Tuple{Vararg{Union{Int16, Int32, Int64, Int8, UInt16, UInt32, UInt64, UInt8, Static.StaticInt{N} where N}, N}} where R where B where C where N, LayoutPointers.AbstractStridedPointer{T, N, C, B, R, X, O} where O<:Tuple{Vararg{Union{Int16, Int32, Int64, Int8, UInt16, UInt32, UInt64, UInt8, Static.StaticInt{N} where N}, N}} where X<:Tuple{Vararg{Union{Int16, Int32, Int64, Int8, UInt16, UInt32, UInt64, UInt8, Static.StaticInt{N} where N}, N}} where R where B where C where N where T, LayoutPointers.AbstractStridedPointer{T, N, C, B, R, X, O} where O<:Tuple{Vararg{Union{Int16, Int32, Int64, Int8, UInt16, UInt32, UInt64, UInt8, Static.StaticInt{N} where N}, N}} where X<:Tuple{Vararg{Union{Int16, Int32, Int64, Int8, UInt16, UInt32, UInt64, UInt8, Static.StaticInt{N} where N}, N}} where R where B where C where N where T, Any, Any, Any, Any, Any, Static.StaticFloat64{W₁}, Static.StaticFloat64{W₂}, Static.StaticFloat64{R₁}, Static.StaticFloat64{R₂}) where {T, W₁, W₂, R₁, R₂} at /home/pkgeval/.julia/packages/Octavian/4f4xi/src/matmul.jl WARNING: llvmcall with integer pointers is deprecated. Use actual pointers instead, replacing i32 or i64 with i8* or ptr in packaloopmul!(LayoutPointers.AbstractStridedPointer{T, N, C, B, R, X, O} where O<:Tuple{Vararg{Union{Int16, Int32, Int64, Int8, UInt16, UInt32, UInt64, UInt8, Static.StaticInt{N} where N}, N}} where X<:Tuple{Vararg{Union{Int16, Int32, Int64, Int8, UInt16, UInt32, UInt64, UInt8, Static.StaticInt{N} where N}, N}} where R where B where C where N, LayoutPointers.AbstractStridedPointer{T, N, C, B, R, X, O} where O<:Tuple{Vararg{Union{Int16, Int32, Int64, Int8, UInt16, UInt32, UInt64, UInt8, Static.StaticInt{N} where N}, N}} where X<:Tuple{Vararg{Union{Int16, Int32, Int64, Int8, UInt16, UInt32, UInt64, UInt8, Static.StaticInt{N} where N}, N}} where R where B where C where N where T, LayoutPointers.AbstractStridedPointer{T, N, C, B, R, X, O} where O<:Tuple{Vararg{Union{Int16, Int32, Int64, Int8, UInt16, UInt32, UInt64, UInt8, Static.StaticInt{N} where N}, N}} where X<:Tuple{Vararg{Union{Int16, Int32, Int64, Int8, UInt16, UInt32, UInt64, UInt8, Static.StaticInt{N} where N}, N}} where R where B where C where N where T, Any, Any, Any, Any, Any) where {T} at /home/pkgeval/.julia/packages/Octavian/4f4xi/src/macrokernels.jl WARNING: llvmcall with integer pointers is deprecated. Use actual pointers instead, replacing i32 or i64 with i8* or ptr in sync_mul!(LayoutPointers.AbstractStridedPointer{T, N, C, B, R, X, O} where O<:Tuple{Vararg{Union{Int16, Int32, Int64, Int8, UInt16, UInt32, UInt64, UInt8, Static.StaticInt{N} where N}, N}} where X<:Tuple{Vararg{Union{Int16, Int32, Int64, Int8, UInt16, UInt32, UInt64, UInt8, Static.StaticInt{N} where N}, N}} where R where B where C where N, LayoutPointers.AbstractStridedPointer{T, N, C, B, R, X, O} where O<:Tuple{Vararg{Union{Int16, Int32, Int64, Int8, UInt16, UInt32, UInt64, UInt8, Static.StaticInt{N} where N}, N}} where X<:Tuple{Vararg{Union{Int16, Int32, Int64, Int8, UInt16, UInt32, UInt64, UInt8, Static.StaticInt{N} where N}, N}} where R where B where C where N where T, LayoutPointers.AbstractStridedPointer{T, N, C, B, R, X, O} where O<:Tuple{Vararg{Union{Int16, Int32, Int64, Int8, UInt16, UInt32, UInt64, UInt8, Static.StaticInt{N} where N}, N}} where X<:Tuple{Vararg{Union{Int16, Int32, Int64, Int8, UInt16, UInt32, UInt64, UInt8, Static.StaticInt{N} where N}, N}} where R where B where C where N where T, Any, Any, Any, Any, Any, Ptr{UInt32}, Ptr{T} where T, UInt64, UInt64, Static.StaticFloat64{W₁}, Static.StaticFloat64{W₂}, Static.StaticFloat64{R₁}, Static.StaticFloat64{R₂}) where {T, W₁, W₂, R₁, R₂} at /home/pkgeval/.julia/packages/Octavian/4f4xi/src/matmul.jl WARNING: llvmcall with integer pointers is deprecated. Use actual pointers instead, replacing i32 or i64 with i8* or ptr in unsafe_copyto_turbo!(Any, Any, Any, Any) at /home/pkgeval/.julia/packages/Octavian/4f4xi/src/utils.jl WARNING: llvmcall with integer pointers is deprecated. Use actual pointers instead, replacing i32 or i64 with i8* or ptr in matmulsplitn!(LayoutPointers.AbstractStridedPointer{T, N, C, B, R, X, O} where O<:Tuple{Vararg{Union{Int16, Int32, Int64, Int8, UInt16, UInt32, UInt64, UInt8, Static.StaticInt{N} where N}, N}} where X<:Tuple{Vararg{Union{Int16, Int32, Int64, Int8, UInt16, UInt32, UInt64, UInt8, Static.StaticInt{N} where N}, N}} where R where B where C where N, Any, Any, Any, Any, Static.StaticInt{Mc}, Any, Any, Any, Any, Base.Val{PACK}, Any, Any, Any, Any) where {T, Mc, PACK} at /home/pkgeval/.julia/packages/Octavian/4f4xi/src/matmul.jl WARNING: llvmcall with integer pointers is deprecated. Use actual pointers instead, replacing i32 or i64 with i8* or ptr in matmul_pack_A_and_B!(LayoutPointers.AbstractStridedPointer{T, N, C, B, R, X, O} where O<:Tuple{Vararg{Union{Int16, Int32, Int64, Int8, UInt16, UInt32, UInt64, UInt8, Static.StaticInt{N} where N}, N}} where X<:Tuple{Vararg{Union{Int16, Int32, Int64, Int8, UInt16, UInt32, UInt64, UInt8, Static.StaticInt{N} where N}, N}} where R where B where C where N, LayoutPointers.AbstractStridedPointer{T, N, C, B, R, X, O} where O<:Tuple{Vararg{Union{Int16, Int32, Int64, Int8, UInt16, UInt32, UInt64, UInt8, Static.StaticInt{N} where N}, N}} where X<:Tuple{Vararg{Union{Int16, Int32, Int64, Int8, UInt16, UInt32, UInt64, UInt8, Static.StaticInt{N} where N}, N}} where R where B where C where N where T, LayoutPointers.AbstractStridedPointer{T, N, C, B, R, X, O} where O<:Tuple{Vararg{Union{Int16, Int32, Int64, Int8, UInt16, UInt32, UInt64, UInt8, Static.StaticInt{N} where N}, N}} where X<:Tuple{Vararg{Union{Int16, Int32, Int64, Int8, UInt16, UInt32, UInt64, UInt8, Static.StaticInt{N} where N}, N}} where R where B where C where N where T, Any, Any, Any, Any, Any, Any, Static.StaticFloat64{W₁}, Static.StaticFloat64{W₂}, Static.StaticFloat64{R₁}, Static.StaticFloat64{R₂}) where {T, W₁, W₂, R₁, R₂} at /home/pkgeval/.julia/packages/Octavian/4f4xi/src/matmul.jl WARNING: llvmcall with integer pointers is deprecated. Use actual pointers instead, replacing i32 or i64 with i8* or ptr in matmul_st_pack_A_and_B!(LayoutPointers.AbstractStridedPointer{T, N, C, B, R, X, O} where O<:Tuple{Vararg{Union{Int16, Int32, Int64, Int8, UInt16, UInt32, UInt64, UInt8, Static.StaticInt{N} where N}, N}} where X<:Tuple{Vararg{Union{Int16, Int32, Int64, Int8, UInt16, UInt32, UInt64, UInt8, Static.StaticInt{N} where N}, N}} where R where B where C where N, LayoutPointers.AbstractStridedPointer{T, N, C, B, R, X, O} where O<:Tuple{Vararg{Union{Int16, Int32, Int64, Int8, UInt16, UInt32, UInt64, UInt8, Static.StaticInt{N} where N}, N}} where X<:Tuple{Vararg{Union{Int16, Int32, Int64, Int8, UInt16, UInt32, UInt64, UInt8, Static.StaticInt{N} where N}, N}} where R where B where C where N where T, LayoutPointers.AbstractStridedPointer{T, N, C, B, R, X, O} where O<:Tuple{Vararg{Union{Int16, Int32, Int64, Int8, UInt16, UInt32, UInt64, UInt8, Static.StaticInt{N} where N}, N}} where X<:Tuple{Vararg{Union{Int16, Int32, Int64, Int8, UInt16, UInt32, UInt64, UInt8, Static.StaticInt{N} where N}, N}} where R where B where C where N where T, Any, Any, Any, Any, Any, Any, Any, Any, Any, Any) where {T} at /home/pkgeval/.julia/packages/Octavian/4f4xi/src/matmul.jl WARNING: llvmcall with integer pointers is deprecated. Use actual pointers instead, replacing i32 or i64 with i8* or ptr in matmul_only_β!(AbstractArray{T, 2} where T, Any) at /home/pkgeval/.julia/packages/Octavian/4f4xi/src/matmul.jl WARNING: llvmcall with integer pointers is deprecated. Use actual pointers instead, replacing i32 or i64 with i8* or ptr in matmul_only_β!(AbstractArray{T, 2}, Static.StaticInt{0}) where {T} at /home/pkgeval/.julia/packages/Octavian/4f4xi/src/matmul.jl Running /home/pkgeval/.julia/packages/NDTensors/Lb78J/test/test_diag.jl DiagTensor contractions: Error During Test at /home/pkgeval/.julia/packages/NDTensors/Lb78J/test/test_diag.jl:115 Test threw exception Expression: contract(A, (1, -2), t, (-2, 3)) == A Scalar indexing is disallowed. Invocation of getindex resulted in scalar indexing of a GPU array. This is typically caused by calling an iterating implementation of a method. Such implementations *do not* execute on the GPU, but very slowly on the CPU, and therefore should be avoided. If you want to allow scalar iteration, use `allowscalar` or `@allowscalar` to enable scalar iteration globally or for the operations in question. Stacktrace: [1] error(s::String) @ Base ./error.jl:44 [2] errorscalar(op::String) @ GPUArraysCore ~/.julia/packages/GPUArraysCore/aNaXo/src/GPUArraysCore.jl:151 [3] _assertscalar(op::String, behavior::GPUArraysCore.ScalarIndexing) @ GPUArraysCore ~/.julia/packages/GPUArraysCore/aNaXo/src/GPUArraysCore.jl:124 [4] assertscalar(op::String) @ GPUArraysCore ~/.julia/packages/GPUArraysCore/aNaXo/src/GPUArraysCore.jl:112 [5] getindex @ ~/.julia/packages/GPUArrays/u6tui/src/host/indexing.jl:50 [inlined] [6] scalar_getindex @ ~/.julia/packages/GPUArrays/u6tui/src/host/indexing.jl:36 [inlined] [7] _getindex @ ~/.julia/packages/GPUArrays/u6tui/src/host/indexing.jl:19 [inlined] [8] getindex @ ~/.julia/packages/GPUArrays/u6tui/src/host/indexing.jl:17 [inlined] [9] __matmul3x3_elements(tA::Char, A::JLArrays.JLArray{Float64, 2}) @ LinearAlgebra /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:1267 [10] __matmul3x3_elements @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:1304 [inlined] [11] _matmul3x3_elements @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:1261 [inlined] [12] matmul3x3!(C::JLArrays.JLArray{Float64, 2}, tA::Char, tB::Char, A::JLArrays.JLArray{Float64, 2}, B::JLArrays.JLArray{Float64, 2}, α::Float64, β::Float64) @ LinearAlgebra /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:1327 [13] matmul2x2or3x3_nonzeroalpha! @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:483 [inlined] [14] gemm_wrapper!(C::JLArrays.JLArray{Float64, 2}, tA::Char, tB::Char, A::JLArrays.JLArray{Float64, 2}, B::JLArrays.JLArray{Float64, 2}, α::Float64, β::Float64) @ LinearAlgebra /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:894 [15] _syrk_herk_gemm_wrapper! @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:531 [inlined] [16] generic_matmatmul_wrapper!(C::JLArrays.JLArray{Float64, 2}, tA::Char, tB::Char, A::JLArrays.JLArray{Float64, 2}, B::JLArrays.JLArray{Float64, 2}, α::Float64, β::Float64, val::Val{LinearAlgebra.BlasFlag.GEMM}) @ LinearAlgebra /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:511 [17] _mul! @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:333 [inlined] [18] mul! @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:302 [inlined] [19] mul! @ ~/.julia/packages/NDTensors/Lb78J/src/lib/Expose/src/functions/mul.jl:2 [inlined] [20] mul!! @ ~/.julia/packages/NDTensors/Lb78J/src/abstractarray/mul.jl:2 [inlined] [21] mul!! @ ~/.julia/packages/NDTensors/Lb78J/src/abstractarray/mul.jl:10 [inlined] [22] _contract!(CT::JLArrays.JLArray{Float64, 2}, AT::JLArrays.JLArray{Float64, 2}, BT::JLArrays.JLArray{Float64, 2}, props::NDTensors.ContractionProperties{2, 2, 2}, α::Bool, β::Bool) @ NDTensors ~/.julia/packages/NDTensors/Lb78J/src/abstractarray/tensoralgebra/contract.jl:174 [23] _contract! @ ~/.julia/packages/NDTensors/Lb78J/src/dense/tensoralgebra/contract.jl:230 [inlined] [24] contract!(R::NDTensors.DenseTensor{Float64, 2, Tuple{Int64, Int64}, NDTensors.Dense{Float64, JLArrays.JLArray{Float64, 1}}}, labelsR::Tuple{Int64, Int64}, T1::NDTensors.DenseTensor{Float64, 2, Tuple{Int64, Int64}, NDTensors.Dense{Float64, JLArrays.JLArray{Float64, 1}}}, labelsT1::Tuple{Int64, Int64}, T2::NDTensors.DenseTensor{Float64, 2, Tuple{Int64, Int64}, NDTensors.Dense{Float64, JLArrays.JLArray{Float64, 1}}}, labelsT2::Tuple{Int64, Int64}, α::Bool, β::Bool) @ NDTensors ~/.julia/packages/NDTensors/Lb78J/src/dense/tensoralgebra/contract.jl:213 [25] contract! @ ~/.julia/packages/NDTensors/Lb78J/src/tensoroperations/generic_tensor_operations.jl:165 [inlined] [26] #contract!#1 @ ~/.julia/packages/NDTensors/Lb78J/ext/NDTensorsGPUArraysCoreExt/contract.jl:21 [inlined] [27] contract! @ ~/.julia/packages/NDTensors/Lb78J/ext/NDTensorsGPUArraysCoreExt/contract.jl:7 [inlined] ┌[28] contract! │ @ ~/.julia/packages/NDTensors/Lb78J/ext/NDTensorsGPUArraysCoreExt/contract.jl:37 [inlined] ╰──── repeated 2 times [30] _contract!! @ ~/.julia/packages/NDTensors/Lb78J/src/tensoroperations/generic_tensor_operations.jl:143 [inlined] [31] _contract!! @ ~/.julia/packages/NDTensors/Lb78J/src/tensoroperations/generic_tensor_operations.jl:131 [inlined] [32] contract!!(output_tensor::NDTensors.DenseTensor{Float64, 2, Tuple{Int64, Int64}, NDTensors.Dense{Float64, JLArrays.JLArray{Float64, 1}}}, labelsoutput_tensor::Tuple{Int64, Int64}, tensor1::NDTensors.DenseTensor{Float64, 2, Tuple{Int64, Int64}, NDTensors.Dense{Float64, JLArrays.JLArray{Float64, 1}}}, labelstensor1::Tuple{Int64, Int64}, tensor2::NDTensors.DiagTensor{Float64, 2, Tuple{Int64, Int64}, NDTensors.Diag{Float64, JLArrays.JLArray{Float64, 1}}}, labelstensor2::Tuple{Int64, Int64}, α::Int64, β::Int64) @ NDTensors ~/.julia/packages/NDTensors/Lb78J/src/tensoroperations/generic_tensor_operations.jl:219 [33] contract!! @ ~/.julia/packages/NDTensors/Lb78J/src/tensoroperations/generic_tensor_operations.jl:188 [inlined] [34] contract(tensor1::NDTensors.DenseTensor{Float64, 2, Tuple{Int64, Int64}, NDTensors.Dense{Float64, JLArrays.JLArray{Float64, 1}}}, labelstensor1::Tuple{Int64, Int64}, tensor2::NDTensors.DiagTensor{Float64, 2, Tuple{Int64, Int64}, NDTensors.Diag{Float64, JLArrays.JLArray{Float64, 1}}}, labelstensor2::Tuple{Int64, Int64}, labelsoutput_tensor::Tuple{Int64, Int64}) @ NDTensors ~/.julia/packages/NDTensors/Lb78J/src/tensoroperations/generic_tensor_operations.jl:113 [35] contract(::Type{NDTensors.CanContract{NDTensors.DenseTensor{Float64, 2, Tuple{Int64, Int64}, NDTensors.Dense{Float64, JLArrays.JLArray{Float64, 1}}}, NDTensors.DiagTensor{Float64, 2, Tuple{Int64, Int64}, NDTensors.Diag{Float64, JLArrays.JLArray{Float64, 1}}}}}, tensor1::NDTensors.DenseTensor{Float64, 2, Tuple{Int64, Int64}, NDTensors.Dense{Float64, JLArrays.JLArray{Float64, 1}}}, labels_tensor1::Tuple{Int64, Int64}, tensor2::NDTensors.DiagTensor{Float64, 2, Tuple{Int64, Int64}, NDTensors.Diag{Float64, JLArrays.JLArray{Float64, 1}}}, labels_tensor2::Tuple{Int64, Int64}) @ NDTensors ~/.julia/packages/NDTensors/Lb78J/src/tensoroperations/generic_tensor_operations.jl:91 [36] contract(tensor1::NDTensors.DenseTensor{Float64, 2, Tuple{Int64, Int64}, NDTensors.Dense{Float64, JLArrays.JLArray{Float64, 1}}}, labels_tensor1::Tuple{Int64, Int64}, tensor2::NDTensors.DiagTensor{Float64, 2, Tuple{Int64, Int64}, NDTensors.Diag{Float64, JLArrays.JLArray{Float64, 1}}}, labels_tensor2::Tuple{Int64, Int64}) @ NDTensors ~/.julia/packages/SimpleTraits/7VJph/src/SimpleTraits.jl:332 [37] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:742 [inlined] [38] macro expansion @ ~/.julia/packages/NDTensors/Lb78J/test/test_diag.jl:115 [inlined] [39] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:2018 [inlined] [40] top-level scope @ ~/.julia/packages/NDTensors/Lb78J/test/test_diag.jl:2041 DiagTensor contractions: Error During Test at /home/pkgeval/.julia/packages/NDTensors/Lb78J/test/test_diag.jl:116 Test threw exception Expression: contract(A, (-2, 1), t, (-2, 3)) == transpose(A) Scalar indexing is disallowed. Invocation of getindex resulted in scalar indexing of a GPU array. This is typically caused by calling an iterating implementation of a method. Such implementations *do not* execute on the GPU, but very slowly on the CPU, and therefore should be avoided. If you want to allow scalar iteration, use `allowscalar` or `@allowscalar` to enable scalar iteration globally or for the operations in question. Stacktrace: [1] error(s::String) @ Base ./error.jl:44 [2] errorscalar(op::String) @ GPUArraysCore ~/.julia/packages/GPUArraysCore/aNaXo/src/GPUArraysCore.jl:151 [3] _assertscalar(op::String, behavior::GPUArraysCore.ScalarIndexing) @ GPUArraysCore ~/.julia/packages/GPUArraysCore/aNaXo/src/GPUArraysCore.jl:124 [4] assertscalar(op::String) @ GPUArraysCore ~/.julia/packages/GPUArraysCore/aNaXo/src/GPUArraysCore.jl:112 [5] getindex @ ~/.julia/packages/GPUArrays/u6tui/src/host/indexing.jl:50 [inlined] [6] scalar_getindex @ ~/.julia/packages/GPUArrays/u6tui/src/host/indexing.jl:36 [inlined] [7] _getindex @ ~/.julia/packages/GPUArrays/u6tui/src/host/indexing.jl:19 [inlined] [8] getindex @ ~/.julia/packages/GPUArrays/u6tui/src/host/indexing.jl:17 [inlined] [9] __matmul3x3_elements(tA::Char, A::JLArrays.JLArray{Float64, 2}) @ LinearAlgebra /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:1272 [10] __matmul3x3_elements @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:1304 [inlined] [11] _matmul3x3_elements @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:1261 [inlined] [12] matmul3x3!(C::JLArrays.JLArray{Float64, 2}, tA::Char, tB::Char, A::JLArrays.JLArray{Float64, 2}, B::JLArrays.JLArray{Float64, 2}, α::Float64, β::Float64) @ LinearAlgebra /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:1327 [13] matmul2x2or3x3_nonzeroalpha! @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:483 [inlined] [14] gemm_wrapper!(C::JLArrays.JLArray{Float64, 2}, tA::Char, tB::Char, A::JLArrays.JLArray{Float64, 2}, B::JLArrays.JLArray{Float64, 2}, α::Float64, β::Float64) @ LinearAlgebra /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:894 [15] _syrk_herk_gemm_wrapper! @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:519 [inlined] [16] generic_matmatmul_wrapper!(C::JLArrays.JLArray{Float64, 2}, tA::Char, tB::Char, A::JLArrays.JLArray{Float64, 2}, B::JLArrays.JLArray{Float64, 2}, α::Float64, β::Float64, val::Val{LinearAlgebra.BlasFlag.SYRK}) @ LinearAlgebra /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:511 [17] _mul! @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:333 [inlined] [18] mul! @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:302 [inlined] [19] mul! @ ~/.julia/packages/NDTensors/Lb78J/ext/NDTensorsJLArraysExt/mul.jl:12 [inlined] [20] mul!!(CM::LinearAlgebra.Transpose{Float64, JLArrays.JLArray{Float64, 2}}, AM::LinearAlgebra.Transpose{Float64, JLArrays.JLArray{Float64, 2}}, BM::JLArrays.JLArray{Float64, 2}, α::Float64, β::Float64) @ NDTensors ~/.julia/packages/NDTensors/Lb78J/src/abstractarray/mul.jl:2 [21] _contract!(CT::JLArrays.JLArray{Float64, 2}, AT::JLArrays.JLArray{Float64, 2}, BT::JLArrays.JLArray{Float64, 2}, props::NDTensors.ContractionProperties{2, 2, 2}, α::Bool, β::Bool) @ NDTensors ~/.julia/packages/NDTensors/Lb78J/src/abstractarray/tensoralgebra/contract.jl:174 [22] _contract! @ ~/.julia/packages/NDTensors/Lb78J/src/dense/tensoralgebra/contract.jl:230 [inlined] [23] contract!(R::NDTensors.DenseTensor{Float64, 2, Tuple{Int64, Int64}, NDTensors.Dense{Float64, JLArrays.JLArray{Float64, 1}}}, labelsR::Tuple{Int64, Int64}, T1::NDTensors.DenseTensor{Float64, 2, Tuple{Int64, Int64}, NDTensors.Dense{Float64, JLArrays.JLArray{Float64, 1}}}, labelsT1::Tuple{Int64, Int64}, T2::NDTensors.DenseTensor{Float64, 2, Tuple{Int64, Int64}, NDTensors.Dense{Float64, JLArrays.JLArray{Float64, 1}}}, labelsT2::Tuple{Int64, Int64}, α::Bool, β::Bool) @ NDTensors ~/.julia/packages/NDTensors/Lb78J/src/dense/tensoralgebra/contract.jl:213 [24] contract! @ ~/.julia/packages/NDTensors/Lb78J/src/tensoroperations/generic_tensor_operations.jl:165 [inlined] [25] #contract!#1 @ ~/.julia/packages/NDTensors/Lb78J/ext/NDTensorsGPUArraysCoreExt/contract.jl:21 [inlined] [26] contract! @ ~/.julia/packages/NDTensors/Lb78J/ext/NDTensorsGPUArraysCoreExt/contract.jl:7 [inlined] ┌[27] contract! │ @ ~/.julia/packages/NDTensors/Lb78J/ext/NDTensorsGPUArraysCoreExt/contract.jl:37 [inlined] ╰──── repeated 2 times [29] _contract!! @ ~/.julia/packages/NDTensors/Lb78J/src/tensoroperations/generic_tensor_operations.jl:143 [inlined] [30] _contract!! @ ~/.julia/packages/NDTensors/Lb78J/src/tensoroperations/generic_tensor_operations.jl:131 [inlined] [31] contract!!(output_tensor::NDTensors.DenseTensor{Float64, 2, Tuple{Int64, Int64}, NDTensors.Dense{Float64, JLArrays.JLArray{Float64, 1}}}, labelsoutput_tensor::Tuple{Int64, Int64}, tensor1::NDTensors.DenseTensor{Float64, 2, Tuple{Int64, Int64}, NDTensors.Dense{Float64, JLArrays.JLArray{Float64, 1}}}, labelstensor1::Tuple{Int64, Int64}, tensor2::NDTensors.DiagTensor{Float64, 2, Tuple{Int64, Int64}, NDTensors.Diag{Float64, JLArrays.JLArray{Float64, 1}}}, labelstensor2::Tuple{Int64, Int64}, α::Int64, β::Int64) @ NDTensors ~/.julia/packages/NDTensors/Lb78J/src/tensoroperations/generic_tensor_operations.jl:219 [32] contract!! @ ~/.julia/packages/NDTensors/Lb78J/src/tensoroperations/generic_tensor_operations.jl:188 [inlined] [33] contract(tensor1::NDTensors.DenseTensor{Float64, 2, Tuple{Int64, Int64}, NDTensors.Dense{Float64, JLArrays.JLArray{Float64, 1}}}, labelstensor1::Tuple{Int64, Int64}, tensor2::NDTensors.DiagTensor{Float64, 2, Tuple{Int64, Int64}, NDTensors.Diag{Float64, JLArrays.JLArray{Float64, 1}}}, labelstensor2::Tuple{Int64, Int64}, labelsoutput_tensor::Tuple{Int64, Int64}) @ NDTensors ~/.julia/packages/NDTensors/Lb78J/src/tensoroperations/generic_tensor_operations.jl:113 [34] contract(::Type{NDTensors.CanContract{NDTensors.DenseTensor{Float64, 2, Tuple{Int64, Int64}, NDTensors.Dense{Float64, JLArrays.JLArray{Float64, 1}}}, NDTensors.DiagTensor{Float64, 2, Tuple{Int64, Int64}, NDTensors.Diag{Float64, JLArrays.JLArray{Float64, 1}}}}}, tensor1::NDTensors.DenseTensor{Float64, 2, Tuple{Int64, Int64}, NDTensors.Dense{Float64, JLArrays.JLArray{Float64, 1}}}, labels_tensor1::Tuple{Int64, Int64}, tensor2::NDTensors.DiagTensor{Float64, 2, Tuple{Int64, Int64}, NDTensors.Diag{Float64, JLArrays.JLArray{Float64, 1}}}, labels_tensor2::Tuple{Int64, Int64}) @ NDTensors ~/.julia/packages/NDTensors/Lb78J/src/tensoroperations/generic_tensor_operations.jl:91 [35] contract(tensor1::NDTensors.DenseTensor{Float64, 2, Tuple{Int64, Int64}, NDTensors.Dense{Float64, JLArrays.JLArray{Float64, 1}}}, labels_tensor1::Tuple{Int64, Int64}, tensor2::NDTensors.DiagTensor{Float64, 2, Tuple{Int64, Int64}, NDTensors.Diag{Float64, JLArrays.JLArray{Float64, 1}}}, labels_tensor2::Tuple{Int64, Int64}) @ NDTensors ~/.julia/packages/SimpleTraits/7VJph/src/SimpleTraits.jl:332 [36] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:742 [inlined] [37] macro expansion @ ~/.julia/packages/NDTensors/Lb78J/test/test_diag.jl:116 [inlined] [38] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:2018 [inlined] [39] top-level scope @ ~/.julia/packages/NDTensors/Lb78J/test/test_diag.jl:2041 Running /home/pkgeval/.julia/packages/NDTensors/Lb78J/test/test_diagblocksparse.jl DiagBlockSparse contract: Error During Test at /home/pkgeval/.julia/packages/NDTensors/Lb78J/test/test_diagblocksparse.jl:60 Got exception outside of a @test Scalar indexing is disallowed. Invocation of getindex resulted in scalar indexing of a GPU array. This is typically caused by calling an iterating implementation of a method. Such implementations *do not* execute on the GPU, but very slowly on the CPU, and therefore should be avoided. If you want to allow scalar iteration, use `allowscalar` or `@allowscalar` to enable scalar iteration globally or for the operations in question. Stacktrace: [1] error(s::String) @ Base ./error.jl:44 [2] errorscalar(op::String) @ GPUArraysCore ~/.julia/packages/GPUArraysCore/aNaXo/src/GPUArraysCore.jl:151 [3] _assertscalar(op::String, behavior::GPUArraysCore.ScalarIndexing) @ GPUArraysCore ~/.julia/packages/GPUArraysCore/aNaXo/src/GPUArraysCore.jl:124 [4] assertscalar(op::String) @ GPUArraysCore ~/.julia/packages/GPUArraysCore/aNaXo/src/GPUArraysCore.jl:112 [5] getindex @ ~/.julia/packages/GPUArrays/u6tui/src/host/indexing.jl:50 [inlined] [6] scalar_getindex @ ~/.julia/packages/GPUArrays/u6tui/src/host/indexing.jl:36 [inlined] [7] _getindex @ ~/.julia/packages/GPUArrays/u6tui/src/host/indexing.jl:19 [inlined] [8] getindex @ ~/.julia/packages/GPUArrays/u6tui/src/host/indexing.jl:17 [inlined] [9] __matmul2x2_elements(tA::Char, A::JLArrays.JLArray{Float64, 2}) @ LinearAlgebra /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:1205 [10] __matmul2x2_elements @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:1234 [inlined] [11] _matmul2x2_elements @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:1199 [inlined] [12] matmul2x2!(C::JLArrays.JLArray{Float64, 2}, tA::Char, tB::Char, A::JLArrays.JLArray{Float64, 2}, B::JLArrays.JLArray{Float64, 2}, α::Float64, β::Float64) @ LinearAlgebra /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:1248 [13] matmul2x2or3x3_nonzeroalpha! @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:479 [inlined] [14] gemm_wrapper!(C::JLArrays.JLArray{Float64, 2}, tA::Char, tB::Char, A::JLArrays.JLArray{Float64, 2}, B::JLArrays.JLArray{Float64, 2}, α::Float64, β::Float64) @ LinearAlgebra /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:894 [15] _syrk_herk_gemm_wrapper! @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:519 [inlined] [16] generic_matmatmul_wrapper!(C::JLArrays.JLArray{Float64, 2}, tA::Char, tB::Char, A::JLArrays.JLArray{Float64, 2}, B::JLArrays.JLArray{Float64, 2}, α::Float64, β::Float64, val::Val{LinearAlgebra.BlasFlag.SYRK}) @ LinearAlgebra /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:511 [17] _mul! @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:333 [inlined] [18] mul! @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/matmul.jl:302 [inlined] [19] mul! @ ~/.julia/packages/NDTensors/Lb78J/ext/NDTensorsJLArraysExt/mul.jl:12 [inlined] [20] mul!!(CM::LinearAlgebra.Transpose{Float64, JLArrays.JLArray{Float64, 2}}, AM::JLArrays.JLArray{Float64, 2}, BM::LinearAlgebra.Transpose{Float64, JLArrays.JLArray{Float64, 2}}, α::Float64, β::Float64) @ NDTensors ~/.julia/packages/NDTensors/Lb78J/src/abstractarray/mul.jl:2 [21] _contract!(CT::JLArrays.JLArray{Float64, 2}, AT::JLArrays.JLArray{Float64, 2}, BT::JLArrays.JLArray{Float64, 2}, props::NDTensors.ContractionProperties{2, 2, 2}, α::Float64, β::Float64) @ NDTensors ~/.julia/packages/NDTensors/Lb78J/src/abstractarray/tensoralgebra/contract.jl:174 [22] _contract! @ ~/.julia/packages/NDTensors/Lb78J/src/dense/tensoralgebra/contract.jl:230 [inlined] [23] contract!(R::NDTensors.DenseTensor{Float64, 2, Tuple{Int64, Int64}, NDTensors.Dense{Float64, JLArrays.JLArray{Float64, 1}}}, labelsR::Tuple{Int64, Int64}, T1::NDTensors.DenseTensor{Float64, 2, Tuple{Int64, Int64}, NDTensors.Dense{Float64, JLArrays.JLArray{Float64, 1}}}, labelsT1::Tuple{Int64, Int64}, T2::NDTensors.DenseTensor{Float64, 2, Tuple{Int64, Int64}, NDTensors.Dense{Float64, JLArrays.JLArray{Float64, 1}}}, labelsT2::Tuple{Int64, Int64}, α::Float64, β::Float64) @ NDTensors ~/.julia/packages/NDTensors/Lb78J/src/dense/tensoralgebra/contract.jl:213 [24] contract! @ ~/.julia/packages/NDTensors/Lb78J/src/tensoroperations/generic_tensor_operations.jl:165 [inlined] [25] contract!(output_tensor::NDTensors.Expose.Exposed{JLArrays.JLArray{Float64, 1}, NDTensors.DenseTensor{Float64, 2, Tuple{Int64, Int64}, NDTensors.Dense{Float64, JLArrays.JLArray{Float64, 1}}}}, labelsoutput_tensor::Tuple{Int64, Int64}, tensor1::NDTensors.Expose.Exposed{Float64, NDTensors.DiagTensor{Float64, 2, Tuple{Int64, Int64}, NDTensors.Diag{Float64, Float64}}}, labelstensor1::Tuple{Int64, Int64}, tensor2::NDTensors.Expose.Exposed{JLArrays.JLArray{Float64, 1}, NDTensors.DenseTensor{Float64, 2, Tuple{Int64, Int64}, NDTensors.Dense{Float64, JLArrays.JLArray{Float64, 1}}}}, labelstensor2::Tuple{Int64, Int64}, α::Float64, β::Float64) @ NDTensorsGPUArraysCoreExt ~/.julia/packages/NDTensors/Lb78J/ext/NDTensorsGPUArraysCoreExt/contract.jl:61 [26] contract! @ ~/.julia/packages/NDTensors/Lb78J/ext/NDTensorsGPUArraysCoreExt/contract.jl:77 [inlined] [27] contract!(R::NDTensors.BlockSparseTensor{Float64, 2, Tuple{Vector{Int64}, Vector{Int64}}, NDTensors.BlockSparse{Float64, JLArrays.JLArray{Float64, 1}, 2}}, labelsR::Tuple{Int64, Int64}, T1::NDTensors.BlockSparseTensor{Float64, 2, Tuple{Vector{Int64}, Vector{Int64}}, NDTensors.BlockSparse{Float64, JLArrays.JLArray{Float64, 1}, 2}}, labelsT1::Tuple{Int64, Int64}, T2::NDTensors.DiagBlockSparseTensor{Float64, 2, Tuple{Vector{Int64}, Vector{Int64}}, NDTensors.DiagBlockSparse{Float64, Float64, 2}}, labelsT2::Tuple{Int64, Int64}, contraction_plan::Vector{Tuple{NDTensors.Block{2}, NDTensors.Block{2}, NDTensors.Block{2}}}) @ NDTensors ~/.julia/packages/NDTensors/Lb78J/src/blocksparse/diagblocksparse.jl:671 [28] contract(T1::NDTensors.BlockSparseTensor{Float64, 2, Tuple{Vector{Int64}, Vector{Int64}}, NDTensors.BlockSparse{Float64, JLArrays.JLArray{Float64, 1}, 2}}, labelsT1::Tuple{Int64, Int64}, T2::NDTensors.DiagBlockSparseTensor{Float64, 2, Tuple{Vector{Int64}, Vector{Int64}}, NDTensors.DiagBlockSparse{Float64, Float64, 2}}, labelsT2::Tuple{Int64, Int64}, labelsR::Tuple{Int64, Int64}) @ NDTensors ~/.julia/packages/NDTensors/Lb78J/src/blocksparse/diagblocksparse.jl:621 [29] contract(T1::NDTensors.BlockSparseTensor{Float64, 2, Tuple{Vector{Int64}, Vector{Int64}}, NDTensors.BlockSparse{Float64, JLArrays.JLArray{Float64, 1}, 2}}, labelsT1::Tuple{Int64, Int64}, T2::NDTensors.DiagBlockSparseTensor{Float64, 2, Tuple{Vector{Int64}, Vector{Int64}}, NDTensors.DiagBlockSparse{Float64, Float64, 2}}, labelsT2::Tuple{Int64, Int64}) @ NDTensors ~/.julia/packages/NDTensors/Lb78J/src/blocksparse/diagblocksparse.jl:620 [30] macro expansion @ ~/.julia/packages/NDTensors/Lb78J/test/test_diagblocksparse.jl:67 [inlined] [31] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:2018 [inlined] [32] top-level scope @ ~/.julia/packages/NDTensors/Lb78J/test/test_diagblocksparse.jl:2041 [33] eval(m::Module, e::Any) @ Core ./boot.jl:489 [34] top-level scope @ ~/.julia/packages/NDTensors/Lb78J/test/test_diagblocksparse.jl:1 [35] include(mapexpr::Function, mod::Module, _path::String) @ Base ./Base.jl:312 [36] top-level scope @ ~/.julia/packages/SafeTestsets/raUNr/src/SafeTestsets.jl:4 [37] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1929 [inlined] [38] macro expansion @ ~/.julia/packages/NDTensors/Lb78J/test/runtests.jl:7 [inlined] [39] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1929 [inlined] [40] macro expansion @ ~/.julia/packages/NDTensors/Lb78J/test/runtests.jl:13 [inlined] [41] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:2018 [inlined] [42] macro expansion @ ~/.julia/packages/NDTensors/Lb78J/test/runtests.jl:15 [inlined] [43] eval(m::Module, e::Any) @ Core ./boot.jl:489 [44] top-level scope @ ~/.julia/packages/NDTensors/Lb78J/test/runtests.jl:28 [45] include(mapexpr::Function, mod::Module, _path::String) @ Base ./Base.jl:312 [46] top-level scope @ none:6 [47] eval(m::Module, e::Any) @ Core ./boot.jl:489 [48] exec_options(opts::Base.JLOptions) @ Base ./client.jl:286 [49] _start() @ Base ./client.jl:553 Running /home/pkgeval/.julia/packages/NDTensors/Lb78J/test/test_emptynumber.jl Running /home/pkgeval/.julia/packages/NDTensors/Lb78J/test/test_emptystorage.jl Running /home/pkgeval/.julia/packages/NDTensors/Lb78J/test/test_linearalgebra.jl Running /home/pkgeval/.julia/packages/NDTensors/Lb78J/test/test_tupletools.jl Running /home/pkgeval/.julia/packages/NDTensors/Lb78J/test/lib/runtests.jl NDTensors.BackendSelection.Algorithm type , NamedTuple() Testing Expose jl, Float32: Error During Test at /home/pkgeval/.julia/packages/NDTensors/Lb78J/src/lib/Expose/test/runtests.jl:23 Got exception outside of a @test Scalar indexing is disallowed. Invocation of getindex resulted in scalar indexing of a GPU array. This is typically caused by calling an iterating implementation of a method. Such implementations *do not* execute on the GPU, but very slowly on the CPU, and therefore should be avoided. If you want to allow scalar iteration, use `allowscalar` or `@allowscalar` to enable scalar iteration globally or for the operations in question. Stacktrace: [1] error(s::String) @ Base ./error.jl:44 [2] errorscalar(op::String) @ GPUArraysCore ~/.julia/packages/GPUArraysCore/aNaXo/src/GPUArraysCore.jl:151 [3] _assertscalar(op::String, behavior::GPUArraysCore.ScalarIndexing) @ GPUArraysCore ~/.julia/packages/GPUArraysCore/aNaXo/src/GPUArraysCore.jl:124 [4] assertscalar(op::String) @ GPUArraysCore ~/.julia/packages/GPUArraysCore/aNaXo/src/GPUArraysCore.jl:112 [5] getindex @ ~/.julia/packages/GPUArrays/u6tui/src/host/indexing.jl:50 [inlined] [6] _count_svdvals(S::JLArrays.JLArray{Float32, 1}, atol::Int64, rtol::Int64) @ LinearAlgebra /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/svd.jl:137 [7] svd!(A::JLArrays.JLArray{Float32, 2}; full::Bool, alg::LinearAlgebra.DivideAndConquer, atol::Int64, rtol::Int64) @ LinearAlgebra /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/svd.jl:107 [8] svd! @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/svd.jl:100 [inlined] [9] #svd#122 @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/svd.jl:194 [inlined] [10] svd @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/svd.jl:193 [inlined] [11] #svd#1 @ ~/.julia/packages/NDTensors/Lb78J/src/lib/Expose/src/functions/linearalgebra.jl:28 [inlined] [12] svd(E::NDTensors.Expose.Exposed{JLArrays.JLArray{Float32, 2}, JLArrays.JLArray{Float32, 2}}) @ NDTensors.Expose ~/.julia/packages/NDTensors/Lb78J/src/lib/Expose/src/functions/linearalgebra.jl:27 [13] macro expansion @ ~/.julia/packages/NDTensors/Lb78J/src/lib/Expose/test/runtests.jl:99 [inlined] [14] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:2018 [inlined] [15] top-level scope @ ~/.julia/packages/NDTensors/Lb78J/src/lib/Expose/test/runtests.jl:2041 [16] eval(m::Module, e::Any) @ Core ./boot.jl:489 [17] top-level scope @ ~/.julia/packages/NDTensors/Lb78J/src/lib/Expose/test/runtests.jl:1 [18] include(mapexpr::Function, mod::Module, _path::String) @ Base ./Base.jl:312 [19] macro expansion @ ~/.julia/packages/NDTensors/Lb78J/test/lib/runtests.jl:12 [inlined] [20] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:2018 [inlined] [21] top-level scope @ ~/.julia/packages/NDTensors/Lb78J/test/lib/runtests.jl:2041 [22] eval(m::Module, e::Any) @ Core ./boot.jl:489 [23] top-level scope @ ~/.julia/packages/NDTensors/Lb78J/test/lib/runtests.jl:1 [24] include(mapexpr::Function, mod::Module, _path::String) @ Base ./Base.jl:312 [25] top-level scope @ ~/.julia/packages/SafeTestsets/raUNr/src/SafeTestsets.jl:4 [26] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1929 [inlined] [27] macro expansion @ ~/.julia/packages/NDTensors/Lb78J/test/runtests.jl:7 [inlined] [28] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1929 [inlined] [29] macro expansion @ ~/.julia/packages/NDTensors/Lb78J/test/runtests.jl:13 [inlined] [30] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:2018 [inlined] [31] macro expansion @ ~/.julia/packages/NDTensors/Lb78J/test/runtests.jl:15 [inlined] [32] eval(m::Module, e::Any) @ Core ./boot.jl:489 [33] top-level scope @ ~/.julia/packages/NDTensors/Lb78J/test/runtests.jl:28 [34] include(mapexpr::Function, mod::Module, _path::String) @ Base ./Base.jl:312 [35] top-level scope @ none:6 [36] eval(m::Module, e::Any) @ Core ./boot.jl:489 [37] exec_options(opts::Base.JLOptions) @ Base ./client.jl:286 [38] _start() @ Base ./client.jl:553 Testing Expose jl, ComplexF32: Error During Test at /home/pkgeval/.julia/packages/NDTensors/Lb78J/src/lib/Expose/test/runtests.jl:23 Got exception outside of a @test Scalar indexing is disallowed. Invocation of getindex resulted in scalar indexing of a GPU array. This is typically caused by calling an iterating implementation of a method. Such implementations *do not* execute on the GPU, but very slowly on the CPU, and therefore should be avoided. If you want to allow scalar iteration, use `allowscalar` or `@allowscalar` to enable scalar iteration globally or for the operations in question. Stacktrace: [1] error(s::String) @ Base ./error.jl:44 [2] errorscalar(op::String) @ GPUArraysCore ~/.julia/packages/GPUArraysCore/aNaXo/src/GPUArraysCore.jl:151 [3] _assertscalar(op::String, behavior::GPUArraysCore.ScalarIndexing) @ GPUArraysCore ~/.julia/packages/GPUArraysCore/aNaXo/src/GPUArraysCore.jl:124 [4] assertscalar(op::String) @ GPUArraysCore ~/.julia/packages/GPUArraysCore/aNaXo/src/GPUArraysCore.jl:112 [5] getindex @ ~/.julia/packages/GPUArrays/u6tui/src/host/indexing.jl:50 [inlined] [6] _count_svdvals(S::JLArrays.JLArray{Float32, 1}, atol::Int64, rtol::Int64) @ LinearAlgebra /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/svd.jl:137 [7] svd!(A::JLArrays.JLArray{ComplexF32, 2}; full::Bool, alg::LinearAlgebra.DivideAndConquer, atol::Int64, rtol::Int64) @ LinearAlgebra /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/svd.jl:107 [8] svd! @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/svd.jl:100 [inlined] [9] #svd#122 @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/svd.jl:194 [inlined] [10] svd @ /opt/julia/share/julia/stdlib/v1.13/LinearAlgebra/src/svd.jl:193 [inlined] [11] #svd#1 @ ~/.julia/packages/NDTensors/Lb78J/src/lib/Expose/src/functions/linearalgebra.jl:28 [inlined] [12] svd(E::NDTensors.Expose.Exposed{JLArrays.JLArray{ComplexF32, 2}, JLArrays.JLArray{ComplexF32, 2}}) @ NDTensors.Expose ~/.julia/packages/NDTensors/Lb78J/src/lib/Expose/src/functions/linearalgebra.jl:27 [13] macro expansion @ ~/.julia/packages/NDTensors/Lb78J/src/lib/Expose/test/runtests.jl:99 [inlined] [14] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:2018 [inlined] [15] top-level scope @ ~/.julia/packages/NDTensors/Lb78J/src/lib/Expose/test/runtests.jl:2041 [16] eval(m::Module, e::Any) @ Core ./boot.jl:489 [17] top-level scope @ ~/.julia/packages/NDTensors/Lb78J/src/lib/Expose/test/runtests.jl:1 [18] include(mapexpr::Function, mod::Module, _path::String) @ Base ./Base.jl:312 [19] macro expansion @ ~/.julia/packages/NDTensors/Lb78J/test/lib/runtests.jl:12 [inlined] [20] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:2018 [inlined] [21] top-level scope @ ~/.julia/packages/NDTensors/Lb78J/test/lib/runtests.jl:2041 [22] eval(m::Module, e::Any) @ Core ./boot.jl:489 [23] top-level scope @ ~/.julia/packages/NDTensors/Lb78J/test/lib/runtests.jl:1 [24] include(mapexpr::Function, mod::Module, _path::String) @ Base ./Base.jl:312 [25] top-level scope @ ~/.julia/packages/SafeTestsets/raUNr/src/SafeTestsets.jl:4 [26] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1929 [inlined] [27] macro expansion @ ~/.julia/packages/NDTensors/Lb78J/test/runtests.jl:7 [inlined] [28] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:1929 [inlined] [29] macro expansion @ ~/.julia/packages/NDTensors/Lb78J/test/runtests.jl:13 [inlined] [30] macro expansion @ /opt/julia/share/julia/stdlib/v1.13/Test/src/Test.jl:2018 [inlined] [31] macro expansion @ ~/.julia/packages/NDTensors/Lb78J/test/runtests.jl:15 [inlined] [32] eval(m::Module, e::Any) @ Core ./boot.jl:489 [33] top-level scope @ ~/.julia/packages/NDTensors/Lb78J/test/runtests.jl:28 [34] include(mapexpr::Function, mod::Module, _path::String) @ Base ./Base.jl:312 [35] top-level scope @ none:6 [36] eval(m::Module, e::Any) @ Core ./boot.jl:489 [37] exec_options(opts::Base.JLOptions) @ Base ./client.jl:286 [38] _start() @ Base ./client.jl:553 Test Summary: | Pass Error Total Time NDTensors | 2708 12 2720 19m46.6s /home/pkgeval/.julia/packages/NDTensors/Lb78J/test | 2708 12 2720 19m44.5s Test /home/pkgeval/.julia/packages/NDTensors/Lb78J/test/test_blocksparse.jl | 1019 6 1025 3m41.2s BlockSparseTensor basic functionality | 1019 6 1025 3m32.4s test device: cpu, eltype: Float32 | 230 230 23.7s test device: cpu, eltype: Float64 | 230 230 11.1s test device: jl, eltype: Float32 | 230 230 31.8s test device: jl, eltype: Float64 | 230 230 28.6s BlockSparseTensor setindex! add block | 79 79 0.1s svd on cpu, eltype: Float32 | 5 5 19.8s svd on cpu, eltype: Float64 | 5 5 18.1s svd on jl, eltype: Float32 | 2 3 5 28.1s svd example 1 | 1 1 15.6s svd example 2 | 1 1 0.1s svd example 3 | 1 1 0.1s svd example 4 | 1 1 12.3s svd example 5 | 1 1 0.0s svd on jl, eltype: Float64 | 2 3 5 23.8s svd example 1 | 1 1 11.6s svd example 2 | 1 1 0.1s svd example 3 | 1 1 0.1s svd example 4 | 1 1 11.9s svd example 5 | 1 1 0.0s exp, eltype: Float32 | 3 3 14.0s exp, eltype: Float64 | 3 3 13.0s Test /home/pkgeval/.julia/packages/NDTensors/Lb78J/test/test_combiner.jl | 144 144 1m27.3s Test /home/pkgeval/.julia/packages/NDTensors/Lb78J/test/test_dense.jl | 544 1 545 3m48.2s Dense Tensors | 544 1 545 3m47.6s test device: cpu | 260 260 28.0s test device: jl | 259 1 260 59.0s DenseTensor basic functionality | 116 1 117 10.8s Random constructor | 8 8 0.1s Complex Valued Tensors | 72 72 0.1s Custom inds types | 21 21 1.2s generic contraction | 36 36 0.0s Contraction with size 1 block and NaN | 6 6 46.8s Contract with exotic types | 14 14 18.0s change backends | 6 6 1.5s change backends | 5 5 2m00.8s Test /home/pkgeval/.julia/packages/NDTensors/Lb78J/test/test_diag.jl | 264 2 266 7m04.7s DiagTensor basic functionality | 248 248 6m37.6s DiagTensor contractions | 9 9 8.8s DiagTensor contractions | 7 2 9 15.9s Test /home/pkgeval/.julia/packages/NDTensors/Lb78J/test/test_diagblocksparse.jl | 28 1 29 1m13.6s UniformDiagBlockSparseTensor basic functionality | 10 10 1.1s DiagBlockSparse off-diagonal (eltype=Float32) | 2 2 12.9s DiagBlockSparse off-diagonal (eltype=Float64) | 2 2 9.8s DiagBlockSparse off-diagonal (eltype=ComplexF32) | 2 2 11.7s DiagBlockSparse off-diagonal (eltype=ComplexF64) | 2 2 12.2s DiagBlockSparse contract | 6 6 9.2s DiagBlockSparse contract | 1 1 3.9s UniformDiagBlockSparse norm | 2 2 0.0s DiagBlockSparse denseblocks | 2 2 0.3s Test /home/pkgeval/.julia/packages/NDTensors/Lb78J/test/test_emptynumber.jl | 28 28 0.6s Test /home/pkgeval/.julia/packages/NDTensors/Lb78J/test/test_emptystorage.jl | 24 24 2.5s Test /home/pkgeval/.julia/packages/NDTensors/Lb78J/test/test_linearalgebra.jl | 454 454 55.3s Test /home/pkgeval/.julia/packages/NDTensors/Lb78J/test/test_tupletools.jl | 7 7 0.2s Test /home/pkgeval/.julia/packages/NDTensors/Lb78J/test/lib/runtests.jl | 196 2 198 1m30.7s Test NDTensors lib AMDGPUExtensions | 2 2 0.1s Test NDTensors lib BackendSelection | 12 12 0.8s Test NDTensors lib CUDAExtensions | 2 2 0.0s Test NDTensors lib GPUArraysCoreExtensions | 1 1 0.0s Test NDTensors lib MetalExtensions | 1 1 0.0s Test NDTensors lib Expose | 178 2 180 1m29.5s Testing Expose cpu, Float32 | 64 64 19.8s Testing Expose cpu, ComplexF32 | 64 64 20.6s Testing Expose jl, Float32 | 25 1 26 20.9s Testing Expose jl, ComplexF32 | 25 1 26 21.6s RNG of the outermost testset: Random.Xoshiro(0x50049fcbc282aa98, 0x8966d9952758dc85, 0x95225284a4e2fe68, 0xcacaff182832a4c0, 0x6171c2cb8a5eb96f) ERROR: LoadError: Some tests did not pass: 2708 passed, 0 failed, 12 errored, 0 broken. in expression starting at /home/pkgeval/.julia/packages/NDTensors/Lb78J/test/runtests.jl:3 Testing failed after 1199.2s ERROR: LoadError: Package NDTensors errored during testing Stacktrace: [1] pkgerror(msg::String) @ Pkg.Types /opt/julia/share/julia/stdlib/v1.13/Pkg/src/Types.jl:68 [2] test(ctx::Pkg.Types.Context, pkgs::Vector{PackageSpec}; coverage::Bool, julia_args::Cmd, test_args::Cmd, test_fn::Nothing, force_latest_compatible_version::Bool, allow_earlier_backwards_compatible_versions::Bool, allow_reresolve::Bool) @ Pkg.Operations /opt/julia/share/julia/stdlib/v1.13/Pkg/src/Operations.jl:2672 [3] test @ /opt/julia/share/julia/stdlib/v1.13/Pkg/src/Operations.jl:2521 [inlined] [4] test(ctx::Pkg.Types.Context, pkgs::Vector{PackageSpec}; coverage::Bool, test_fn::Nothing, julia_args::Cmd, test_args::Cmd, force_latest_compatible_version::Bool, allow_earlier_backwards_compatible_versions::Bool, allow_reresolve::Bool, kwargs::@Kwargs{io::IOContext{IO}}) @ Pkg.API /opt/julia/share/julia/stdlib/v1.13/Pkg/src/API.jl:538 [5] kwcall(::@NamedTuple{julia_args::Cmd, io::IOContext{IO}}, ::typeof(Pkg.API.test), ctx::Pkg.Types.Context, pkgs::Vector{PackageSpec}) @ Pkg.API /opt/julia/share/julia/stdlib/v1.13/Pkg/src/API.jl:515 [6] test(pkgs::Vector{PackageSpec}; io::IOContext{IO}, kwargs::@Kwargs{julia_args::Cmd}) @ Pkg.API /opt/julia/share/julia/stdlib/v1.13/Pkg/src/API.jl:168 [7] kwcall(::@NamedTuple{julia_args::Cmd}, ::typeof(Pkg.API.test), pkgs::Vector{PackageSpec}) @ Pkg.API /opt/julia/share/julia/stdlib/v1.13/Pkg/src/API.jl:157 [8] test(pkgs::Vector{String}; kwargs::@Kwargs{julia_args::Cmd}) @ Pkg.API /opt/julia/share/julia/stdlib/v1.13/Pkg/src/API.jl:156 [9] test @ /opt/julia/share/julia/stdlib/v1.13/Pkg/src/API.jl:156 [inlined] [10] kwcall(::@NamedTuple{julia_args::Cmd}, ::typeof(Pkg.API.test), pkg::String) @ Pkg.API /opt/julia/share/julia/stdlib/v1.13/Pkg/src/API.jl:155 [11] top-level scope @ /PkgEval.jl/scripts/evaluate.jl:219 [12] include(mod::Module, _path::String) @ Base ./Base.jl:311 [13] exec_options(opts::Base.JLOptions) @ Base ./client.jl:320 [14] _start() @ Base ./client.jl:553 in expression starting at /PkgEval.jl/scripts/evaluate.jl:210 PkgEval failed after 1293.09s: package tests unexpectedly errored