Yes, above is helping. only some tests are failing now
Status `C:\Users\andre\AppData\Local\Temp\jl_zJKTY5\Manifest.toml`
[621f4979] AbstractFFTs v0.5.0
[79e6a3ab] Adapt v2.0.2
[b99e7846] BinaryProvider v0.5.10
[fa961155] CEnum v0.4.1
[052768ef] CUDA v1.2.0 [`C:\Src\cuda\cuda\dev\CUDA`]
[bbf7d656] CommonSubexpressions v0.3.0
[e66e0078] CompilerSupportLibraries_jll v0.3.3+0
[adafc99b] CpuId v0.2.2
[864edb3b] DataStructures v0.17.19
[163ba53b] DiffResults v1.0.2
[b552c78f] DiffRules v1.0.1
[ffbed154] DocStringExtensions v0.8.2
[e2ba6199] ExprTools v0.1.1
[7a1cc6ca] FFTW v1.2.2
[f5851436] FFTW_jll v3.3.9+5
[1a297f60] FillArrays v0.8.13
[f6369f11] ForwardDiff v0.10.12
[0c68f7d7] GPUArrays v4.0.1 #master (https://github.com/JuliaGPU/GPUArrays.jl)
[61eb1bfa] GPUCompiler v0.5.3
[1d5cc7b8] IntelOpenMP_jll v2018.0.3+0
[929cbde3] LLVM v2.0.0
[bdcacae8] LoopVectorization v0.8.19
[856f044c] MKL_jll v2020.1.216+0
[1914dd2f] MacroTools v0.5.5
[a6bfbf70] NNPACK_jll v2018.6.22+0
[872c559c] NNlib v0.7.3
[77ba4419] NaNMath v0.3.4
[6fe1bfb0] OffsetArrays v1.1.1
[efe28fd5] OpenSpecFun_jll v0.5.3+3
[bac558e1] OrderedCollections v1.3.0
[189a3867] Reexport v0.2.0
[ae029012] Requires v1.0.1
[21efa798] SIMDPirates v0.8.17
[476501e8] SLEEFPirates v0.5.5
[276daf66] SpecialFunctions v0.10.3
[90137ffa] StaticArrays v0.12.4
[a759f4b9] TimerOutputs v0.5.6
[3a884ed6] UnPack v1.0.1
[3d5dd08c] VectorizationBase v0.12.25
[2a0f44e3] Base64
[ade2ca70] Dates
[8ba89e20] Distributed
[b77e0a4c] InteractiveUtils
[76f85450] LibGit2
[8f399da3] Libdl
[37e2e46d] LinearAlgebra
[56ddb016] Logging
[d6f4376e] Markdown
[44cfe95a] Pkg
[de0858da] Printf
[3fa0cd96] REPL
[9a3f8284] Random
[ea8e919c] SHA
[9e88b42a] Serialization
[6462fe0b] Sockets
[2f01184e] SparseArrays
[10745b16] Statistics
[8dfed614] Test
[cf7118a7] UUIDs
[4ec0a83e] Unicode
┌ Info: System information:
│ CUDA toolkit 11.0.194, artifact installation
│ CUDA driver 11.0.0
│ NVIDIA driver 451.67.0
│
│ Libraries:
│ - CUBLAS: 11.1.0
│ - CURAND: 10.2.1
│ - CUFFT: 10.2.0
│ - CUSOLVER: 10.5.0
│ - CUSPARSE: 11.1.0
│ - CUPTI: 13.0.0
│ - NVML: 11.0.0+451.67
│ - CUDNN: missing
│ - CUTENSOR: missing
│
│ Toolchain:
│ - Julia: 1.4.2
│ - LLVM: 8.0.1
│ - PTX ISA support: 3.2, 4.0, 4.1, 4.2, 4.3, 5.0, 6.0, 6.1, 6.3
│ - Device support: sm_35, sm_37, sm_50, sm_52, sm_53, sm_60, sm_61, sm_62, sm_70, sm_72, sm_75
│
│ Environment:
│ - JULIA_CUDA_VERBOSE: true
│
│ 2 device(s):
│ - GeForce RTX 2070 (sm_75, 7.843 GiB / 8.000 GiB available)
└ - GeForce RTX 2080 Ti (sm_75, 10.790 GiB / 11.000 GiB available)
[ Info: Testing using 1 device(s): 2. GeForce RTX 2080 Ti (UUID 4d228dab-1287-5f37-f4db-e53f82e2bc60)
[ Info: Skipping the following tests: cudnn, cutensor, device\wmma
| | ---------------- GPU ---------------- | ---------------- CPU ---------------- |
Test (Worker) | Time (s) | GC (s) | GC % | Alloc (MB) | RSS (MB) | GC (s) | GC % | Alloc (MB) | RSS (MB) |
initialization (2) | 3.65 | 0.00 | 0.0 | 0.00 | N/A | 0.05 | 1.4 | 199.20 | 562.26 |
apiutils (3) | 0.81 | 0.00 | 0.0 | 0.00 | N/A | 0.03 | 3.5 | 85.70 | 552.87 |
iterator (3) | 2.71 | 0.27 | 10.0 | 1.25 | N/A | 0.09 | 3.4 | 281.28 | 552.87 |
memory (3) | 1.53 | 0.00 | 0.0 | 0.00 | N/A | 0.33 | 21.2 | 142.82 | 552.87 |
codegen (6) | failed at 2020-07-24T20:12:39.053
nnlib (3) | 13.92 | 0.15 | 1.1 | 0.00 | N/A | 0.59 | 4.2 | 1550.51 | 980.92 |
nvml (3) | 0.51 | 0.00 | 0.0 | 0.00 | N/A | 0.02 | 3.4 | 58.52 | 986.84 |
nvtx (3) | 0.49 | 0.00 | 0.0 | 0.00 | N/A | 0.03 | 6.8 | 63.11 | 990.04 |
pointer (3) | 0.11 | 0.00 | 0.0 | 0.00 | N/A | 0.00 | 0.0 | 7.06 | 990.66 |
curand (9) | 23.48 | 0.26 | 1.1 | 0.02 | N/A | 1.18 | 5.0 | 2416.18 | 634.97 |
statistics (3) | 16.18 | 0.00 | 0.0 | 0.00 | N/A | 0.81 | 5.0 | 1582.73 | 1046.86 |
utils (3) | 0.85 | 0.00 | 0.0 | 0.00 | N/A | 0.02 | 2.7 | 118.93 | 1046.86 |
cufft (8) | 38.56 | 0.27 | 0.7 | 133.23 | N/A | 1.97 | 5.1 | 4257.42 | 1189.55 |
cudadrv\context (3) | 0.58 | 0.00 | 0.0 | 0.00 | N/A | 0.02 | 2.8 | 32.43 | 1099.50 |
cudadrv\devices (8) | 0.34 | 0.00 | 0.0 | 0.00 | N/A | 0.00 | 0.0 | 40.12 | 1189.55 |
broadcast (5) | 39.32 | 0.27 | 0.7 | 0.00 | N/A | 1.56 | 4.0 | 3605.98 | 677.96 |
cudadrv\errors (3) | 0.15 | 0.00 | 0.0 | 0.00 | N/A | 0.00 | 0.0 | 21.96 | 1099.50 |
cudadrv\events (8) | 0.15 | 0.00 | 0.0 | 0.00 | N/A | 0.00 | 0.0 | 14.93 | 1189.55 |
cudadrv\module (8) | 0.37 | 0.00 | 0.0 | 0.00 | N/A | 0.01 | 3.2 | 46.07 | 1189.55 |
cudadrv\occupancy (8) | 0.11 | 0.00 | 0.0 | 0.00 | N/A | 0.00 | 0.0 | 8.55 | 1189.55 |
cudadrv\profile (8) | 0.30 | 0.00 | 0.0 | 0.00 | N/A | 0.01 | 4.8 | 49.13 | 1189.55 |
cudadrv\execution (5) | 1.01 | 0.00 | 0.0 | 0.00 | N/A | 0.05 | 5.4 | 108.96 | 677.96 |
cudadrv\version (5) | 0.01 | 0.00 | 0.0 | 0.00 | N/A | 0.00 | 0.0 | 0.08 | 677.96 |
cudadrv\stream (8) | 0.21 | 0.00 | 0.0 | 0.00 | N/A | 0.02 | 9.0 | 20.06 | 1189.55 |
cudadrv\memory (3) | 2.10 | 0.00 | 0.0 | 0.00 | N/A | 0.07 | 3.5 | 206.39 | 1099.50 |
device\array (8) | 3.58 | 0.00 | 0.0 | 0.00 | N/A | 0.10 | 2.8 | 339.27 | 1194.57 |
texture (9) | 26.26 | 0.00 | 0.0 | 0.08 | N/A | 0.92 | 3.5 | 2530.68 | 720.41 |
device\pointer (8) | 7.09 | 0.00 | 0.0 | 0.00 | N/A | 0.20 | 2.9 | 568.32 | 1208.79 |
gpuarrays/math (8) | 2.93 | 0.00 | 0.0 | 0.00 | N/A | 0.09 | 3.2 | 306.95 | 1217.26 |
gpuarrays/input output (8) | 1.19 | 0.00 | 0.0 | 0.00 | N/A | 0.04 | 3.7 | 126.49 | 1217.26 |
cusparse (11) | 56.99 | 0.32 | 0.6 | 4.46 | N/A | 2.41 | 4.2 | 6253.25 | 967.00 |
threading (14) | 16.95 | 4.44 | 26.2 | 56.25 | N/A | 0.83 | 4.9 | 1801.75 | 1274.68 |
forwarddiff (2) | 58.55 | 0.30 | 0.5 | 0.00 | N/A | 1.32 | 2.3 | 3334.10 | 737.97 |
gpuarrays/interface (11) | 3.34 | 0.00 | 0.0 | 0.00 | N/A | 0.11 | 3.4 | 253.39 | 975.28 |
cusolver\cusparse (5) | 19.97 | 0.00 | 0.0 | 0.19 | N/A | 0.84 | 4.2 | 2063.14 | 1441.06 |
gpuarrays/value constructors (8) | 7.89 | 0.00 | 0.0 | 0.00 | N/A | 0.17 | 2.2 | 573.97 | 1238.47 |
gpuarrays/fft (8) | 1.24 | 0.00 | 0.1 | 6.01 | N/A | 0.04 | 3.3 | 133.59 | 1247.52 |
gpuarrays/conversions (5) | 4.46 | 0.00 | 0.0 | 0.01 | N/A | 0.19 | 4.3 | 626.51 | 1448.38 |
gpuarrays/constructors (8) | 1.42 | 0.00 | 0.2 | 0.04 | N/A | 0.01 | 0.9 | 76.52 | 1249.82 |
gpuarrays/indexing (9) | 17.71 | 0.00 | 0.0 | 0.12 | N/A | 0.57 | 3.2 | 1458.04 | 750.68 |
gpuarrays/uniformscaling (2) | 8.12 | 0.00 | 0.0 | 0.01 | N/A | 0.26 | 3.2 | 762.20 | 737.97 |
array (4) | 73.13 | 0.20 | 0.3 | 5.20 | N/A | 2.67 | 3.7 | 6785.38 | 784.63 |
cublas (7) | 75.23 | 0.35 | 0.5 | 11.12 | N/A | 3.19 | 4.2 | 8831.23 | 1199.72 |
gpuarrays/iterator constructors (14) | 19.32 | 0.00 | 0.0 | 0.02 | N/A | 0.62 | 3.2 | 1829.22 | 1278.43 |
gpuarrays/base (8) | 15.86 | 0.00 | 0.0 | 17.61 | N/A | 0.60 | 3.8 | 1872.71 | 1348.77 |
cusolver (10) | 83.64 | 0.42 | 0.5 | 1229.85 | N/A | 3.46 | 4.1 | 8887.82 | 1574.50 |
gpuarrays/random (5) | 19.85 | 0.00 | 0.0 | 0.00 | N/A | 0.53 | 2.7 | 1656.92 | 1505.24 |
device\intrinsics (3) | 67.58 | 0.00 | 0.0 | 0.01 | N/A | 1.30 | 1.9 | 5025.86 | 1201.33 |
gpuarrays/linear algebra (11) | 48.93 | 0.01 | 0.0 | 1.42 | N/A | 1.53 | 3.1 | 5037.09 | 1463.40 |
examples (12) | 124.64 | 0.00 | 0.0 | 0.00 | N/A | 0.08 | 0.1 | 132.75 | 562.43 |
gpuarrays/broadcasting (2) | 60.48 | 0.00 | 0.0 | 1.19 | N/A | 2.02 | 3.3 | 7448.78 | 853.78 |
execution (13) | failed at 2020-07-24T20:14:35.877
gpuarrays/mapreduce essentials (9) | 89.02 | 0.01 | 0.0 | 3.19 | N/A | 2.43 | 2.7 | 10171.24 | 980.95 |
gpuarrays/mapreduce (old tests) (7) | 92.07 | 0.01 | 0.0 | 130.20 | N/A | 2.21 | 2.4 | 10276.32 | 1506.56 |
gpuarrays/mapreduce derivatives (4) | 126.02 | 0.01 | 0.0 | 3.06 | N/A | 3.02 | 2.4 | 13688.27 | 1117.89 |
Worker 6 failed running test codegen:
Some tests did not pass: 17 passed, 1 failed, 0 errored, 0 broken.
codegen: Test Failed at C:\Src\cuda\cuda\dev\CUDA\test\codegen.jl:78
Expression: !(occursin("gpu_report_exception", ir))
Evaluated: !(occursin("gpu_report_exception", "\n; @ C:\\Src\\\cuda\\cuda\\dev\\CUDA\\test\\codegen.jl:76 within `bar'\n; Function Attrs: uwtable\ndefine i64 @julia_bar_19727(i64) local_unnamed_addr #0 {\ntop:\n; ┌ @ operators.jl:294 within `>'\n; │┌ @ int.jl:49 within `<'\n %1 = icmp sgt i64 %0, 0\n; └└\n; ┌ @ C:\\Src\\\cuda\\cuda\\dev\\CUDA\\src\\device\\llvm.jl:5 within `assume'\n %2 = call fastcc %jl_value_t addrspace(10)* @gpu_gc_pool_alloc(i64 16)\n %3 = addrspacecast %jl_value_t addrspace(10)* %2 to %jl_value_t addrspace(11)*\n %4 = bitcast %jl_value_t addrspace(10)* %2 to %jl_value_t addrspace(10)* addrspace(10)*\n store %jl_value_t addrspace(10)* addrspacecast (%jl_value_t* inttoptr (i64 1023020288 to %jl_value_t*) to %jl_value_t addrspace(10)*), %jl_value_t addrspace(10)* addrspace(10)* %4, align 8\n %5 = bitcast %jl_value_t addrspace(11)* %3 to i8 addrspace(11)*\n %6 = getelementptr inbounds i8, i8 addrspace(11)* %5, i64 8\n %7
= bitcast i8 addrspace(11)* %6 to %jl_value_t addrspace(10)* addrspace(11)*\n store %jl_value_t addrspace(10)* addrspacecast (%jl_value_t* inttoptr (i64 370085648 to %jl_value_t*) to %jl_value_t addrspace(10)*), %jl_value_t addrspace(10)* addrspace(11)* %7, align 8\n call void @llvm.assume(i1 %1)\n; └\n; ┌ @ div.jl:227 within `cld'\n; │┌ @ div.jl:270 within `div' @ div.jl:215 @ int.jl:230\n %8 = icmp eq i64 %0, 0\n br i1 %8, label %fail, label %pass\n\nfail:
; preds = %top\n call fastcc void @gpu_report_exception(i64 ptrtoint ([10 x i8]* @exception to
i64))\n call fastcc void @gpu_signal_exception()\n call void asm sideeffect \"exit;\", \"\"() #1\n br label %pass\n\npass: ; preds = %top, %fail\n %9 = udiv i64 42, %0\n; ││ @ div.jl:271 within `div'\n; ││┌ @ int.jl:54 within `*'\n %10 = mul i64 %9, %0\n; ││└\n; ││┌ @ operators.jl:193 within `!='\n; │││┌ @ promotion.jl:398 within `=='\n %11 = icmp ne i64 %10, 42\n; ││└└\n; ││┌ @ int.jl:858 within `+'\n; │││┌ @ int.jl:442 within `rem'\n; ││││┌ @ number.jl:7 within `convert'\n; │││││┌ @ boot.jl:707 within `Int64'\n; ││││││┌ @ boot.jl:634 within `toInt64'\n %12 = zext i1 %11 to i64\n; │││└└└└\n; │││ @ int.jl:860 within `+' @ int.jl:53\n %13 = add nuw nsw i64 %9, %12\n; └└└\n ret i64 %13\n}\n"))
Stacktrace:
[1] record(::Test.DefaultTestSet, ::Union{Test.Error, Test.Fail}) at D:\buildbot\worker\package_win64\build\usr\share\julia\stdlib\v1.4\Test\src\Test.jl:775
[2] top-level scope at C:\Src\cuda\cuda\dev\CUDA\test\runtests.jl:444
[3] include(::String) at .\client.jl:439
[4] top-level scope at none:6
[5] eval(::Module, ::Any) at .\boot.jl:331
[6] exec_options(::Base.JLOptions) at .\client.jl:264
[7] _start() at .\client.jl:484
Worker 13 failed running test execution:
Some tests did not pass: 72 passed, 8 failed, 0 errored, 0 broken.
execution: Test Failed at C:\Src\cuda\cuda\dev\CUDA\test\execution.jl:568
Expression: code == 1
Evaluated: 0 == 1
Stacktrace:
[1] record(::Test.DefaultTestSet, ::Union{Test.Error, Test.Fail}) at D:\buildbot\worker\package_win64\build\usr\share\julia\stdlib\v1.4\Test\src\Test.jl:775
[2] top-level scope at C:\Src\cuda\cuda\dev\CUDA\test\runtests.jl:444
[3] include(::String) at .\client.jl:439
[4] top-level scope at none:6
[5] eval(::Module, ::Any) at .\boot.jl:331
[6] exec_options(::Base.JLOptions) at .\client.jl:264
[7] _start() at .\client.jl:484
execution: Test Failed at C:\Src\cuda\cuda\dev\CUDA\test\execution.jl:569
Expression: occursin("ERROR: KernelException: exception thrown during kernel execution on device", err)
Evaluated: occursin("ERROR: KernelException: exception thrown during kernel execution on device", "error in running finalizer: CUDA.KernelException(dev=CUDA.CuDevice(handle=0))\n")
Stacktrace:
[1] record(::Test.DefaultTestSet, ::Union{Test.Error, Test.Fail}) at D:\buildbot\worker\package_win64\build\usr\share\julia\stdlib\v1.4\Test\src\Test.jl:775
[2] top-level scope at C:\Src\cuda\cuda\dev\CUDA\test\runtests.jl:444
[3] include(::String) at .\client.jl:439
[4] top-level scope at none:6
[5] eval(::Module, ::Any) at .\boot.jl:331
[6] exec_options(::Base.JLOptions) at .\client.jl:264
[7] _start() at .\client.jl:484
execution: Test Failed at C:\Src\cuda\cuda\dev\CUDA\test\execution.jl:575
Expression: code == 1
Evaluated: 0 == 1
Stacktrace:
[1] record(::Test.DefaultTestSet, ::Union{Test.Error, Test.Fail}) at D:\buildbot\worker\package_win64\build\usr\share\julia\stdlib\v1.4\Test\src\Test.jl:775
[2] top-level scope at C:\Src\cuda\cuda\dev\CUDA\test\runtests.jl:444
[3] include(::String) at .\client.jl:439
[4] top-level scope at none:6
[5] eval(::Module, ::Any) at .\boot.jl:331
[6] exec_options(::Base.JLOptions) at .\client.jl:264
[7] _start() at .\client.jl:484
execution: Test Failed at C:\Src\cuda\cuda\dev\CUDA\test\execution.jl:576
Expression: occursin("ERROR: KernelException: exception thrown during kernel execution on device", err)
Evaluated: occursin("ERROR: KernelException: exception thrown during kernel execution on device", "error in running finalizer: CUDA.KernelException(dev=CUDA.CuDevice(handle=0))\n")
Stacktrace:
[1] record(::Test.DefaultTestSet, ::Union{Test.Error, Test.Fail}) at D:\buildbot\worker\package_win64\build\usr\share\julia\stdlib\v1.4\Test\src\Test.jl:775
[2] top-level scope at C:\Src\cuda\cuda\dev\CUDA\test\runtests.jl:444
[3] include(::String) at .\client.jl:439
[4] top-level scope at none:6
[5] eval(::Module, ::Any) at .\boot.jl:331
[6] exec_options(::Base.JLOptions) at .\client.jl:264
[7] _start() at .\client.jl:484
execution: Test Failed at C:\Src\cuda\cuda\dev\CUDA\test\execution.jl:582
Expression: code == 1
Evaluated: 0 == 1
Stacktrace:
[1] record(::Test.DefaultTestSet, ::Union{Test.Error, Test.Fail}) at D:\buildbot\worker\package_win64\build\usr\share\julia\stdlib\v1.4\Test\src\Test.jl:775
[2] top-level scope at C:\Src\cuda\cuda\dev\CUDA\test\runtests.jl:444
[3] include(::String) at .\client.jl:439
[4] top-level scope at none:6
[5] eval(::Module, ::Any) at .\boot.jl:331
[6] exec_options(::Base.JLOptions) at .\client.jl:264
[7] _start() at .\client.jl:484
execution: Test Failed at C:\Src\cuda\cuda\dev\CUDA\test\execution.jl:583
Expression: occursin("ERROR: KernelException: exception thrown during kernel execution on device", err)
Evaluated: occursin("ERROR: KernelException: exception thrown during kernel execution on device", "error in running finalizer: CUDA.KernelException(dev=CUDA.CuDevice(handle=0))\n")
Stacktrace:
[1] record(::Test.DefaultTestSet, ::Union{Test.Error, Test.Fail}) at D:\buildbot\worker\package_win64\build\usr\share\julia\stdlib\v1.4\Test\src\Test.jl:775
[2] top-level scope at C:\Src\cuda\cuda\dev\CUDA\test\runtests.jl:444
[3] include(::String) at .\client.jl:439
[4] top-level scope at none:6
[5] eval(::Module, ::Any) at .\boot.jl:331
[6] exec_options(::Base.JLOptions) at .\client.jl:264
[7] _start() at .\client.jl:484
execution: Test Failed at C:\Src\cuda\cuda\dev\CUDA\test\execution.jl:609
Expression: code == 1
Evaluated: 0 == 1
Stacktrace:
[1] record(::Test.DefaultTestSet, ::Union{Test.Error, Test.Fail}) at D:\buildbot\worker\package_win64\build\usr\share\julia\stdlib\v1.4\Test\src\Test.jl:775
[2] top-level scope at C:\Src\cuda\cuda\dev\CUDA\test\runtests.jl:444
[3] include(::String) at .\client.jl:439
[4] top-level scope at none:6
[5] eval(::Module, ::Any) at .\boot.jl:331
[6] exec_options(::Base.JLOptions) at .\client.jl:264
[7] _start() at .\client.jl:484
execution: Test Failed at C:\Src\cuda\cuda\dev\CUDA\test\execution.jl:610
Expression: occursin("ERROR: KernelException: exception thrown during kernel execution on device", err)
Evaluated: occursin("ERROR: KernelException: exception thrown during kernel execution on device", "error in running finalizer: CUDA.KernelException(dev=CUDA.CuDevice(handle=0))\n")
Stacktrace:
[1] record(::Test.DefaultTestSet, ::Union{Test.Error, Test.Fail}) at D:\buildbot\worker\package_win64\build\usr\share\julia\stdlib\v1.4\Test\src\Test.jl:775
[2] top-level scope at C:\Src\cuda\cuda\dev\CUDA\test\runtests.jl:444
[3] include(::String) at .\client.jl:439
[4] top-level scope at none:6
[5] eval(::Module, ::Any) at .\boot.jl:331
[6] exec_options(::Base.JLOptions) at .\client.jl:264
[7] _start() at .\client.jl:484
Test Summary: | Pass Fail Broken Total
Overall | 8167 9 2 8178
initialization | 11 11
apiutils | 15 15
iterator | 30 30
memory | 10 10
codegen | 17 1 18
nnlib | 3 3
nvml | 7 7
nvtx | No tests
pointer | 13 13
curand | 102 102
statistics | 12 12
utils | 5 5
cufft | 151 151
cudadrv\context | 12 12
cudadrv\devices | 5 5
broadcast | 29 29
cudadrv\errors | 6 6
cudadrv\events | 6 6
cudadrv\module | 11 11
cudadrv\occupancy | 1 1
cudadrv\profile | 2 2
cudadrv\execution | 15 15
cudadrv\version | 3 3
cudadrv\stream | 7 7
cudadrv\memory | 50 1 51
device\array | 20 20
texture | 26 1 27
device\pointer | 57 57
gpuarrays/math | 8 8
gpuarrays/input output | 5 5
cusparse | 468 468
threading | No tests
forwarddiff | 106 106
gpuarrays/interface | 7 7
cusolver\cusparse | 84 84
gpuarrays/value constructors | 120 120
gpuarrays/fft | 12 12
gpuarrays/conversions | 72 72
gpuarrays/constructors | 335 335
gpuarrays/indexing | 113 113
gpuarrays/uniformscaling | 56 56
array | 154 154
cublas | 1885 1885
gpuarrays/iterator constructors | 24 24
gpuarrays/base | 38 38
cusolver | 1493 1493
gpuarrays/random | 40 40
device\intrinsics | 265 265
gpuarrays/linear algebra | 393 393
examples | 7 7
gpuarrays/broadcasting | 155 155
execution | 72 8 80
gpuarrays/mapreduce essentials | 522 522
gpuarrays/mapreduce (old tests) | 297 297
gpuarrays/mapreduce derivatives | 810 810
FAILURE
Error in testset codegen:
Test Failed at C:\Src\cuda\cuda\dev\CUDA\test\codegen.jl:78
Expression: !(occursin("gpu_report_exception", ir))
Evaluated: !(occursin("gpu_report_exception", "\n; @ C:\\Src\\\cuda\\cuda\\dev\\CUDA\\test\\codegen.jl:76 within `bar'\n; Function Attrs: uwtable\ndefine i64 @julia_bar_19727(i64) local_unnamed_addr #0 {\ntop:\n; ┌ @ operators.jl:294 within `>'\n; │┌ @ int.jl:49 within `<'\n %1 = icmp sgt i64 %0, 0\n; └└\n; ┌ @ C:\\Src\\\cuda\\cuda\\dev\\CUDA\\src\\device\\llvm.jl:5 within `assume'\n %2 = call fastcc %jl_value_t addrspace(10)* @gpu_gc_pool_alloc(i64 16)\n %3 = addrspacecast %jl_value_t addrspace(10)* %2 to %jl_value_t addrspace(11)*\n %4 = bitcast %jl_value_t addrspace(10)* %2 to %jl_value_t addrspace(10)* addrspace(10)*\n store %jl_value_t addrspace(10)* addrspacecast (%jl_value_t* inttoptr (i64 1023020288 to %jl_value_t*) to %jl_value_t addrspace(10)*), %jl_value_t addrspace(10)* addrspace(10)* %4, align 8\n %5 = bitcast %jl_value_t addrspace(11)* %3 to i8 addrspace(11)*\n %6 = getelementptr inbounds i8, i8 addrspace(11)* %5, i64 8\n %7
= bitcast i8 addrspace(11)* %6 to %jl_value_t addrspace(10)* addrspace(11)*\n store %jl_value_t addrspace(10)* addrspacecast (%jl_value_t* inttoptr (i64 370085648 to %jl_value_t*) to %jl_value_t addrspace(10)*), %jl_value_t addrspace(10)* addrspace(11)* %7, align 8\n call void @llvm.assume(i1 %1)\n; └\n; ┌ @ div.jl:227 within `cld'\n; │┌ @ div.jl:270 within `div' @ div.jl:215 @ int.jl:230\n %8 = icmp eq i64 %0, 0\n br i1 %8, label %fail, label %pass\n\nfail:
; preds = %top\n call fastcc void @gpu_report_exception(i64 ptrtoint ([10 x i8]* @exception to
i64))\n call fastcc void @gpu_signal_exception()\n call void asm sideeffect \"exit;\", \"\"() #1\n br label %pass\n\npass: ; preds = %top, %fail\n %9 = udiv i64 42, %0\n; ││ @ div.jl:271 within `div'\n; ││┌ @ int.jl:54 within `*'\n %10 = mul i64 %9, %0\n; ││└\n; ││┌ @ operators.jl:193 within `!='\n; │││┌ @ promotion.jl:398 within `=='\n %11 = icmp ne i64 %10, 42\n; ││└└\n; ││┌ @ int.jl:858 within `+'\n; │││┌ @ int.jl:442 within `rem'\n; ││││┌ @ number.jl:7 within `convert'\n; │││││┌ @ boot.jl:707 within `Int64'\n; ││││││┌ @ boot.jl:634 within `toInt64'\n %12 = zext i1 %11 to i64\n; │││└└└└\n; │││ @ int.jl:860 within `+' @ int.jl:53\n %13 = add nuw nsw i64 %9, %12\n; └└└\n ret i64 %13\n}\n"))
Error in testset execution:
Test Failed at C:\Src\cuda\cuda\dev\CUDA\test\execution.jl:568
Expression: code == 1
Evaluated: 0 == 1
Error in testset execution:
Test Failed at C:\Src\cuda\cuda\dev\CUDA\test\execution.jl:569
Expression: occursin("ERROR: KernelException: exception thrown during kernel execution on device", err)
Evaluated: occursin("ERROR: KernelException: exception thrown during kernel execution on device", "error in running finalizer: CUDA.KernelException(dev=CUDA.CuDevice(handle=0))\n")
Error in testset execution:
Test Failed at C:\Src\cuda\cuda\dev\CUDA\test\execution.jl:575
Expression: code == 1
Evaluated: 0 == 1
Error in testset execution:
Test Failed at C:\Src\cuda\cuda\dev\CUDA\test\execution.jl:576
Expression: occursin("ERROR: KernelException: exception thrown during kernel execution on device", err)
Evaluated: occursin("ERROR: KernelException: exception thrown during kernel execution on device", "error in running finalizer: CUDA.KernelException(dev=CUDA.CuDevice(handle=0))\n")
Error in testset execution:
Test Failed at C:\Src\cuda\cuda\dev\CUDA\test\execution.jl:582
Expression: code == 1
Evaluated: 0 == 1
Error in testset execution:
Test Failed at C:\Src\cuda\cuda\dev\CUDA\test\execution.jl:583
Expression: occursin("ERROR: KernelException: exception thrown during kernel execution on device", err)
Evaluated: occursin("ERROR: KernelException: exception thrown during kernel execution on device", "error in running finalizer: CUDA.KernelException(dev=CUDA.CuDevice(handle=0))\n")
Error in testset execution:
Test Failed at C:\Src\cuda\cuda\dev\CUDA\test\execution.jl:609
Expression: code == 1
Evaluated: 0 == 1
Error in testset execution:
Test Failed at C:\Src\cuda\cuda\dev\CUDA\test\execution.jl:610
Expression: occursin("ERROR: KernelException: exception thrown during kernel execution on device", err)
Evaluated: occursin("ERROR: KernelException: exception thrown during kernel execution on device", "error in running finalizer: CUDA.KernelException(dev=CUDA.CuDevice(handle=0))\n")
ERROR: LoadError: Test run finished with errors
in expression starting at C:\Src\cuda\cuda\dev\CUDA\test\runtests.jl:475