Am Samstag, den 18.08.2018, 09:27 +0200 schrieb Thomas Trepl: > Am Freitag, den 17.08.2018, 21:27 +0100 schrieb Ken Moffat: > > On Fri, Aug 17, 2018 at 08:57:02PM +0200, Thomas Trepl wrote: > > > > > > Rerun the tests here with "ninja check{,-clang{,-tooling}}" - not > > > "make". I used > > > > > > CC=gcc CXX=g++ \ > > > cmake -DCMAKE_INSTALL_PREFIX=/usr \ > > > -DLLVM_ENABLE_FFI=ON \ > > > -DCMAKE_BUILD_TYPE=Release \ > > > -DLLVM_BUILD_LLVM_DYLIB=ON \ > > > -DLLVM_TARGETS_TO_BUILD="host;AMDGPU" \ > > > -DLLVM_BUILD_TESTS=ON \ > > > -Wno-dev -G Ninja .. > > > > > > The summary looks fine to me: > > > > > > [0/3] Running the LLVM regression tests > > > -- Testing: 23298 tests, 4 threads -- > > > Testing: 0 .. 10.. 20.. 30.. 40.. 50.. 60.. 70.. 80.. 90.. > > > Testing Time: 106.64s > > > Expected Passes : 15113 > > > Expected Failures : 56 > > > Unsupported Tests : 8129 > > > [1/3] Running lit suite /tmp/llvm/build/llvm- > > > 6.0.1.src/tools/clang/test/Tooling > > > llvm-lit: /tmp/llvm/build/llvm- > > > 6.0.1.src/utils/lit/lit/llvm/config.py:334: note: using clang: > > > /tmp/l > > > lvm/build/llvm-6.0.1.src/build/bin/clang > > > -- Testing: 26 tests, 4 threads -- > > > Testing: 0 .. 10.. 20.. 30.. 40.. 50.. 60.. 70.. 80.. 90.. > > > Testing Time: 3.32s > > > Expected Passes : 26 > > > [2/3] Running the Clang regression tests > > > llvm-lit: /tmp/llvm/build/llvm- > > > 6.0.1.src/utils/lit/lit/llvm/config.py:334: note: using clang: > > > /tmp/llvm/build/llvm-6.0.1.src/build/bin/clang > > > -- Testing: 11832 tests, 4 threads -- > > > Testing: 0 .. 10.. 20.. 30.. 40.. 50.. 60.. 70.. 80.. 90.. > > > Testing Time: 135.52s > > > Expected Passes : 11572 > > > Expected Failures : 18 > > > Unsupported Tests : 242 > > > ... > > > > > > When running "ninja check-all" it looks like (same build > > > instructions, > > > clean build): > > > > > > ... > > > Testing: 0 .. 10.. 20.. 30.. 40.. 50.. 60.. 70.. 80.. 90.. > > > > > > 1 warning(s) in tests. > > > Testing Time: 615.44s > > > ******************** > > > Failing Tests (8): > > > LeakSanitizer-AddressSanitizer-x86_64 :: > > > TestCases/Linux/use_tls_dynamic.cc > > > LeakSanitizer-Standalone-x86_64 :: > > > TestCases/Linux/use_tls_dynamic.cc > > > MemorySanitizer-X86_64 :: Linux/sunrpc.cc > > > MemorySanitizer-X86_64 :: Linux/sunrpc_bytes.cc > > > MemorySanitizer-X86_64 :: Linux/sunrpc_string.cc > > > MemorySanitizer-X86_64 :: dtls_test.c > > > SanitizerCommon-lsan-x86_64-Linux :: > > > Posix/sanitizer_set_death_callback_test.cc > > > ThreadSanitizer-x86_64 :: sunrpc.cc > > > > > > Expected Passes : 29119 > > > Expected Failures : 103 > > > Unsupported Tests : 8914 > > > Unexpected Failures: 8 > > > FAILED: CMakeFiles/check-all > > > ... > > > > > > So, comparable to Ken's results but different anyhow - all > > > failures > > > have to with the rpc header files. Interesting that cmake checks > > > for > > > them and states that they are not found but continues (see the > > > first > > > two lines of following grep output), so i think they are no hard > > > prerequisites. The tests seems not taking care of not having the > > > headers available. I simply did a grep on the my log file: > > > > > > # grep "rpc/.*not found" llvm-check-all.log > > > -- Looking for rpc/xdr.h - not found > > > -- Looking for tirpc/rpc/xdr.h - not found> > > > /home/lfs/tmp/llvm/build/llvm-6.0.1.src/projects/compiler- > > > rt/test/msan/Linux/sunrpc.cc:15:10: fatal error: 'rpc/xdr.h' file > > > not > > > found > > > /home/lfs/tmp/llvm/build/llvm-6.0.1.src/projects/compiler- > > > rt/test/msan/Linux/sunrpc_bytes.cc:8:10: fatal error: 'rpc/xdr.h' > > > file > > > not found > > > /home/lfs/tmp/llvm/build/llvm-6.0.1.src/projects/compiler- > > > rt/test/msan/Linux/sunrpc_string.cc:8:10: fatal error: > > > 'rpc/xdr.h' > > > file > > > not found > > > /home/lfs/tmp/llvm/build/llvm-6.0.1.src/projects/compiler- > > > rt/test/tsan/sunrpc.cc:4:10: fatal error: 'rpc/types.h' file not > > > found > > > > > > Can we assume that those unexpected failures are cause by a flaw > > > in > > > the > > > test suite? > > > > > > > I think so. I was going to suggest dropping back to the targets DJ > > was using, but I see that there were a lot fewer expected passes > > there (11572 instead of 29119). As to the marginally different > > number of unexpected failures, probably a minor difference in what > > we have installed. > > > > > > Btw, just redoing build with 'make' to see if the hang is > > > reproducable > > > there. If not than it may have been caused by whatever reason... > > > > > > > Currently rerunning the base build (i.e. no docs) with > > hyperthreading (should be slower) and ninja -j4. It is > > consistently > > running at 400%, unlike the build on 4 real cores which mostly only > > ran at 100% and 200%. > > > > So, it now looks as if ninja makes a muh better job of the build > > than cmake's Makefiles. OTOH, building docs might be a lot > > messier. > > > > Still thinking about what will suit _me_ best, but if your run with > > 'make' works, don't let me stop you updating with your preference. > > 'make'-based build & test finished now with the same results as > 'ninja'-based. Dunno what the hang caused, the current build run on a > 8.3-rc1 system while the system where the hang occured was a 8.2 (but > heavily tweaked and updated). > > If there are no arguments against, i'd like to update the > instructions > to use ninja as it took 16.8 SBU while make needed 24.1 which is a > significant improvement. Building the docs runs well if sphinx is > installed, the only exception is that i had to clear the build- > directory before running the "cmake -DLLVM_ENABLE_SPHINX=ON..." > stuff. Argh, SBUs are wrong, got exactly same SBUs as Ken when calculating them correctly: SBU=29 (ninja, not 16.8) and 41.7 (make, not 24.1)
-- Thomas -- http://lists.linuxfromscratch.org/listinfo/blfs-dev FAQ: http://www.linuxfromscratch.org/blfs/faq.html Unsubscribe: See the above information page