[package - 133releng-armv7-default][misc/py-llama-cpp-python] Failed for py311-llama-cpp-python-0.3.2 in build
- Go to: [ bottom of page ] [ top of archives ] [ this month ]
Date: Tue, 03 Dec 2024 16:57:42 UTC
You are receiving this mail as a port that you maintain is failing to build on the FreeBSD package build server. Please investigate the failure and submit a PR to fix build. Maintainer: yuri@FreeBSD.org Log URL: https://pkg-status.freebsd.org/ampere3/data/133releng-armv7-default/884c365e2398/logs/py311-llama-cpp-python-0.3.2.log Build URL: https://pkg-status.freebsd.org/ampere3/build.html?mastername=133releng-armv7-default&build=884c365e2398 Log: =>> Building misc/py-llama-cpp-python build started at Tue Dec 3 16:56:35 UTC 2024 port directory: /usr/ports/misc/py-llama-cpp-python package name: py311-llama-cpp-python-0.3.2 building for: FreeBSD 133releng-armv7-default-job-13 13.3-RELEASE-p8 FreeBSD 13.3-RELEASE-p8 1303001 arm maintained by: yuri@FreeBSD.org Makefile datestamp: -rw-r--r-- 1 root wheel 1736 Nov 28 01:06 /usr/ports/misc/py-llama-cpp-python/Makefile Ports top last git commit: 884c365e239 Ports top unclean checkout: no Port dir last git commit: f6f441f6827 Port dir unclean checkout: no Poudriere version: poudriere-git-3.4.2 Host OSVERSION: 1500023 Jail OSVERSION: 1303001 Job Id: 13 ---Begin Environment--- SHELL=/bin/csh OSVERSION=1303001 UNAME_v=FreeBSD 13.3-RELEASE-p8 1303001 UNAME_r=13.3-RELEASE-p8 BLOCKSIZE=K MAIL=/var/mail/root MM_CHARSET=UTF-8 LANG=C.UTF-8 STATUS=1 HOME=/root PATH=/sbin:/bin:/usr/sbin:/usr/bin:/usr/local/sbin:/usr/local/bin:/root/bin MAKE_OBJDIR_CHECK_WRITABLE=0 LOCALBASE=/usr/local USER=root POUDRIERE_NAME=poudriere-git LIBEXECPREFIX=/usr/local/libexec/poudriere POUDRIERE_VERSION=3.4.2 MASTERMNT=/usr/local/poudriere/data/.m/133releng-armv7-default/ref LC_COLLATE=C POUDRIERE_BUILD_TYPE=bulk PACKAGE_BUILDING=yes SAVED_TERM= OUTPUT_REDIRECTED_STDERR=4 OUTPUT_REDIRECTED=1 PWD=/usr/local/poudriere/data/.m/133releng-armv7-default/13/.p OUTPUT_REDIRECTED_STDOUT=3 P_PORTS_FEATURES=FLAVORS SUBPACKAGES SELECTED_OPTIONS MASTERNAME=133releng-armv7-default SCRIPTPREFIX=/usr/local/share/poudriere SCRIPTNAME=bulk.sh OLDPWD=/usr/local/poudriere/data/.m/133releng-armv7-default/ref/.p/pool POUDRIERE_PKGNAME=poudriere-git-3.4.2 SCRIPTPATH=/usr/local/share/poudriere/bulk.sh POUDRIEREPATH=/usr/local/bin/poudriere ---End Environment--- ---Begin Poudriere Port Flags/Env--- PORT_FLAGS= PKGENV= FLAVOR=py311 MAKE_ARGS= FLAVOR=py311 ---End Poudriere Port Flags/Env--- ---Begin OPTIONS List--- ---End OPTIONS List--- --MAINTAINER-- yuri@FreeBSD.org --End MAINTAINER-- --CONFIGURE_ARGS-- --End CONFIGURE_ARGS-- --CONFIGURE_ENV-- PYTHON="/usr/local/bin/python3.11" XDG_DATA_HOME=/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311 XDG_CONFIG_HOME=/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311 XDG_CACHE_HOME=/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/.cache HOME=/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311 TMPDIR="/tmp" PATH=/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/.bin:/sbin:/bin:/usr/sbin:/usr/bin:/usr/local/sbin:/usr/local/bin:/root/bin PKG_CONFIG_LIBDIR=/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/.pkgconfig:/usr/local/libdata/pkgconfig:/usr/local/share/pkgconfig:/usr/libdata/pkgconfig SHELL=/bin/sh CONFIG_SHELL=/bin/sh --End CONFIGURE_ENV-- --MAKE_ENV-- XDG_DATA_HOME=/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311 XDG_CONFIG_HOME=/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311 XDG_CACHE_HOME=/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/.cache HOME=/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311 TMPDIR="/tmp" PATH=/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/.bin:/sbin:/bin:/usr/sbin:/usr/bin:/usr/local/sbin:/usr/local/bin:/root/bin PKG_CONFIG_LIBDIR=/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/.pkgconfig:/usr/local/libdata/pkgconfig:/usr/local/share/pkgconfig:/usr/libdata/pkgconfig MK_DEBUG_FILES=no MK_KERNEL_SYMBOLS=no SHELL=/bin/sh NO_LINT=YES PREFIX=/usr/local LOCALBASE=/usr/local CC="cc" CFLAGS="-O2 -pipe -fstack-protector-strong -fno-strict-aliasing " CPP="cpp" CPPFLAGS="" LDFLAGS=" -fstack-protector-strong " LIBS="" CXX="c++" CXXFLAGS="-O2 -pipe -fstack-protector-strong -fno-strict-aliasing " BSD_INSTALL_PROGRAM="install -s -m 555" BSD_INSTALL_LIB="install -s -m 0644" BSD_INSTALL_SCRIPT="install -m 555" BSD_INSTALL_DATA="install -m 0644" BSD_INSTALL_MAN="install -m 444" --End MAKE_ENV-- --PLIST_SUB-- PYTHON_INCLUDEDIR=include/python3.11 PYTHON_LIBDIR=lib/python3.11 PYTHON_PLATFORM=freebsd13 PYTHON_SITELIBDIR=lib/python3.11/site-packages PYTHON_SUFFIX=311 PYTHON_EXT_SUFFIX=.cpython-311 PYTHON_VER=3.11 PYTHON_VERSION=python3.11 PYTHON2="@comment " PYTHON3="" OSREL=13.3 PREFIX=%D LOCALBASE=/usr/local RESETPREFIX=/usr/local LIB32DIR=lib DOCSDIR="share/doc/llama-cpp-python" EXAMPLESDIR="share/examples/llama-cpp-python" DATADIR="share/llama-cpp-python" WWWDIR="www/llama-cpp-python" ETCDIR="etc/llama-cpp-python" --End PLIST_SUB-- --SUB_LIST-- PYTHON_INCLUDEDIR=/usr/local/include/python3.11 PYTHON_LIBDIR=/usr/local/lib/python3.11 PYTHON_PLATFORM=freebsd13 PYTHON_SITELIBDIR=/usr/local/lib/python3.11/site-packages PYTHON_SUFFIX=311 PYTHON_EXT_SUFFIX=.cpython-311 PYTHON_VER=3.11 PYTHON_VERSION=python3.11 PYTHON2="@comment " PYTHON3="" PREFIX=/usr/local LOCALBASE=/usr/local DATADIR=/usr/local/share/llama-cpp-python DOCSDIR=/usr/local/share/doc/llama-cpp-python EXAMPLESDIR=/usr/local/share/examples/llama-cpp-python WWWDIR=/usr/local/www/llama-cpp-python ETCDIR=/usr/local/etc/llama-cpp-python --End SUB_LIST-- ---Begin make.conf--- # XXX: We really need this but cannot use it while 'make checksum' does not # try the next mirror on checksum failure. It currently retries the same # failed mirror and then fails rather then trying another. It *does* # try the next if the size is mismatched though. #MASTER_SITE_FREEBSD=yes # Build ALLOW_MAKE_JOBS_PACKAGES with 3 jobs MAKE_JOBS_NUMBER=3 USE_PACKAGE_DEPENDS=yes BATCH=yes WRKDIRPREFIX=/wrkdirs PORTSDIR=/usr/ports PACKAGES=/packages DISTDIR=/distfiles PACKAGE_BUILDING=yes PACKAGE_BUILDING_FLAVORS=yes #### #### # XXX: We really need this but cannot use it while 'make checksum' does not # try the next mirror on checksum failure. It currently retries the same # failed mirror and then fails rather then trying another. It *does* # try the next if the size is mismatched though. #MASTER_SITE_FREEBSD=yes # Build ALLOW_MAKE_JOBS_PACKAGES with 3 jobs MAKE_JOBS_NUMBER=3 #### Misc Poudriere #### .include "/etc/make.conf.ports_env" GID=0 UID=0 ---End make.conf--- --Resource limits-- cpu time (seconds, -t) unlimited file size (512-blocks, -f) unlimited data seg size (kbytes, -d) 524288 stack size (kbytes, -s) 65536 core file size (512-blocks, -c) unlimited max memory size (kbytes, -m) unlimited locked memory (kbytes, -l) unlimited max user processes (-u) 89999 open files (-n) 8192 virtual mem size (kbytes, -v) unlimited swap limit (kbytes, -w) unlimited socket buffer size (bytes, -b) unlimited pseudo-terminals (-p) unlimited kqueues (-k) unlimited umtx shared locks (-o) unlimited --End resource limits-- =======================<phase: check-sanity >============================ ===== env: NO_DEPENDS=yes USER=root UID=0 GID=0 ===> License MIT accepted by the user =========================================================================== =======================<phase: pkg-depends >============================ ===== env: USE_PACKAGE_DEPENDS_ONLY=1 USER=root UID=0 GID=0 ===> py311-llama-cpp-python-0.3.2 depends on file: /usr/local/sbin/pkg - not found ===> Installing existing package /packages/All/pkg-1.21.3.pkg [133releng-armv7-default-job-13] Installing pkg-1.21.3... [133releng-armv7-default-job-13] Extracting pkg-1.21.3: .......... done ===> py311-llama-cpp-python-0.3.2 depends on file: /usr/local/sbin/pkg - found ===> Returning to build of py311-llama-cpp-python-0.3.2 =========================================================================== =======================<phase: fetch-depends >============================ ===== env: USE_PACKAGE_DEPENDS_ONLY=1 USER=root UID=0 GID=0 =========================================================================== =======================<phase: fetch >============================ ===== env: NO_DEPENDS=yes USER=root UID=0 GID=0 ===> License MIT accepted by the user ===> Fetching all distfiles required by py311-llama-cpp-python-0.3.2 for building =========================================================================== =======================<phase: checksum >============================ ===== env: NO_DEPENDS=yes USER=root UID=0 GID=0 ===> License MIT accepted by the user ===> Fetching all distfiles required by py311-llama-cpp-python-0.3.2 for building => SHA256 Checksum OK for abetlen-llama-cpp-python-v0.3.2_GH0.tar.gz. => SHA256 Checksum OK for ggerganov-llama.cpp-74d73dc_GH0.tar.gz. =========================================================================== =======================<phase: extract-depends>============================ ===== env: USE_PACKAGE_DEPENDS_ONLY=1 USER=root UID=0 GID=0 =========================================================================== =======================<phase: extract >============================ ===== env: NO_DEPENDS=yes USER=root UID=0 GID=0 ===> License MIT accepted by the user ===> Fetching all distfiles required by py311-llama-cpp-python-0.3.2 for building ===> Extracting for py311-llama-cpp-python-0.3.2 => SHA256 Checksum OK for abetlen-llama-cpp-python-v0.3.2_GH0.tar.gz. => SHA256 Checksum OK for ggerganov-llama.cpp-74d73dc_GH0.tar.gz. =========================================================================== =======================<phase: patch-depends >============================ ===== env: USE_PACKAGE_DEPENDS_ONLY=1 USER=root UID=0 GID=0 =========================================================================== =======================<phase: patch >============================ ===== env: NO_DEPENDS=yes USER=root UID=0 GID=0 ===> Patching for py311-llama-cpp-python-0.3.2 =========================================================================== =======================<phase: build-depends >============================ ===== env: USE_PACKAGE_DEPENDS_ONLY=1 USER=root UID=0 GID=0 ===> py311-llama-cpp-python-0.3.2 depends on package: py311-scikit-build-core>0 - not found ===> Installing existing package /packages/All/py311-scikit-build-core-0.10.7.pkg [133releng-armv7-default-job-13] Installing py311-scikit-build-core-0.10.7... [133releng-armv7-default-job-13] `-- Installing py311-packaging-24.2... [133releng-armv7-default-job-13] | `-- Installing python311-3.11.10... [133releng-armv7-default-job-13] | | `-- Installing gettext-runtime-0.22.5... [133releng-armv7-default-job-13] | | `-- Installing indexinfo-0.3.1... [133releng-armv7-default-job-13] | | `-- Extracting indexinfo-0.3.1: .... done [133releng-armv7-default-job-13] | | `-- Extracting gettext-runtime-0.22.5: .......... done [133releng-armv7-default-job-13] | | `-- Installing libffi-3.4.6... [133releng-armv7-default-job-13] | | `-- Extracting libffi-3.4.6: .......... done [133releng-armv7-default-job-13] | | `-- Installing mpdecimal-4.0.0... [133releng-armv7-default-job-13] | | `-- Extracting mpdecimal-4.0.0: .......... done [133releng-armv7-default-job-13] | | `-- Installing readline-8.2.13_2... [133releng-armv7-default-job-13] | | `-- Extracting readline-8.2.13_2: .......... done [133releng-armv7-default-job-13] | `-- Extracting python311-3.11.10: .......... done [133releng-armv7-default-job-13] `-- Extracting py311-packaging-24.2: .......... done [133releng-armv7-default-job-13] `-- Installing py311-pathspec-0.12.1... <snip> -- ARM detected -- Performing Test COMPILER_SUPPORTS_FP16_FORMAT_I3E -- Performing Test COMPILER_SUPPORTS_FP16_FORMAT_I3E - Failed -- Using runtime weight conversion of Q4_0 to Q4_0_x_x to enable optimized GEMM/GEMV kernels -- Including CPU backend CMake Warning at vendor/llama.cpp/ggml/src/ggml-amx/CMakeLists.txt:106 (message): AMX requires x86 and gcc version > 11.0. Turning off GGML_AMX. -- Looking for pthread_create in pthreads -- Looking for pthread_create in pthreads - not found -- Looking for pthread_create in pthread -- Looking for pthread_create in pthread - found CMake Warning at vendor/llama.cpp/common/CMakeLists.txt:30 (message): Git repository not found; to enable automatic generation of build info, make sure Git is installed and the project is a Git repository. CMake Warning (dev) at CMakeLists.txt:13 (install): Target llama has PUBLIC_HEADER files but no PUBLIC_HEADER DESTINATION. Call Stack (most recent call first): CMakeLists.txt:80 (llama_cpp_python_install_target) This warning is for project developers. Use -Wno-dev to suppress it. CMake Warning (dev) at CMakeLists.txt:21 (install): Target llama has PUBLIC_HEADER files but no PUBLIC_HEADER DESTINATION. Call Stack (most recent call first): CMakeLists.txt:80 (llama_cpp_python_install_target) This warning is for project developers. Use -Wno-dev to suppress it. CMake Warning (dev) at CMakeLists.txt:13 (install): Target ggml has PUBLIC_HEADER files but no PUBLIC_HEADER DESTINATION. Call Stack (most recent call first): CMakeLists.txt:81 (llama_cpp_python_install_target) This warning is for project developers. Use -Wno-dev to suppress it. CMake Warning (dev) at CMakeLists.txt:21 (install): Target ggml has PUBLIC_HEADER files but no PUBLIC_HEADER DESTINATION. Call Stack (most recent call first): CMakeLists.txt:81 (llama_cpp_python_install_target) This warning is for project developers. Use -Wno-dev to suppress it. -- Configuring done (4.2s) -- Generating done (0.1s) -- Build files have been written to: /tmp/tmpaz5rnqxo/build *** Building project with Unix Makefiles... Change Dir: '/tmp/tmpaz5rnqxo/build' Run Build Command(s): /usr/local/bin/cmake -E env VERBOSE=1 /usr/bin/make -f Makefile /usr/local/bin/cmake -S/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.2 -B/tmp/tmpaz5rnqxo/build --check-build-system CMakeFiles/Makefile.cmake 0 /usr/local/bin/cmake -E cmake_progress_start /tmp/tmpaz5rnqxo/build/CMakeFiles /tmp/tmpaz5rnqxo/build//CMakeFiles/progress.marks /usr/bin/make -f CMakeFiles/Makefile2 all /usr/bin/make -f vendor/llama.cpp/ggml/src/CMakeFiles/ggml-base.dir/build.make vendor/llama.cpp/ggml/src/CMakeFiles/ggml-base.dir/depend cd /tmp/tmpaz5rnqxo/build && /usr/local/bin/cmake -E cmake_depends "Unix Makefiles" /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.2 /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.2/vendor/llama.cpp/ggml/src /tmp/tmpaz5rnqxo/build /tmp/tmpaz5rnqxo/build/vendor/llama.cpp/ggml/src /tmp/tmpaz5rnqxo/build/vendor/llama.cpp/ggml/src/CMakeFiles/ggml-base.dir/DependInfo.cmake "--color=" /usr/bin/make -f vendor/llama.cpp/ggml/src/CMakeFiles/ggml-base.dir/build.make vendor/llama.cpp/ggml/src/CMakeFiles/ggml-base.dir/build [ 2%] Building C object vendor/llama.cpp/ggml/src/CMakeFiles/ggml-base.dir/ggml.c.o cd /tmp/tmpaz5rnqxo/build/vendor/llama.cpp/ggml/src && /usr/bin/cc -DGGML_BUILD -DGGML_SCHED_MAX_COPIES=4 -DGGML_SHARED -D_XOPEN_SOURCE=600 -D__BSD_VISIBLE -Dggml_base_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.2/vendor/llama.cpp/ggml/src/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.2/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -std=gnu11 -fPIC -Wshadow -Wstrict-prototypes -Wpointer-arith -Wmissing-prototypes -Werror=implicit-int -Werror=implicit-function-declaration -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wdouble-promotion -pthread -MD -MT vendor/llama.cpp/ggml/src/CMakeFiles/ggml-base.dir/ggml.c.o -MF CMakeFiles/ggml-base.dir/ggml.c.o.d -o CMakeFiles/ggml-base.dir/ggml.c.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.2/vendor/llama. cpp/ggml/src/ggml.c [ 5%] Building C object vendor/llama.cpp/ggml/src/CMakeFiles/ggml-base.dir/ggml-alloc.c.o cd /tmp/tmpaz5rnqxo/build/vendor/llama.cpp/ggml/src && /usr/bin/cc -DGGML_BUILD -DGGML_SCHED_MAX_COPIES=4 -DGGML_SHARED -D_XOPEN_SOURCE=600 -D__BSD_VISIBLE -Dggml_base_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.2/vendor/llama.cpp/ggml/src/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.2/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -std=gnu11 -fPIC -Wshadow -Wstrict-prototypes -Wpointer-arith -Wmissing-prototypes -Werror=implicit-int -Werror=implicit-function-declaration -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wdouble-promotion -pthread -MD -MT vendor/llama.cpp/ggml/src/CMakeFiles/ggml-base.dir/ggml-alloc.c.o -MF CMakeFiles/ggml-base.dir/ggml-alloc.c.o.d -o CMakeFiles/ggml-base.dir/ggml-alloc.c.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0 .3.2/vendor/llama.cpp/ggml/src/ggml-alloc.c [ 7%] Building CXX object vendor/llama.cpp/ggml/src/CMakeFiles/ggml-base.dir/ggml-backend.cpp.o cd /tmp/tmpaz5rnqxo/build/vendor/llama.cpp/ggml/src && /usr/bin/c++ -DGGML_BUILD -DGGML_SCHED_MAX_COPIES=4 -DGGML_SHARED -D_XOPEN_SOURCE=600 -D__BSD_VISIBLE -Dggml_base_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.2/vendor/llama.cpp/ggml/src/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.2/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -std=gnu++11 -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -pthread -MD -MT vendor/llama.cpp/ggml/src/CMakeFiles/ggml-base.dir/ggml-backend.cpp.o -MF CMakeFiles/ggml-base.dir/ggml-backend.cpp.o.d -o CMakeFiles/ggml-base.dir/ggml-backend.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.2/vendor/llama.cpp/ggml/src/ggml-backend.cpp [ 10%] Building CXX object vendor/llama.cpp/ggml/src/CMakeFiles/ggml-base.dir/ggml-threading.cpp.o cd /tmp/tmpaz5rnqxo/build/vendor/llama.cpp/ggml/src && /usr/bin/c++ -DGGML_BUILD -DGGML_SCHED_MAX_COPIES=4 -DGGML_SHARED -D_XOPEN_SOURCE=600 -D__BSD_VISIBLE -Dggml_base_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.2/vendor/llama.cpp/ggml/src/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.2/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -std=gnu++11 -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -pthread -MD -MT vendor/llama.cpp/ggml/src/CMakeFiles/ggml-base.dir/ggml-threading.cpp.o -MF CMakeFiles/ggml-base.dir/ggml-threading.cpp.o.d -o CMakeFiles/ggml-base.dir/ggml-threading.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.2/vendor/llama.cpp/ggml/src/ggml-threading.cpp [ 12%] Building C object vendor/llama.cpp/ggml/src/CMakeFiles/ggml-base.dir/ggml-quants.c.o cd /tmp/tmpaz5rnqxo/build/vendor/llama.cpp/ggml/src && /usr/bin/cc -DGGML_BUILD -DGGML_SCHED_MAX_COPIES=4 -DGGML_SHARED -D_XOPEN_SOURCE=600 -D__BSD_VISIBLE -Dggml_base_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.2/vendor/llama.cpp/ggml/src/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.2/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -std=gnu11 -fPIC -Wshadow -Wstrict-prototypes -Wpointer-arith -Wmissing-prototypes -Werror=implicit-int -Werror=implicit-function-declaration -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wdouble-promotion -pthread -MD -MT vendor/llama.cpp/ggml/src/CMakeFiles/ggml-base.dir/ggml-quants.c.o -MF CMakeFiles/ggml-base.dir/ggml-quants.c.o.d -o CMakeFiles/ggml-base.dir/ggml-quants.c.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-pytho n-0.3.2/vendor/llama.cpp/ggml/src/ggml-quants.c [ 15%] Building C object vendor/llama.cpp/ggml/src/CMakeFiles/ggml-base.dir/ggml-aarch64.c.o cd /tmp/tmpaz5rnqxo/build/vendor/llama.cpp/ggml/src && /usr/bin/cc -DGGML_BUILD -DGGML_SCHED_MAX_COPIES=4 -DGGML_SHARED -D_XOPEN_SOURCE=600 -D__BSD_VISIBLE -Dggml_base_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.2/vendor/llama.cpp/ggml/src/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.2/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -std=gnu11 -fPIC -Wshadow -Wstrict-prototypes -Wpointer-arith -Wmissing-prototypes -Werror=implicit-int -Werror=implicit-function-declaration -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wdouble-promotion -pthread -MD -MT vendor/llama.cpp/ggml/src/CMakeFiles/ggml-base.dir/ggml-aarch64.c.o -MF CMakeFiles/ggml-base.dir/ggml-aarch64.c.o.d -o CMakeFiles/ggml-base.dir/ggml-aarch64.c.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-py thon-0.3.2/vendor/llama.cpp/ggml/src/ggml-aarch64.c [ 17%] Linking CXX shared library libggml-base.so cd /tmp/tmpaz5rnqxo/build/vendor/llama.cpp/ggml/src && /usr/local/bin/cmake -E cmake_link_script CMakeFiles/ggml-base.dir/link.txt --verbose=1 /usr/bin/c++ -fPIC -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -Xlinker --dependency-file -Xlinker CMakeFiles/ggml-base.dir/link.d -fstack-protector-strong -shared -Wl,-soname,libggml-base.so -o libggml-base.so "CMakeFiles/ggml-base.dir/ggml.c.o" "CMakeFiles/ggml-base.dir/ggml-alloc.c.o" "CMakeFiles/ggml-base.dir/ggml-backend.cpp.o" "CMakeFiles/ggml-base.dir/ggml-threading.cpp.o" "CMakeFiles/ggml-base.dir/ggml-quants.c.o" "CMakeFiles/ggml-base.dir/ggml-aarch64.c.o" -Wl,-rpath,"\$ORIGIN" -lm -pthread [ 17%] Built target ggml-base /usr/bin/make -f vendor/llama.cpp/ggml/src/ggml-cpu/CMakeFiles/ggml-cpu.dir/build.make vendor/llama.cpp/ggml/src/ggml-cpu/CMakeFiles/ggml-cpu.dir/depend cd /tmp/tmpaz5rnqxo/build && /usr/local/bin/cmake -E cmake_depends "Unix Makefiles" /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.2 /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.2/vendor/llama.cpp/ggml/src/ggml-cpu /tmp/tmpaz5rnqxo/build /tmp/tmpaz5rnqxo/build/vendor/llama.cpp/ggml/src/ggml-cpu /tmp/tmpaz5rnqxo/build/vendor/llama.cpp/ggml/src/ggml-cpu/CMakeFiles/ggml-cpu.dir/DependInfo.cmake "--color=" /usr/bin/make -f vendor/llama.cpp/ggml/src/ggml-cpu/CMakeFiles/ggml-cpu.dir/build.make vendor/llama.cpp/ggml/src/ggml-cpu/CMakeFiles/ggml-cpu.dir/build [ 20%] Building C object vendor/llama.cpp/ggml/src/ggml-cpu/CMakeFiles/ggml-cpu.dir/ggml-cpu.c.o cd /tmp/tmpaz5rnqxo/build/vendor/llama.cpp/ggml/src/ggml-cpu && /usr/bin/cc -DGGML_BACKEND_BUILD -DGGML_BACKEND_SHARED -DGGML_SCHED_MAX_COPIES=4 -DGGML_SHARED -DGGML_USE_CPU_AARCH64 -DGGML_USE_LLAMAFILE -D_XOPEN_SOURCE=600 -D__BSD_VISIBLE -Dggml_cpu_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.2/vendor/llama.cpp/ggml/src/ggml-cpu/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.2/vendor/llama.cpp/ggml/src/ggml-cpu/.. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.2/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -std=gnu11 -fPIC -Wshadow -Wstrict-prototypes -Wpointer-arith -Wmissing-prototypes -Werror=implicit-int -Werror=implicit-function-declaration -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wdouble-promotion -mfpu=neon-fp-armv8 -mno-unaligned-acce ss -funsafe-math-optimizations -MD -MT vendor/llama.cpp/ggml/src/ggml-cpu/CMakeFiles/ggml-cpu.dir/ggml-cpu.c.o -MF CMakeFiles/ggml-cpu.dir/ggml-cpu.c.o.d -o CMakeFiles/ggml-cpu.dir/ggml-cpu.c.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.2/vendor/llama.cpp/ggml/src/ggml-cpu/ggml-cpu.c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.2/vendor/llama.cpp/ggml/src/ggml-cpu/ggml-cpu.c:245:9: warning: 'CACHE_LINE_SIZE' macro redefined [-Wmacro-redefined] 245 | #define CACHE_LINE_SIZE 64 | ^ /usr/include/machine/param.h:102:9: note: previous definition is here 102 | #define CACHE_LINE_SIZE (1 << CACHE_LINE_SHIFT) | ^ /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.2/vendor/llama.cpp/ggml/src/ggml-cpu/ggml-cpu.c:1534:5: warning: implicit conversion increases floating-point precision: 'float' to 'ggml_float' (aka 'double') [-Wdouble-promotion] 1534 | GGML_F16_VEC_REDUCE(sumf, sum); | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.2/vendor/llama.cpp/ggml/src/ggml-cpu/ggml-cpu.c:582:41: note: expanded from macro 'GGML_F16_VEC_REDUCE' 582 | #define GGML_F16_VEC_REDUCE GGML_F32Cx4_REDUCE | ^ /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.2/vendor/llama.cpp/ggml/src/ggml-cpu/ggml-cpu.c:572:38: note: expanded from macro 'GGML_F32Cx4_REDUCE' 572 | #define GGML_F32Cx4_REDUCE GGML_F32x4_REDUCE | ^ /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.2/vendor/llama.cpp/ggml/src/ggml-cpu/ggml-cpu.c:502:13: note: expanded from macro 'GGML_F32x4_REDUCE' 502 | (res) = GGML_F32x4_REDUCE_ONE((x)[0]); \ | ~ ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.2/vendor/llama.cpp/ggml/src/ggml-cpu/ggml-cpu.c:487:34: note: expanded from macro 'GGML_F32x4_REDUCE_ONE' 487 | #define GGML_F32x4_REDUCE_ONE(x) vaddvq_f32(x) | ^~~~~~~~~~~~~ /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.2/vendor/llama.cpp/ggml/src/ggml-cpu/ggml-cpu.c:1582:9: warning: implicit conversion increases floating-point precision: 'float' to 'ggml_float' (aka 'double') [-Wdouble-promotion] 1582 | GGML_F16_VEC_REDUCE(sumf[k], sum[k]); | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.2/vendor/llama.cpp/ggml/src/ggml-cpu/ggml-cpu.c:582:41: note: expanded from macro 'GGML_F16_VEC_REDUCE' 582 | #define GGML_F16_VEC_REDUCE GGML_F32Cx4_REDUCE | ^ /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.2/vendor/llama.cpp/ggml/src/ggml-cpu/ggml-cpu.c:572:38: note: expanded from macro 'GGML_F32Cx4_REDUCE' 572 | #define GGML_F32Cx4_REDUCE GGML_F32x4_REDUCE | ^ /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.2/vendor/llama.cpp/ggml/src/ggml-cpu/ggml-cpu.c:502:13: note: expanded from macro 'GGML_F32x4_REDUCE' 502 | (res) = GGML_F32x4_REDUCE_ONE((x)[0]); \ | ~ ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.2/vendor/llama.cpp/ggml/src/ggml-cpu/ggml-cpu.c:487:34: note: expanded from macro 'GGML_F32x4_REDUCE_ONE' 487 | #define GGML_F32x4_REDUCE_ONE(x) vaddvq_f32(x) | ^~~~~~~~~~~~~ fatal error: error in backend: Cannot select: 0x270f5aa0: v4f32 = fmaxnum 0x27a93190, 0x269f01e0 0x27a93190: v4f32,i32,ch = ARMISD::VLD1_UPD<(load (s128) from %ir.16869, align 4)> 0x269ad6fc, 0x269f0640, Constant:i32<16>, Constant:i32<1> 0x269f0640: i32,ch = CopyFromReg 0x269ad6fc, Register:i32 %6764 0x270f5be0: i32 = Register %6764 0x27a93b40: i32 = Constant<16> 0x26974f50: i32 = Constant<1> 0x269f01e0: v4f32 = bitcast 0x270f55f0 0x270f55f0: v4i32 = ARMISD::VMOVIMM TargetConstant:i32<0> 0x27a933c0: i32 = TargetConstant<0> In function: ggml_graph_compute_thread PLEASE submit a bug report to https://bugs.freebsd.org/submit/ and include the crash backtrace, preprocessed source, and associated run script. Stack dump: 0. Program arguments: /usr/bin/cc -DGGML_BACKEND_BUILD -DGGML_BACKEND_SHARED -DGGML_SCHED_MAX_COPIES=4 -DGGML_SHARED -DGGML_USE_CPU_AARCH64 -DGGML_USE_LLAMAFILE -D_XOPEN_SOURCE=600 -D__BSD_VISIBLE -Dggml_cpu_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.2/vendor/llama.cpp/ggml/src/ggml-cpu/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.2/vendor/llama.cpp/ggml/src/ggml-cpu/.. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.2/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -std=gnu11 -fPIC -Wshadow -Wstrict-prototypes -Wpointer-arith -Wmissing-prototypes -Werror=implicit-int -Werror=implicit-function-declaration -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wdouble-promotion -mfpu=neon-fp-armv8 -mno-unaligned-access -funsafe-math-optimizations -MD -MT vend or/llama.cpp/ggml/src/ggml-cpu/CMakeFiles/ggml-cpu.dir/ggml-cpu.c.o -MF CMakeFiles/ggml-cpu.dir/ggml-cpu.c.o.d -o CMakeFiles/ggml-cpu.dir/ggml-cpu.c.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.2/vendor/llama.cpp/ggml/src/ggml-cpu/ggml-cpu.c 1. <eof> parser at end of file 2. Code generation 3. Running pass 'Function Pass Manager' on module '/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.2/vendor/llama.cpp/ggml/src/ggml-cpu/ggml-cpu.c'. 4. Running pass 'ARM Instruction Selection' on function '@ggml_graph_compute_thread' #0 0x046f3e28 (/usr/bin/cc+0x46f3e28) #1 0x046f1cf0 (/usr/bin/cc+0x46f1cf0) #2 0x046b81a4 (/usr/bin/cc+0x46b81a4) #3 0x046b8160 (/usr/bin/cc+0x46b8160) #4 0x046e2ab0 (/usr/bin/cc+0x46e2ab0) #5 0x019cc89c (/usr/bin/cc+0x19cc89c) #6 0x046bf248 (/usr/bin/cc+0x46bf248) #7 0x04a2df4c (/usr/bin/cc+0x4a2df4c) #8 0x04a2d4d0 (/usr/bin/cc+0x4a2d4d0) #9 0x04ee11cc (/usr/bin/cc+0x4ee11cc) #10 0x04a25330 (/usr/bin/cc+0x4a25330) #11 0x04a24b84 (/usr/bin/cc+0x4a24b84) #12 0x04a247d8 (/usr/bin/cc+0x4a247d8) #13 0x04a24048 (/usr/bin/cc+0x4a24048) #14 0x04a223fc (/usr/bin/cc+0x4a223fc) #15 0x04edc618 (/usr/bin/cc+0x4edc618) #16 0x03eab248 (/usr/bin/cc+0x3eab248) #17 0x0431ae08 (/usr/bin/cc+0x431ae08) #18 0x04320e04 (/usr/bin/cc+0x4320e04) #19 0x0431b3e0 (/usr/bin/cc+0x431b3e0) #20 0x022968bc (/usr/bin/cc+0x22968bc) #21 0x025a7d6c (/usr/bin/cc+0x25a7d6c) #22 0x02aefc2c (/usr/bin/cc+0x2aefc2c) #23 0x024e61e4 (/usr/bin/cc+0x24e61e4) #24 0x0246ade4 (/usr/bin/cc+0x246ade4) #25 0x025a13e8 (/usr/bin/cc+0x25a13e8) #26 0x019cc240 (/usr/bin/cc+0x19cc240) #27 0x019d9774 (/usr/bin/cc+0x19d9774) #28 0x02314fc4 (/usr/bin/cc+0x2314fc4) #29 0x046b813c (/usr/bin/cc+0x46b813c) #30 0x02314780 (/usr/bin/cc+0x2314780) #31 0x022df1a4 (/usr/bin/cc+0x22df1a4) #32 0x022df430 (/usr/bin/cc+0x22df430) #33 0x022f7640 (/usr/bin/cc+0x22f7640) #34 0x019d9020 (/usr/bin/cc+0x19d9020) cc: error: clang frontend command failed with exit code 70 (use -v to see invocation) FreeBSD clang version 17.0.6 (https://github.com/llvm/llvm-project.git llvmorg-17.0.6-0-g6009708b4367) Target: armv7-unknown-freebsd13.3-gnueabihf Thread model: posix InstalledDir: /usr/bin cc: note: diagnostic msg: ******************** PLEASE ATTACH THE FOLLOWING FILES TO THE BUG REPORT: Preprocessed source(s) and associated run script(s) are located at: cc: note: diagnostic msg: /tmp/ggml-cpu-6ef35f.c cc: note: diagnostic msg: /tmp/ggml-cpu-6ef35f.sh cc: note: diagnostic msg: ******************** *** Error code 1 Stop. make[2]: stopped in /tmp/tmpaz5rnqxo/build *** Error code 1 Stop. make[1]: stopped in /tmp/tmpaz5rnqxo/build *** Error code 1 Stop. make: stopped in /tmp/tmpaz5rnqxo/build *** CMake build failed ERROR Backend subprocess exited when trying to invoke build_wheel *** Error code 1 Stop. make: stopped in /usr/ports/misc/py-llama-cpp-python