[package - 140i386-default][misc/llama-cpp] Failed for llama-cpp-3658 in build
- Go to: [ bottom of page ] [ top of archives ] [ this month ]
Date: Thu, 05 Sep 2024 03:23:00 UTC
You are receiving this mail as a port that you maintain is failing to build on the FreeBSD package build server. Please investigate the failure and submit a PR to fix build. Maintainer: yuri@FreeBSD.org Log URL: https://pkg-status.freebsd.org/beefy21/data/140i386-default/b9994ae18767/logs/llama-cpp-3658.log Build URL: https://pkg-status.freebsd.org/beefy21/build.html?mastername=140i386-default&build=b9994ae18767 Log: =>> Building misc/llama-cpp build started at Thu Sep 5 03:22:07 UTC 2024 port directory: /usr/ports/misc/llama-cpp package name: llama-cpp-3658 building for: FreeBSD 140i386-default-job-29 14.0-RELEASE-p10 FreeBSD 14.0-RELEASE-p10 i386 maintained by: yuri@FreeBSD.org Makefile datestamp: -rw-r--r-- 1 root wheel 1329 Sep 5 01:01 /usr/ports/misc/llama-cpp/Makefile Ports top last git commit: b9994ae187 Ports top unclean checkout: no Port dir last git commit: 2dffd523cf Port dir unclean checkout: no Poudriere version: poudriere-git-3.4.2 Host OSVERSION: 1500023 Jail OSVERSION: 1400097 Job Id: 29 ---Begin Environment--- SHELL=/bin/sh BLOCKSIZE=K MAIL=/var/mail/root MM_CHARSET=UTF-8 LANG=C.UTF-8 OSVERSION=1400097 STATUS=1 HOME=/root PATH=/sbin:/bin:/usr/sbin:/usr/bin:/usr/local/sbin:/usr/local/bin:/root/bin MAKE_OBJDIR_CHECK_WRITABLE=0 UNAME_m=i386 UNAME_p=i386 UNAME_r=14.0-RELEASE-p10 LOCALBASE=/usr/local UNAME_v=FreeBSD 14.0-RELEASE-p10 USER=root POUDRIERE_NAME=poudriere-git LIBEXECPREFIX=/usr/local/libexec/poudriere POUDRIERE_VERSION=3.4.2 MASTERMNT=/usr/local/poudriere/data/.m/140i386-default/ref LC_COLLATE=C POUDRIERE_BUILD_TYPE=bulk PACKAGE_BUILDING=yes SAVED_TERM= OUTPUT_REDIRECTED_STDERR=4 OUTPUT_REDIRECTED=1 PWD=/usr/local/poudriere/data/.m/140i386-default/29/.p OUTPUT_REDIRECTED_STDOUT=3 P_PORTS_FEATURES=FLAVORS SUBPACKAGES SELECTED_OPTIONS MASTERNAME=140i386-default SCRIPTPREFIX=/usr/local/share/poudriere SCRIPTNAME=bulk.sh OLDPWD=/usr/local/poudriere/data/.m/140i386-default/ref/.p/pool POUDRIERE_PKGNAME=poudriere-git-3.4.2 SCRIPTPATH=/usr/local/share/poudriere/bulk.sh POUDRIEREPATH=/usr/local/bin/poudriere ---End Environment--- ---Begin Poudriere Port Flags/Env--- PORT_FLAGS= PKGENV= FLAVOR= MAKE_ARGS= ---End Poudriere Port Flags/Env--- ---Begin OPTIONS List--- ===> The following configuration options are available for llama-cpp-3658: EXAMPLES=on: Build and/or install examples VULKAN=on: Vulkan GPU offload support ===> Use 'make config' to modify these settings ---End OPTIONS List--- --MAINTAINER-- yuri@FreeBSD.org --End MAINTAINER-- --CONFIGURE_ARGS-- --End CONFIGURE_ARGS-- --CONFIGURE_ENV-- PYTHON="/usr/local/bin/python3.11" XDG_DATA_HOME=/wrkdirs/usr/ports/misc/llama-cpp/work XDG_CONFIG_HOME=/wrkdirs/usr/ports/misc/llama-cpp/work XDG_CACHE_HOME=/wrkdirs/usr/ports/misc/llama-cpp/work/.cache HOME=/wrkdirs/usr/ports/misc/llama-cpp/work TMPDIR="/tmp" PATH=/wrkdirs/usr/ports/misc/llama-cpp/work/.bin:/sbin:/bin:/usr/sbin:/usr/bin:/usr/local/sbin:/usr/local/bin:/root/bin PKG_CONFIG_LIBDIR=/wrkdirs/usr/ports/misc/llama-cpp/work/.pkgconfig:/usr/local/libdata/pkgconfig:/usr/local/share/pkgconfig:/usr/libdata/pkgconfig SHELL=/bin/sh CONFIG_SHELL=/bin/sh --End CONFIGURE_ENV-- --MAKE_ENV-- NINJA_STATUS="[%p %s/%t] " XDG_DATA_HOME=/wrkdirs/usr/ports/misc/llama-cpp/work XDG_CONFIG_HOME=/wrkdirs/usr/ports/misc/llama-cpp/work XDG_CACHE_HOME=/wrkdirs/usr/ports/misc/llama-cpp/work/.cache HOME=/wrkdirs/usr/ports/misc/llama-cpp/work TMPDIR="/tmp" PATH=/wrkdirs/usr/ports/misc/llama-cpp/work/.bin:/sbin:/bin:/usr/sbin:/usr/bin:/usr/local/sbin:/usr/local/bin:/root/bin PKG_CONFIG_LIBDIR=/wrkdirs/usr/ports/misc/llama-cpp/work/.pkgconfig:/usr/local/libdata/pkgconfig:/usr/local/share/pkgconfig:/usr/libdata/pkgconfig MK_DEBUG_FILES=no MK_KERNEL_SYMBOLS=no SHELL=/bin/sh NO_LINT=YES DESTDIR=/wrkdirs/usr/ports/misc/llama-cpp/work/stage PREFIX=/usr/local LOCALBASE=/usr/local CC="cc" CFLAGS="-O2 -pipe -fstack-protector-strong -fno-strict-aliasing " CPP="cpp" CPPFLAGS="" LDFLAGS=" -fstack-protector-strong " LIBS="" CXX="c++" CXXFLAGS="-O2 -pipe -fstack-protector-strong -fno-strict-aliasing " BSD_INSTALL_PROGRAM="install -s -m 555" BSD_INSTALL_LIB="install -s -m 0644" B SD_INSTALL_SCRIPT="install -m 555" BSD_INSTALL_DATA="install -m 0644" BSD_INSTALL_MAN="install -m 444" --End MAKE_ENV-- --PLIST_SUB-- PORTEXAMPLES="" EXAMPLES="" NO_EXAMPLES="@comment " VULKAN="" NO_VULKAN="@comment " CMAKE_BUILD_TYPE="release" PYTHON_INCLUDEDIR=include/python3.11 PYTHON_LIBDIR=lib/python3.11 PYTHON_PLATFORM=freebsd14 PYTHON_SITELIBDIR=lib/python3.11/site-packages PYTHON_SUFFIX=311 PYTHON_EXT_SUFFIX=.cpython-311 PYTHON_VER=3.11 PYTHON_VERSION=python3.11 PYTHON2="@comment " PYTHON3="" OSREL=14.0 PREFIX=%D LOCALBASE=/usr/local RESETPREFIX=/usr/local LIB32DIR=lib DOCSDIR="share/doc/llama-cpp" EXAMPLESDIR="share/examples/llama-cpp" DATADIR="share/llama-cpp" WWWDIR="www/llama-cpp" ETCDIR="etc/llama-cpp" --End PLIST_SUB-- --SUB_LIST-- EXAMPLES="" NO_EXAMPLES="@comment " VULKAN="" NO_VULKAN="@comment " PYTHON_INCLUDEDIR=/usr/local/include/python3.11 PYTHON_LIBDIR=/usr/local/lib/python3.11 PYTHON_PLATFORM=freebsd14 PYTHON_SITELIBDIR=/usr/local/lib/python3.11/site-packages PYTHON_SUFFIX=311 PYTHON_EXT_SUFFIX=.cpython-311 PYTHON_VER=3.11 PYTHON_VERSION=python3.11 PYTHON2="@comment " PYTHON3="" PREFIX=/usr/local LOCALBASE=/usr/local DATADIR=/usr/local/share/llama-cpp DOCSDIR=/usr/local/share/doc/llama-cpp EXAMPLESDIR=/usr/local/share/examples/llama-cpp WWWDIR=/usr/local/www/llama-cpp ETCDIR=/usr/local/etc/llama-cpp --End SUB_LIST-- ---Begin make.conf--- USE_PACKAGE_DEPENDS=yes BATCH=yes WRKDIRPREFIX=/wrkdirs PORTSDIR=/usr/ports PACKAGES=/packages DISTDIR=/distfiles PACKAGE_BUILDING=yes PACKAGE_BUILDING_FLAVORS=yes MACHINE=i386 MACHINE_ARCH=i386 ARCH=${MACHINE_ARCH} #### #### # XXX: We really need this but cannot use it while 'make checksum' does not # try the next mirror on checksum failure. It currently retries the same # failed mirror and then fails rather then trying another. It *does* # try the next if the size is mismatched though. #MASTER_SITE_FREEBSD=yes # Build ALLOW_MAKE_JOBS_PACKAGES with 3 jobs MAKE_JOBS_NUMBER=3 #### Misc Poudriere #### .include "/etc/make.conf.ports_env" GID=0 UID=0 DISABLE_MAKE_JOBS=poudriere ---End make.conf--- --Resource limits-- cpu time (seconds, -t) unlimited file size (512-blocks, -f) unlimited data seg size (kbytes, -d) 524288 stack size (kbytes, -s) 65536 core file size (512-blocks, -c) unlimited max memory size (kbytes, -m) unlimited locked memory (kbytes, -l) unlimited max user processes (-u) 89999 open files (-n) 8192 virtual mem size (kbytes, -v) unlimited swap limit (kbytes, -w) unlimited socket buffer size (bytes, -b) unlimited pseudo-terminals (-p) unlimited kqueues (-k) unlimited umtx shared locks (-o) unlimited --End resource limits-- =======================<phase: check-sanity >============================ ===== env: NO_DEPENDS=yes USER=root UID=0 GID=0 ===> License MIT accepted by the user =========================================================================== =======================<phase: pkg-depends >============================ ===== env: USE_PACKAGE_DEPENDS_ONLY=1 USER=root UID=0 GID=0 ===> llama-cpp-3658 depends on file: /usr/local/sbin/pkg - not found ===> Installing existing package /packages/All/pkg-1.21.3.pkg [140i386-default-job-29] Installing pkg-1.21.3... [140i386-default-job-29] Extracting pkg-1.21.3: .......... done ===> llama-cpp-3658 depends on file: /usr/local/sbin/pkg - found ===> Returning to build of llama-cpp-3658 =========================================================================== =======================<phase: fetch-depends >============================ ===== env: USE_PACKAGE_DEPENDS_ONLY=1 USER=root UID=0 GID=0 =========================================================================== =======================<phase: fetch >============================ ===== env: NO_DEPENDS=yes USER=root UID=0 GID=0 ===> License MIT accepted by the user => ggerganov-llama.cpp-b3658_GH0.tar.gz doesn't seem to exist in /portdistfiles/. => Attempting to fetch https://codeload.github.com/ggerganov/llama.cpp/tar.gz/b3658?dummy=/ggerganov-llama.cpp-b3658_GH0.tar.gz fetch: https://codeload.github.com/ggerganov/llama.cpp/tar.gz/b3658?dummy=/ggerganov-llama.cpp-b3658_GH0.tar.gz: size unknown fetch: https://codeload.github.com/ggerganov/llama.cpp/tar.gz/b3658?dummy=/ggerganov-llama.cpp-b3658_GH0.tar.gz: size of remote file is not known ggerganov-llama.cpp-b3658_GH0.tar.gz 18 MB 8031 kBps 02s ===> Fetching all distfiles required by llama-cpp-3658 for building =========================================================================== =======================<phase: checksum >============================ ===== env: NO_DEPENDS=yes USER=root UID=0 GID=0 ===> License MIT accepted by the user ===> Fetching all distfiles required by llama-cpp-3658 for building => SHA256 Checksum OK for ggerganov-llama.cpp-b3658_GH0.tar.gz. => SHA256 Checksum OK for nomic-ai-kompute-4565194_GH0.tar.gz. => SHA256 Checksum OK for 121f915a09c1117d34aff6e8faf6d252aaf11027.patch. =========================================================================== =======================<phase: extract-depends>============================ ===== env: USE_PACKAGE_DEPENDS_ONLY=1 USER=root UID=0 GID=0 =========================================================================== =======================<phase: extract >============================ ===== env: NO_DEPENDS=yes USER=root UID=0 GID=0 ===> License MIT accepted by the user ===> Fetching all distfiles required by llama-cpp-3658 for building ===> Extracting for llama-cpp-3658 => SHA256 Checksum OK for ggerganov-llama.cpp-b3658_GH0.tar.gz. => SHA256 Checksum OK for nomic-ai-kompute-4565194_GH0.tar.gz. => SHA256 Checksum OK for 121f915a09c1117d34aff6e8faf6d252aaf11027.patch. =========================================================================== =======================<phase: patch-depends >============================ ===== env: USE_PACKAGE_DEPENDS_ONLY=1 USER=root UID=0 GID=0 =========================================================================== =======================<phase: patch >============================ ===== env: NO_DEPENDS=yes USER=root UID=0 GID=0 ===> Patching for llama-cpp-3658 ===> Applying distribution patches for llama-cpp-3658 =========================================================================== =======================<phase: build-depends >============================ ===== env: USE_PACKAGE_DEPENDS_ONLY=1 USER=root UID=0 GID=0 ===> llama-cpp-3658 depends on executable: glslc - not found ===> Installing existing package /packages/All/shaderc-2024.1.pkg [140i386-default-job-29] Installing shaderc-2024.1... [140i386-default-job-29] Extracting shaderc-2024.1: .......... done ===> llama-cpp-3658 depends on executable: glslc - found ===> Returning to build of llama-cpp-3658 ===> llama-cpp-3658 depends on package: vulkan-headers>0 - not found ===> Installing existing package /packages/All/vulkan-headers-1.3.295.pkg [140i386-default-job-29] Installing vulkan-headers-1.3.295... [140i386-default-job-29] Extracting vulkan-headers-1.3.295: .......... done ===> llama-cpp-3658 depends on package: vulkan-headers>0 - found ===> Returning to build of llama-cpp-3658 ===> llama-cpp-3658 depends on file: /usr/local/bin/cmake - not found ===> Installing existing package /packages/All/cmake-core-3.30.3.pkg [140i386-default-job-29] Installing cmake-core-3.30.3... [140i386-default-job-29] `-- Installing expat-2.6.2... [140i386-default-job-29] `-- Extracting expat-2.6.2: .......... done [140i386-default-job-29] `-- Installing jsoncpp-1.9.5... [140i386-default-job-29] `-- Extracting jsoncpp-1.9.5: .......... done [140i386-default-job-29] `-- Installing libuv-1.48.0... [140i386-default-job-29] `-- Extracting libuv-1.48.0: .......... done [140i386-default-job-29] `-- Installing rhash-1.4.4_1... [140i386-default-job-29] | `-- Installing gettext-runtime-0.22.5... [140i386-default-job-29] | | `-- Installing indexinfo-0.3.1... [140i386-default-job-29] | | `-- Extracting indexinfo-0.3.1: .... done [140i386-default-job-29] | `-- Extracting gettext-runtime-0.22.5: .......... done [140i386-default-job-29] `-- Extracting rhash-1.4.4_1: .......... done [140i386-default-job-29] Extracting cmake-core-3.30.3: .......... done ===> llama-cpp-3658 depends on file: /usr/local/bin/cmake - found ===> Returning to build of llama-cpp-3658 ===> llama-cpp-3658 depends on executable: ninja - not found ===> Installing existing package /packages/All/ninja-1.11.1,4.pkg [140i386-default-job-29] Installing ninja-1.11.1,4... [140i386-default-job-29] `-- Installing python311-3.11.9_1... [140i386-default-job-29] | `-- Installing libffi-3.4.6... [140i386-default-job-29] | `-- Extracting libffi-3.4.6: .......... done [140i386-default-job-29] | `-- Installing mpdecimal-4.0.0... [140i386-default-job-29] | `-- Extracting mpdecimal-4.0.0: .......... done [140i386-default-job-29] | `-- Installing readline-8.2.10... [140i386-default-job-29] | `-- Extracting readline-8.2.10: .......... done [140i386-default-job-29] `-- Extracting python311-3.11.9_1: .......... done [140i386-default-job-29] Extracting ninja-1.11.1,4: ........ done ===== Message from python311-3.11.9_1: -- Note that some standard Python modules are provided as separate ports as they require additional dependencies. They are available as: py311-gdbm databases/py-gdbm@py311 py311-sqlite3 databases/py-sqlite3@py311 py311-tkinter x11-toolkits/py-tkinter@py311 ===> llama-cpp-3658 depends on executable: ninja - found ===> Returning to build of llama-cpp-3658 =========================================================================== =======================<phase: lib-depends >============================ ===== env: USE_PACKAGE_DEPENDS_ONLY=1 USER=root UID=0 GID=0 ===> llama-cpp-3658 depends on shared library: libvulkan.so - not found ===> Installing existing package /packages/All/vulkan-loader-1.3.295.pkg [140i386-default-job-29] Installing vulkan-loader-1.3.295... [140i386-default-job-29] `-- Installing libX11-1.8.9,1... [140i386-default-job-29] | `-- Installing libxcb-1.17.0... [140i386-default-job-29] | | `-- Installing libXau-1.0.11... [140i386-default-job-29] | | `-- Extracting libXau-1.0.11: .......... done [140i386-default-job-29] | | `-- Installing libXdmcp-1.1.5... [140i386-default-job-29] | | `-- Installing xorgproto-2024.1... [140i386-default-job-29] | | `-- Extracting xorgproto-2024.1: .......... done [140i386-default-job-29] | | `-- Extracting libXdmcp-1.1.5: ......... done [140i386-default-job-29] | `-- Extracting libxcb-1.17.0: .......... done [140i386-default-job-29] `-- Extracting libX11-1.8.9,1: .......... done [140i386-default-job-29] `-- Installing libXrandr-1.5.4... [140i386-default-job-29] | `-- Installing libXext-1.3.6,1... [140i386-default-job-29] | `-- Extracting libXext-1.3.6,1: .......... done [140i386-default-job-29] | `-- Installing libXrender-0.9.11... [140i386-default-job-29] | `-- Extracting libXrender-0.9.11: .......... done [140i386-default-job-29] `-- Extracting libXrandr-1.5.4: .......... done [140i386-default-job-29] `-- Installing wayland-1.23.1... [140i386-default-job-29] | `-- Installing libepoll-shim-0.0.20240608... [140i386-default-job-29] | `-- Extracting libepoll-shim-0.0.20240608: .......... done [140i386-default-job-29] | `-- Installing libxml2-2.11.8... [140i386-default-job-29] | `-- Extracting libxml2-2.11.8: .......... done [140i386-default-job-29] `-- Extracting wayland-1.23.1: .......... done [140i386-default-job-29] Extracting vulkan-loader-1.3.295: .......... done ===== Message from wayland-1.23.1: -- Wayland requires XDG_RUNTIME_DIR to be defined to a path that will contain "wayland-%d" unix(4) sockets. This is usually handled by consolekit2 (via ck-launch-session) or pam_xdg (via login). ===> llama-cpp-3658 depends on shared library: libvulkan.so - found (/usr/local/lib/libvulkan.so) ===> Returning to build of llama-cpp-3658 =========================================================================== =======================<phase: configure >============================ ===== env: NO_DEPENDS=yes USER=root UID=0 GID=0 ===> Configuring for llama-cpp-3658 ===> Performing out-of-source build /bin/mkdir -p /wrkdirs/usr/ports/misc/llama-cpp/work/.build -- The C compiler identification is Clang 16.0.6 -- The CXX compiler identification is Clang 16.0.6 -- Detecting C compiler ABI info -- Detecting C compiler ABI info - done -- Check for working C compiler: /usr/bin/cc - skipped -- Detecting C compile features -- Detecting C compile features - done -- Detecting CXX compiler ABI info -- Detecting CXX compiler ABI info - done -- Check for working CXX compiler: /usr/bin/c++ - skipped -- Detecting CXX compile features -- Detecting CXX compile features - done -- Found Git: /wrkdirs/usr/ports/misc/llama-cpp/work/.bin/git -- Performing Test CMAKE_HAVE_LIBC_PTHREAD -- Performing Test CMAKE_HAVE_LIBC_PTHREAD - Failed -- Found Threads: TRUE -- Found OpenMP_C: -fopenmp=libomp (found version "5.0") -- Found OpenMP_CXX: -fopenmp=libomp (found version "5.0") -- Found OpenMP: TRUE (found version "5.0") -- OpenMP found -- Using llamafile -- Found Vulkan: /usr/local/lib/libvulkan.so (found version "1.3.295") found components: glslc missing components: glslangValidator -- Vulkan found -- Warning: ccache not found - consider installing it for faster compilation or disable this warning with GGML_CCACHE=OFF -- CMAKE_SYSTEM_PROCESSOR: i386 -- Unknown architecture -- Looking for pthread_create in pthreads -- Looking for pthread_create in pthreads - not found -- Looking for pthread_create in pthread -- Looking for pthread_create in pthread - found CMake Warning at common/CMakeLists.txt:30 (message): Git repository not found; to enable automatic generation of build info, make sure Git is installed and the project is a Git repository. -- Configuring done (2.2s) -- Generating done (0.1s) CMake Warning: Manually-specified variables were not used by the project: BOOST_PYTHON_SUFFIX CMAKE_COLOR_MAKEFILE CMAKE_MODULE_LINKER_FLAGS CMAKE_VERBOSE_MAKEFILE FETCHCONTENT_FULLY_DISCONNECTED Python3_EXECUTABLE Python_ADDITIONAL_VERSIONS Python_EXECUTABLE -- Build files have been written to: /wrkdirs/usr/ports/misc/llama-cpp/work/.build =========================================================================== =======================<phase: build >============================ ===== env: NO_DEPENDS=yes USER=root UID=0 GID=0 ===> Building for llama-cpp-3658 [ 0% 1/127] /usr/bin/c++ -DGGML_SCHED_MAX_COPIES=4 -DGGML_USE_LLAMAFILE -DGGML_USE_OPENMP -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -DNDEBUG -std=gnu++11 -pthread -MD -MT ggml/src/vulkan-shaders/CMakeFiles/vulkan-shaders-gen.dir/vulkan-shaders-gen.cpp.o -MF ggml/src/vulkan-shaders/CMakeFiles/vulkan-shaders-gen.dir/vulkan-shaders-gen.cpp.o.d -o ggml/src/vulkan-shaders/CMakeFiles/vulkan-shaders-gen.dir/vulkan-shaders-gen.cpp.o -c /wrkdirs/usr/ports/misc/llama-cpp/work/llama.cpp-b3658/ggml/src/vulkan-shaders/vulkan-shaders-gen.cpp [ 1% 2/127] : && /usr/bin/c++ -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -DNDEBUG -fstack-protector-strong ggml/src/vulkan-shaders/CMakeFiles/vulkan-shaders-gen.dir/vulkan-shaders-gen.cpp.o -o bin/vulkan-shaders-gen -pthread && : [ 2% 3/127] cd /wrkdirs/usr/ports/misc/llama-cpp/work/.build/ggml/src && /wrkdirs/usr/ports/misc/llama-cpp/work/.build/bin/vulkan-shaders-gen --glslc /usr/local/bin/glslc --input-dir /wrkdirs/usr/ports/misc/llama-cpp/work/llama.cpp-b3658/ggml/src/vulkan-shaders --output-dir /wrkdirs/usr/ports/misc/llama-cpp/work/.build/ggml/src/vulkan-shaders.spv --target-hpp /wrkdirs/usr/ports/misc/llama-cpp/work/.build/ggml/src/ggml-vulkan-shaders.hpp --target-cpp /wrkdirs/usr/ports/misc/llama-cpp/work/.build/ggml/src/ggml-vulkan-shaders.cpp --no-clean ggml_vulkan: Generating and compiling shaders to SPIR-V [ 3% 4/127] /usr/bin/cc -DGGML_BUILD -DGGML_SCHED_MAX_COPIES=4 -DGGML_SHARED -DGGML_USE_LLAMAFILE -DGGML_USE_OPENMP -DGGML_USE_VULKAN -D_XOPEN_SOURCE=600 -D__BSD_VISIBLE -Dggml_EXPORTS -I/wrkdirs/usr/ports/misc/llama-cpp/work/llama.cpp-b3658/ggml/src/../include -I/wrkdirs/usr/ports/misc/llama-cpp/work/llama.cpp-b3658/ggml/src/. -I/wrkdirs/usr/ports/misc/llama-cpp/work/.build/ggml/src -isystem /usr/local/include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -DNDEBUG -std=gnu11 -fPIC -Wshadow -Wstrict-prototypes -Wpointer-arith -Wmissing-prototypes -Werror=implicit-int -Werror=implicit-function-declaration -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wdouble-promotion -pthread -fopenmp=libomp -MD -MT ggml/src/CMakeFiles/ggml.dir/ggml.c.o -MF ggml/src/CMakeFiles/ggml.dir/ggml.c.o.d -o ggml/src/CMakeFiles/ggml.dir/ggml.c.o -c /wrkdirs/usr/ports/m isc/llama-cpp/work/llama.cpp-b3658/ggml/src/ggml.c /wrkdirs/usr/ports/misc/llama-cpp/work/llama.cpp-b3658/ggml/src/ggml.c:655:9: warning: 'CACHE_LINE_SIZE' macro redefined [-Wmacro-redefined] #define CACHE_LINE_SIZE 64 ^ /usr/include/machine/param.h:85:9: note: previous definition is here #define CACHE_LINE_SIZE (1 << CACHE_LINE_SHIFT) ^ 1 warning generated. [ 3% 5/127] /usr/bin/cc -DGGML_BUILD -DGGML_SCHED_MAX_COPIES=4 -DGGML_SHARED -DGGML_USE_LLAMAFILE -DGGML_USE_OPENMP -DGGML_USE_VULKAN -D_XOPEN_SOURCE=600 -D__BSD_VISIBLE -Dggml_EXPORTS -I/wrkdirs/usr/ports/misc/llama-cpp/work/llama.cpp-b3658/ggml/src/../include -I/wrkdirs/usr/ports/misc/llama-cpp/work/llama.cpp-b3658/ggml/src/. -I/wrkdirs/usr/ports/misc/llama-cpp/work/.build/ggml/src -isystem /usr/local/include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -DNDEBUG -std=gnu11 -fPIC -Wshadow -Wstrict-prototypes -Wpointer-arith -Wmissing-prototypes -Werror=implicit-int -Werror=implicit-function-declaration -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wdouble-promotion -pthread -fopenmp=libomp -MD -MT ggml/src/CMakeFiles/ggml.dir/ggml-alloc.c.o -MF ggml/src/CMakeFiles/ggml.dir/ggml-alloc.c.o.d -o ggml/src/CMakeFiles/ggml.dir/ggml-alloc.c.o -c /w rkdirs/usr/ports/misc/llama-cpp/work/llama.cpp-b3658/ggml/src/ggml-alloc.c [ 4% 6/127] /usr/bin/cc -DGGML_BUILD -DGGML_SCHED_MAX_COPIES=4 -DGGML_SHARED -DGGML_USE_LLAMAFILE -DGGML_USE_OPENMP -DGGML_USE_VULKAN -D_XOPEN_SOURCE=600 -D__BSD_VISIBLE -Dggml_EXPORTS -I/wrkdirs/usr/ports/misc/llama-cpp/work/llama.cpp-b3658/ggml/src/../include -I/wrkdirs/usr/ports/misc/llama-cpp/work/llama.cpp-b3658/ggml/src/. -I/wrkdirs/usr/ports/misc/llama-cpp/work/.build/ggml/src -isystem /usr/local/include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -DNDEBUG -std=gnu11 -fPIC -Wshadow -Wstrict-prototypes -Wpointer-arith -Wmissing-prototypes -Werror=implicit-int -Werror=implicit-function-declaration -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wdouble-promotion -pthread -fopenmp=libomp -MD -MT ggml/src/CMakeFiles/ggml.dir/ggml-backend.c.o -MF ggml/src/CMakeFiles/ggml.dir/ggml-backend.c.o.d -o ggml/src/CMakeFiles/ggml.dir/ggml-backend.c.o -c /wrkdirs/usr/ports/misc/llama-cpp/work/llama.cpp-b3658/ggml/src/ggml-backend.c [ 5% 7/127] /usr/bin/cc -DGGML_BUILD -DGGML_SCHED_MAX_COPIES=4 -DGGML_SHARED -DGGML_USE_LLAMAFILE -DGGML_USE_OPENMP -DGGML_USE_VULKAN -D_XOPEN_SOURCE=600 -D__BSD_VISIBLE -Dggml_EXPORTS -I/wrkdirs/usr/ports/misc/llama-cpp/work/llama.cpp-b3658/ggml/src/../include -I/wrkdirs/usr/ports/misc/llama-cpp/work/llama.cpp-b3658/ggml/src/. -I/wrkdirs/usr/ports/misc/llama-cpp/work/.build/ggml/src -isystem /usr/local/include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -DNDEBUG -std=gnu11 -fPIC -Wshadow -Wstrict-prototypes -Wpointer-arith -Wmissing-prototypes -Werror=implicit-int -Werror=implicit-function-declaration -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wdouble-promotion -pthread -fopenmp=libomp -MD -MT ggml/src/CMakeFiles/ggml.dir/ggml-quants.c.o -MF ggml/src/CMakeFiles/ggml.dir/ggml-quants.c.o.d -o ggml/src/CMakeFiles/ggml.dir/ggml-quants.c.o -c /wrkdirs/usr/ports/misc/llama-cpp/work/llama.cpp-b3658/ggml/src/ggml-quants.c [ 6% 8/127] /usr/bin/c++ -DGGML_BUILD -DGGML_SCHED_MAX_COPIES=4 -DGGML_SHARED -DGGML_USE_LLAMAFILE -DGGML_USE_OPENMP -DGGML_USE_VULKAN -D_XOPEN_SOURCE=600 -D__BSD_VISIBLE -Dggml_EXPORTS -I/wrkdirs/usr/ports/misc/llama-cpp/work/llama.cpp-b3658/ggml/src/../include -I/wrkdirs/usr/ports/misc/llama-cpp/work/llama.cpp-b3658/ggml/src/. -I/wrkdirs/usr/ports/misc/llama-cpp/work/.build/ggml/src -isystem /usr/local/include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -DNDEBUG -std=gnu++11 -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -pthread -fopenmp=libomp -MD -MT ggml/src/CMakeFiles/ggml.dir/ggml-vulkan.cpp.o -MF ggml/src/CMakeFiles/ggml.dir/ggml-vulkan.cpp.o.d -o ggml/src/CMakeFiles/ggml.dir/ggml-vulkan.cpp.o -c /wrkdirs/usr/ports/misc/llama-cpp/work/llama.cpp-b3658/ggml /src/ggml-vulkan.cpp FAILED: ggml/src/CMakeFiles/ggml.dir/ggml-vulkan.cpp.o /usr/bin/c++ -DGGML_BUILD -DGGML_SCHED_MAX_COPIES=4 -DGGML_SHARED -DGGML_USE_LLAMAFILE -DGGML_USE_OPENMP -DGGML_USE_VULKAN -D_XOPEN_SOURCE=600 -D__BSD_VISIBLE -Dggml_EXPORTS -I/wrkdirs/usr/ports/misc/llama-cpp/work/llama.cpp-b3658/ggml/src/../include -I/wrkdirs/usr/ports/misc/llama-cpp/work/llama.cpp-b3658/ggml/src/. -I/wrkdirs/usr/ports/misc/llama-cpp/work/.build/ggml/src -isystem /usr/local/include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -DNDEBUG -std=gnu++11 -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -pthread -fopenmp=libomp -MD -MT ggml/src/CMakeFiles/ggml.dir/ggml-vulkan.cpp.o -MF ggml/src/CMakeFiles/ggml.dir/ggml-vulkan.cpp.o.d -o ggml/src/CMakeFiles/ggml.dir/ggml-vulkan.cpp.o -c /wrkdirs/usr/ports/misc/llama-cpp/work/llama.cpp-b3658/ggml/src/ggml-vulkan.cpp /wrkdirs/usr/ports/misc/llama-cpp/work/llama.cpp-b3658/ggml/src/ggml-vulkan.cpp:2626:5: error: no matching function for call to 'vkCmdCopyBuffer' vkCmdCopyBuffer(subctx->s->buffer, staging->buffer, dst->buffer, 1, &buf_copy); ^~~~~~~~~~~~~~~ /usr/local/include/vulkan/vulkan_core.h:4750:28: note: candidate function not viable: no known conversion from 'vk::Buffer' to 'VkBuffer' (aka 'unsigned long long') for 2nd argument VKAPI_ATTR void VKAPI_CALL vkCmdCopyBuffer( ^ /wrkdirs/usr/ports/misc/llama-cpp/work/llama.cpp-b3658/ggml/src/ggml-vulkan.cpp:2701:5: error: no matching function for call to 'vkCmdCopyBuffer' vkCmdCopyBuffer(subctx->s->buffer, staging_buffer->buffer, dst->buffer, 1, &buf_copy); ^~~~~~~~~~~~~~~ /usr/local/include/vulkan/vulkan_core.h:4750:28: note: candidate function not viable: no known conversion from 'vk::Buffer' to 'VkBuffer' (aka 'unsigned long long') for 2nd argument VKAPI_ATTR void VKAPI_CALL vkCmdCopyBuffer( ^ /wrkdirs/usr/ports/misc/llama-cpp/work/llama.cpp-b3658/ggml/src/ggml-vulkan.cpp:2833:5: error: no matching function for call to 'vkCmdCopyBuffer' vkCmdCopyBuffer(ctx->s->buffer, src->buffer, dst->buffer, 1, &bc); ^~~~~~~~~~~~~~~ /usr/local/include/vulkan/vulkan_core.h:4750:28: note: candidate function not viable: no known conversion from 'vk::Buffer' to 'VkBuffer' (aka 'unsigned long long') for 2nd argument VKAPI_ATTR void VKAPI_CALL vkCmdCopyBuffer( ^ /wrkdirs/usr/ports/misc/llama-cpp/work/llama.cpp-b3658/ggml/src/ggml-vulkan.cpp:6768:61: warning: format specifies type 'long' but the argument has type 'size_t' (aka 'unsigned int') [-Wformat] snprintf(name, sizeof(name), "%s%ld", GGML_VK_NAME, i); ~~~ ^ %zu 1 warning and 3 errors generated. ninja: build stopped: subcommand failed. *** Error code 1 Stop. make: stopped in /usr/ports/misc/llama-cpp