******************************************************************************** conan install llama-cpp/b3542@#b9cee04f44b2458374fd718dd683caf6 --build=llama-cpp -pr C:/J/workspace/prod-v1/bsr/84752/fadbb/profile_windows_16_mdd_vs_debug_64.llama-cpp-shared-True.txt -c tools.system.package_manager:mode=install -c tools.system.package_manager:sudo=True ******************************************************************************** Auto detecting your dev setup to initialize the default profile (C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\profiles\default) Found Visual Studio 17 Default settings os=Windows os_build=Windows arch=x86_64 arch_build=x86_64 compiler=Visual Studio compiler.version=17 build_type=Release *** You can change them in C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\profiles\default *** *** Or override with -s compiler='other' -s ...s*** Configuration: [settings] arch=x86_64 build_type=Debug compiler=Visual Studio compiler.runtime=MDd compiler.version=16 os=Windows [options] llama-cpp:shared=True [build_requires] [env] [conf] tools.system.package_manager:mode=install tools.system.package_manager:sudo=True llama-cpp/b3542: Forced build from source Installing package: llama-cpp/b3542 Requirements llama-cpp/b3542 from local cache - Cache Packages llama-cpp/b3542:4f1710918aa542fccb5a54d7bd712e4b0750b50d - Build Installing (downloading, building) binaries... [HOOK - conan-center.py] pre_source(): [IMMUTABLE SOURCES (KB-H010)] OK llama-cpp/b3542: Configuring sources in C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\source\src llama-cpp/b3542: [HOOK - conan-center.py] post_source(): [LIBCXX MANAGEMENT (KB-H011)] OK [HOOK - conan-center.py] post_source(): [CPPSTD MANAGEMENT (KB-H022)] OK [HOOK - conan-center.py] post_source(): [SHORT_PATHS USAGE (KB-H066)] OK llama-cpp/b3542: Copying sources to build folder llama-cpp/b3542: Building your package in C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d llama-cpp/b3542: Generator txt created conanbuildinfo.txt llama-cpp/b3542: Calling generate() llama-cpp/b3542: Preset 'default' added to CMakePresets.json. Invoke it manually using 'cmake --preset default' llama-cpp/b3542: If your CMake version is not compatible with CMakePresets (<3.19) call cmake like: 'cmake -G "Visual Studio 16 2019" -DCMAKE_TOOLCHAIN_FILE=C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\build\generators\conan_toolchain.cmake -DCMAKE_POLICY_DEFAULT_CMP0091=NEW' llama-cpp/b3542: Aggregating env generators [HOOK - conan-center.py] pre_build(): [FPIC MANAGEMENT (KB-H007)] 'fPIC' option not found [HOOK - conan-center.py] pre_build(): [FPIC MANAGEMENT (KB-H007)] OK llama-cpp/b3542: Calling build() llama-cpp/b3542: CMake command: cmake -G "Visual Studio 16 2019" -DCMAKE_TOOLCHAIN_FILE="C:/J/workspace/prod-v1/bsr/84752/dabcc/.conan/data/llama-cpp/b3542/_/_/build/4f1710918aa542fccb5a54d7bd712e4b0750b50d/build/generators/conan_toolchain.cmake" -DCMAKE_INSTALL_PREFIX="C:/J/workspace/prod-v1/bsr/84752/dabcc/.conan/data/llama-cpp/b3542/_/_/package/4f1710918aa542fccb5a54d7bd712e4b0750b50d" -DCMAKE_POLICY_DEFAULT_CMP0091="NEW" "C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\src" ----Running------ > cmake -G "Visual Studio 16 2019" -DCMAKE_TOOLCHAIN_FILE="C:/J/workspace/prod-v1/bsr/84752/dabcc/.conan/data/llama-cpp/b3542/_/_/build/4f1710918aa542fccb5a54d7bd712e4b0750b50d/build/generators/conan_toolchain.cmake" -DCMAKE_INSTALL_PREFIX="C:/J/workspace/prod-v1/bsr/84752/dabcc/.conan/data/llama-cpp/b3542/_/_/package/4f1710918aa542fccb5a54d7bd712e4b0750b50d" -DCMAKE_POLICY_DEFAULT_CMP0091="NEW" "C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\src" ----------------- -- Using Conan toolchain: C:/J/workspace/prod-v1/bsr/84752/dabcc/.conan/data/llama-cpp/b3542/_/_/build/4f1710918aa542fccb5a54d7bd712e4b0750b50d/build/generators/conan_toolchain.cmake -- Conan toolchain: Setting BUILD_SHARED_LIBS = ON -- The C compiler identification is MSVC 19.29.30148.0 -- The CXX compiler identification is MSVC 19.29.30148.0 -- Detecting C compiler ABI info -- Detecting C compiler ABI info - done -- Check for working C compiler: C:/Program Files (x86)/Microsoft Visual Studio/2019/Community/VC/Tools/MSVC/14.29.30133/bin/Hostx64/x64/cl.exe - skipped -- Detecting C compile features -- Detecting C compile features - done -- Detecting CXX compiler ABI info -- Detecting CXX compiler ABI info - done -- Check for working CXX compiler: C:/Program Files (x86)/Microsoft Visual Studio/2019/Community/VC/Tools/MSVC/14.29.30133/bin/Hostx64/x64/cl.exe - skipped -- Detecting CXX compile features -- Detecting CXX compile features - done -- Found Git: C:/Program Files/Git/cmd/git.exe (found version "2.29.0.windows.1") -- Looking for pthread.h -- Looking for pthread.h - not found -- Found Threads: TRUE -- Found OpenMP_C: -openmp (found version "2.0") -- Found OpenMP_CXX: -openmp (found version "2.0") -- Found OpenMP: TRUE (found version "2.0") -- OpenMP found -- Using llamafile -- Warning: ccache not found - consider installing it for faster compilation or disable this warning with GGML_CCACHE=OFF -- CMAKE_SYSTEM_PROCESSOR: AMD64 -- CMAKE_GENERATOR_PLATFORM: x64 -- x86 detected -- Performing Test HAS_AVX_1 -- Performing Test HAS_AVX_1 - Success -- Performing Test HAS_AVX2_1 -- Performing Test HAS_AVX2_1 - Success -- Performing Test HAS_FMA_1 -- Performing Test HAS_FMA_1 - Success -- Performing Test HAS_AVX512_1 -- Performing Test HAS_AVX512_1 - Failed -- Performing Test HAS_AVX512_2 -- Performing Test HAS_AVX512_2 - Failed -- Configuring done -- Generating done -- Build files have been written to: C:/J/workspace/prod-v1/bsr/84752/dabcc/.conan/data/llama-cpp/b3542/_/_/build/4f1710918aa542fccb5a54d7bd712e4b0750b50d/build llama-cpp/b3542: CMake command: cmake --build "C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\build" --config Debug ----Running------ > cmake --build "C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\build" --config Debug ----------------- Microsoft (R) Build Engine version 16.11.2+f32259642 for .NET Framework Copyright (C) Microsoft Corporation. All rights reserved. Checking Build System Generating build details from Git -- Found Git: C:/Program Files/Git/cmd/git.exe (found version "2.29.0.windows.1") fatal: not a git repository (or any of the parent directories): .git fatal: not a git repository (or any of the parent directories): .git Building Custom Rule C:/J/workspace/prod-v1/bsr/84752/dabcc/.conan/data/llama-cpp/b3542/_/_/build/4f1710918aa542fccb5a54d7bd712e4b0750b50d/src/common/CMakeLists.txt build-info.cpp build_info.vcxproj -> C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\build\common\build_info.dir\Debug\build_info.lib Building Custom Rule C:/J/workspace/prod-v1/bsr/84752/dabcc/.conan/data/llama-cpp/b3542/_/_/build/4f1710918aa542fccb5a54d7bd712e4b0750b50d/src/ggml/src/CMakeLists.txt ggml.c ggml-alloc.c ggml-backend.c ggml-quants.c ggml-aarch64.c C:\Program Files (x86)\Windows Kits\10\Include\10.0.20348.0\ucrt\assert.h(21,1): warning C4005: 'static_assert': macro redefinition (compiling source file C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\src\ggml\src\ggml-aarch64.c) [C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\build\ggml\src\ggml.vcxproj] C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\src\ggml\src\ggml-common.h(62): message : see previous definition of 'static_assert' (compiling source file C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\src\ggml\src\ggml-aarch64.c) [C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\build\ggml\src\ggml.vcxproj] C:\Program Files (x86)\Windows Kits\10\Include\10.0.20348.0\ucrt\assert.h(21,1): warning C4005: 'static_assert': macro redefinition (compiling source file C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\src\ggml\src\ggml-quants.c) [C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\build\ggml\src\ggml.vcxproj] C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\src\ggml\src\ggml-common.h(62): message : see previous definition of 'static_assert' (compiling source file C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\src\ggml\src\ggml-quants.c) [C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\build\ggml\src\ggml.vcxproj] C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\src\ggml\src\ggml-aarch64.c(165,36): warning C4244: '=': conversion from 'float' to 'int8_t', possible loss of data [C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\build\ggml\src\ggml.vcxproj] C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\src\ggml\src\ggml-aarch64.c(80,1): warning C4244: 'initializing': conversion from 'int64_t' to 'const int', possible loss of data [C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\build\ggml\src\ggml.vcxproj] C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\src\ggml\src\ggml-aarch64.c(283,36): warning C4244: '=': conversion from 'float' to 'int8_t', possible loss of data [C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\build\ggml\src\ggml.vcxproj] C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\src\ggml\src\ggml-aarch64.c(174,1): warning C4244: 'initializing': conversion from 'int64_t' to 'const int', possible loss of data [C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\build\ggml\src\ggml.vcxproj] C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\src\ggml\src\ggml-aarch64.c(315,79): warning C4244: '+=': conversion from 'int64_t' to 'int', possible loss of data [C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\build\ggml\src\ggml.vcxproj] C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\src\ggml\src\ggml-aarch64.c(303,1): warning C4244: 'initializing': conversion from 'int64_t' to 'const int', possible loss of data [C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\build\ggml\src\ggml.vcxproj] sgemm.cpp Auto build dll exports Creating library C:/J/workspace/prod-v1/bsr/84752/dabcc/.conan/data/llama-cpp/b3542/_/_/build/4f1710918aa542fccb5a54d7bd712e4b0750b50d/build/ggml/src/Debug/ggml.lib and object C:/J/workspace/prod-v1/bsr/84752/dabcc/.conan/data/llama-cpp/b3542/_/_/build/4f1710918aa542fccb5a54d7bd712e4b0750b50d/build/ggml/src/Debug/ggml.exp ggml.vcxproj -> C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\build\bin\Debug\ggml.dll Building Custom Rule C:/J/workspace/prod-v1/bsr/84752/dabcc/.conan/data/llama-cpp/b3542/_/_/build/4f1710918aa542fccb5a54d7bd712e4b0750b50d/src/src/CMakeLists.txt llama.cpp llama-vocab.cpp llama-grammar.cpp llama-sampling.cpp unicode.cpp unicode-data.cpp C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\src\src\llama-sampling.cpp(26,20): warning C4244: '=': conversion from 'time_t' to 'uint32_t', possible loss of data [C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\build\src\llama.vcxproj] C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\src\src\llama-sampling.cpp(70,29): warning C4267: '=': conversion from 'size_t' to 'int32_t', possible loss of data [C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\build\src\llama.vcxproj] C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\src\src\llama-sampling.cpp(405,33): warning C4244: '=': conversion from 'double' to 'float', possible loss of data [C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\build\src\llama.vcxproj] C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\src\src\llama-sampling.cpp(409,48): warning C4244: '/=': conversion from 'double' to 'float', possible loss of data [C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\build\src\llama.vcxproj] C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\src\src\llama-sampling.cpp(510,34): warning C4244: 'initializing': conversion from 'float' to 'int32_t', possible loss of data [C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\build\src\llama.vcxproj] C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\src\src\llama-vocab.cpp(152,26): warning C4244: 'return': conversion from 'long' to 'uint8_t', possible loss of data [C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\build\src\llama.vcxproj] C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\src\src\llama-sampling.cpp(510,27): warning C4244: 'initializing': conversion from 'float' to 'const int32_t', possible loss of data [C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\build\src\llama.vcxproj] C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\src\src\llama-sampling.cpp(530,61): warning C4244: 'argument': conversion from 'const int32_t' to 'float', possible loss of data [C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\build\src\llama.vcxproj] C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\src\src\llama-vocab.cpp(225,1): warning C4267: 'argument': conversion from 'size_t' to 'int', possible loss of data [C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\build\src\llama.vcxproj] C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\src\src\llama-vocab.cpp(529,1): warning C4267: 'argument': conversion from 'size_t' to 'int', possible loss of data [C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\build\src\llama.vcxproj] C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\src\src\llama-vocab.cpp(569,1): warning C4267: '=': conversion from 'size_t' to 'llm_symbol::index', possible loss of data [C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\build\src\llama.vcxproj] C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\src\src\llama-vocab.cpp(572,1): warning C4267: '=': conversion from 'size_t' to 'int', possible loss of data [C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\build\src\llama.vcxproj] C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\src\src\llama-vocab.cpp(666,1): warning C4267: 'initializing': conversion from 'size_t' to 'int', possible loss of data [C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\build\src\llama.vcxproj] C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\src\src\llama-vocab.cpp(666,1): warning C4267: 'initializing': conversion from 'size_t' to 'const int', possible loss of data [C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\build\src\llama.vcxproj] C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\src\src\llama-vocab.cpp(1529,22): warning C4267: 'return': conversion from 'size_t' to 'int32_t', possible loss of data [C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\build\src\llama.vcxproj] Auto build dll exports Creating library C:/J/workspace/prod-v1/bsr/84752/dabcc/.conan/data/llama-cpp/b3542/_/_/build/4f1710918aa542fccb5a54d7bd712e4b0750b50d/build/src/Debug/llama.lib and object C:/J/workspace/prod-v1/bsr/84752/dabcc/.conan/data/llama-cpp/b3542/_/_/build/4f1710918aa542fccb5a54d7bd712e4b0750b50d/build/src/Debug/llama.exp llama.vcxproj -> C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\build\bin\Debug\llama.dll Building Custom Rule C:/J/workspace/prod-v1/bsr/84752/dabcc/.conan/data/llama-cpp/b3542/_/_/build/4f1710918aa542fccb5a54d7bd712e4b0750b50d/src/common/CMakeLists.txt common.cpp sampling.cpp console.cpp grammar-parser.cpp json-schema-to-grammar.cpp train.cpp ngram-cache.cpp C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\src\common\console.cpp(253,38): warning C4267: 'initializing': conversion from 'size_t' to 'DWORD', possible loss of data [C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\build\common\common.vcxproj] C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\src\common\console.cpp(407,43): warning C4267: 'initializing': conversion from 'size_t' to 'int', possible loss of data [C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\build\common\common.vcxproj] C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\src\common\json-schema-to-grammar.cpp(139,1): warning C4267: 'argument': conversion from 'size_t' to 'int', possible loss of data [C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\build\common\common.vcxproj] C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\src\common\json-schema-to-grammar.cpp(154,1): warning C4267: 'argument': conversion from 'size_t' to 'int', possible loss of data [C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\build\common\common.vcxproj] C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\src\common\json-schema-to-grammar.cpp(234,47): warning C4267: 'argument': conversion from 'size_t' to 'int', possible loss of data [C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\build\common\common.vcxproj] C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\src\common\json-schema-to-grammar.cpp(245,51): warning C4267: 'argument': conversion from 'size_t' to 'int', possible loss of data [C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\build\common\common.vcxproj] C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\src\common\ngram-cache.cpp(20,50): warning C4244: 'argument': conversion from 'int64_t' to 'const int', possible loss of data [C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\build\common\common.vcxproj] C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\src\common\sampling.cpp(105,47): warning C4267: 'initializing': conversion from 'size_t' to 'int', possible loss of data [C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\build\common\common.vcxproj] C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\src\common\sampling.cpp(105,47): warning C4267: 'initializing': conversion from 'size_t' to 'const int', possible loss of data [C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\build\common\common.vcxproj] C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\src\common\ngram-cache.cpp(100,5): warning C4267: 'initializing': conversion from 'size_t' to 'int', possible loss of data [C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\build\common\common.vcxproj] C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\src\common\ngram-cache.cpp(147,36): warning C4267: 'initializing': conversion from 'size_t' to 'int', possible loss of data [C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\build\common\common.vcxproj] C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\src\common\ngram-cache.cpp(147,36): warning C4267: 'initializing': conversion from 'size_t' to 'const int', possible loss of data [C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\build\common\common.vcxproj] C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\src\common\ngram-cache.cpp(156,84): warning C4267: 'initializing': conversion from 'size_t' to 'int', possible loss of data [C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\build\common\common.vcxproj] C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\src\common\ngram-cache.cpp(156,84): warning C4267: 'initializing': conversion from 'size_t' to 'const int', possible loss of data [C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\build\common\common.vcxproj] C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\src\common\ngram-cache.cpp(170,79): warning C4267: 'initializing': conversion from 'size_t' to 'int', possible loss of data [C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\build\common\common.vcxproj] C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\src\common\ngram-cache.cpp(170,79): warning C4267: 'initializing': conversion from 'size_t' to 'const int', possible loss of data [C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\build\common\common.vcxproj] C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\src\common\ngram-cache.cpp(202,52): warning C4267: 'initializing': conversion from 'size_t' to 'int32_t', possible loss of data [C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\build\common\common.vcxproj] C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\src\common\ngram-cache.cpp(202,52): warning C4267: 'initializing': conversion from 'size_t' to 'const int32_t', possible loss of data [C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\build\common\common.vcxproj] C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\src\common\json-schema-to-grammar.cpp(558,60): warning C4101: 'e': unreferenced local variable [C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\build\common\common.vcxproj] common.vcxproj -> C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\build\common\Debug\common.lib Building Custom Rule C:/J/workspace/prod-v1/bsr/84752/dabcc/.conan/data/llama-cpp/b3542/_/_/build/4f1710918aa542fccb5a54d7bd712e4b0750b50d/src/CMakeLists.txt llama-cpp/b3542: Package '4f1710918aa542fccb5a54d7bd712e4b0750b50d' built llama-cpp/b3542: Build folder C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\build llama-cpp/b3542: Generated conaninfo.txt llama-cpp/b3542: Generated conanbuildinfo.txt llama-cpp/b3542: Generating the package llama-cpp/b3542: Package folder C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\package\4f1710918aa542fccb5a54d7bd712e4b0750b50d llama-cpp/b3542: Calling package() llama-cpp/b3542: Copied 1 file: LICENSE llama-cpp/b3542: CMake command: cmake --install "C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\build" --config Debug --prefix "C:/J/workspace/prod-v1/bsr/84752/dabcc/.conan/data/llama-cpp/b3542/_/_/package/4f1710918aa542fccb5a54d7bd712e4b0750b50d" ----Running------ > cmake --install "C:\J\workspace\prod-v1\bsr\84752\dabcc\.conan\data\llama-cpp\b3542\_\_\build\4f1710918aa542fccb5a54d7bd712e4b0750b50d\build" --config Debug --prefix "C:/J/workspace/prod-v1/bsr/84752/dabcc/.conan/data/llama-cpp/b3542/_/_/package/4f1710918aa542fccb5a54d7bd712e4b0750b50d" ----------------- -- Installing: C:/J/workspace/prod-v1/bsr/84752/dabcc/.conan/data/llama-cpp/b3542/_/_/package/4f1710918aa542fccb5a54d7bd712e4b0750b50d/lib/ggml.lib -- Installing: C:/J/workspace/prod-v1/bsr/84752/dabcc/.conan/data/llama-cpp/b3542/_/_/package/4f1710918aa542fccb5a54d7bd712e4b0750b50d/bin/ggml.dll -- Installing: C:/J/workspace/prod-v1/bsr/84752/dabcc/.conan/data/llama-cpp/b3542/_/_/package/4f1710918aa542fccb5a54d7bd712e4b0750b50d/include/ggml.h -- Installing: C:/J/workspace/prod-v1/bsr/84752/dabcc/.conan/data/llama-cpp/b3542/_/_/package/4f1710918aa542fccb5a54d7bd712e4b0750b50d/include/ggml-alloc.h -- Installing: C:/J/workspace/prod-v1/bsr/84752/dabcc/.conan/data/llama-cpp/b3542/_/_/package/4f1710918aa542fccb5a54d7bd712e4b0750b50d/include/ggml-backend.h -- Installing: C:/J/workspace/prod-v1/bsr/84752/dabcc/.conan/data/llama-cpp/b3542/_/_/package/4f1710918aa542fccb5a54d7bd712e4b0750b50d/include/ggml-blas.h -- Installing: C:/J/workspace/prod-v1/bsr/84752/dabcc/.conan/data/llama-cpp/b3542/_/_/package/4f1710918aa542fccb5a54d7bd712e4b0750b50d/include/ggml-cann.h -- Installing: C:/J/workspace/prod-v1/bsr/84752/dabcc/.conan/data/llama-cpp/b3542/_/_/package/4f1710918aa542fccb5a54d7bd712e4b0750b50d/include/ggml-cuda.h -- Up-to-date: C:/J/workspace/prod-v1/bsr/84752/dabcc/.conan/data/llama-cpp/b3542/_/_/package/4f1710918aa542fccb5a54d7bd712e4b0750b50d/include/ggml.h -- Installing: C:/J/workspace/prod-v1/bsr/84752/dabcc/.conan/data/llama-cpp/b3542/_/_/package/4f1710918aa542fccb5a54d7bd712e4b0750b50d/include/ggml-kompute.h -- Installing: C:/J/workspace/prod-v1/bsr/84752/dabcc/.conan/data/llama-cpp/b3542/_/_/package/4f1710918aa542fccb5a54d7bd712e4b0750b50d/include/ggml-metal.h -- Installing: C:/J/workspace/prod-v1/bsr/84752/dabcc/.conan/data/llama-cpp/b3542/_/_/package/4f1710918aa542fccb5a54d7bd712e4b0750b50d/include/ggml-rpc.h -- Installing: C:/J/workspace/prod-v1/bsr/84752/dabcc/.conan/data/llama-cpp/b3542/_/_/package/4f1710918aa542fccb5a54d7bd712e4b0750b50d/include/ggml-sycl.h -- Installing: C:/J/workspace/prod-v1/bsr/84752/dabcc/.conan/data/llama-cpp/b3542/_/_/package/4f1710918aa542fccb5a54d7bd712e4b0750b50d/include/ggml-vulkan.h -- Up-to-date: C:/J/workspace/prod-v1/bsr/84752/dabcc/.conan/data/llama-cpp/b3542/_/_/package/4f1710918aa542fccb5a54d7bd712e4b0750b50d/lib/ggml.lib -- Up-to-date: C:/J/workspace/prod-v1/bsr/84752/dabcc/.conan/data/llama-cpp/b3542/_/_/package/4f1710918aa542fccb5a54d7bd712e4b0750b50d/bin/ggml.dll -- Up-to-date: C:/J/workspace/prod-v1/bsr/84752/dabcc/.conan/data/llama-cpp/b3542/_/_/package/4f1710918aa542fccb5a54d7bd712e4b0750b50d/include/ggml.h -- Up-to-date: C:/J/workspace/prod-v1/bsr/84752/dabcc/.conan/data/llama-cpp/b3542/_/_/package/4f1710918aa542fccb5a54d7bd712e4b0750b50d/include/ggml-alloc.h -- Up-to-date: C:/J/workspace/prod-v1/bsr/84752/dabcc/.conan/data/llama-cpp/b3542/_/_/package/4f1710918aa542fccb5a54d7bd712e4b0750b50d/include/ggml-backend.h -- Up-to-date: C:/J/workspace/prod-v1/bsr/84752/dabcc/.conan/data/llama-cpp/b3542/_/_/package/4f1710918aa542fccb5a54d7bd712e4b0750b50d/include/ggml-blas.h -- Up-to-date: C:/J/workspace/prod-v1/bsr/84752/dabcc/.conan/data/llama-cpp/b3542/_/_/package/4f1710918aa542fccb5a54d7bd712e4b0750b50d/include/ggml-cann.h -- Up-to-date: C:/J/workspace/prod-v1/bsr/84752/dabcc/.conan/data/llama-cpp/b3542/_/_/package/4f1710918aa542fccb5a54d7bd712e4b0750b50d/include/ggml-cuda.h -- Up-to-date: C:/J/workspace/prod-v1/bsr/84752/dabcc/.conan/data/llama-cpp/b3542/_/_/package/4f1710918aa542fccb5a54d7bd712e4b0750b50d/include/ggml.h -- Up-to-date: C:/J/workspace/prod-v1/bsr/84752/dabcc/.conan/data/llama-cpp/b3542/_/_/package/4f1710918aa542fccb5a54d7bd712e4b0750b50d/include/ggml-kompute.h -- Up-to-date: C:/J/workspace/prod-v1/bsr/84752/dabcc/.conan/data/llama-cpp/b3542/_/_/package/4f1710918aa542fccb5a54d7bd712e4b0750b50d/include/ggml-metal.h -- Up-to-date: C:/J/workspace/prod-v1/bsr/84752/dabcc/.conan/data/llama-cpp/b3542/_/_/package/4f1710918aa542fccb5a54d7bd712e4b0750b50d/include/ggml-rpc.h -- Up-to-date: C:/J/workspace/prod-v1/bsr/84752/dabcc/.conan/data/llama-cpp/b3542/_/_/package/4f1710918aa542fccb5a54d7bd712e4b0750b50d/include/ggml-sycl.h -- Up-to-date: C:/J/workspace/prod-v1/bsr/84752/dabcc/.conan/data/llama-cpp/b3542/_/_/package/4f1710918aa542fccb5a54d7bd712e4b0750b50d/include/ggml-vulkan.h -- Installing: C:/J/workspace/prod-v1/bsr/84752/dabcc/.conan/data/llama-cpp/b3542/_/_/package/4f1710918aa542fccb5a54d7bd712e4b0750b50d/lib/llama.lib -- Installing: C:/J/workspace/prod-v1/bsr/84752/dabcc/.conan/data/llama-cpp/b3542/_/_/package/4f1710918aa542fccb5a54d7bd712e4b0750b50d/bin/llama.dll -- Installing: C:/J/workspace/prod-v1/bsr/84752/dabcc/.conan/data/llama-cpp/b3542/_/_/package/4f1710918aa542fccb5a54d7bd712e4b0750b50d/include/llama.h -- Installing: C:/J/workspace/prod-v1/bsr/84752/dabcc/.conan/data/llama-cpp/b3542/_/_/package/4f1710918aa542fccb5a54d7bd712e4b0750b50d/lib/cmake/llama/llama-config.cmake -- Installing: C:/J/workspace/prod-v1/bsr/84752/dabcc/.conan/data/llama-cpp/b3542/_/_/package/4f1710918aa542fccb5a54d7bd712e4b0750b50d/lib/cmake/llama/llama-version.cmake -- Installing: C:/J/workspace/prod-v1/bsr/84752/dabcc/.conan/data/llama-cpp/b3542/_/_/package/4f1710918aa542fccb5a54d7bd712e4b0750b50d/bin/convert_hf_to_gguf.py -- Installing: C:/J/workspace/prod-v1/bsr/84752/dabcc/.conan/data/llama-cpp/b3542/_/_/package/4f1710918aa542fccb5a54d7bd712e4b0750b50d/lib/pkgconfig/llama.pc llama-cpp/b3542: Copied 1 file: .editorconfig llama-cpp/b3542: Copied 16 '.gguf' files llama-cpp/b3542: Copied 13 '.inp' files llama-cpp/b3542: Copied 13 '.out' files llama-cpp/b3542: Copied 2 '.hpp' files: base64.hpp, json.hpp llama-cpp/b3542: Copied 9 '.h' files llama-cpp/b3542: Copied 2 '.lib' files: build_info.lib, common.lib llama-cpp/b3542: Copied 1 '.cmake' file: llama-cpp-cuda-static.cmake [HOOK - conan-center.py] post_package(): [PACKAGE LICENSE (KB-H012)] OK [HOOK - conan-center.py] post_package(): [DEFAULT PACKAGE LAYOUT (KB-H013)] OK [HOOK - conan-center.py] post_package(): [MATCHING CONFIGURATION (KB-H014)] OK [HOOK - conan-center.py] post_package(): [SHARED ARTIFACTS (KB-H015)] OK [HOOK - conan-center.py] post_package(): [STATIC ARTIFACTS (KB-H074)] OK [HOOK - conan-center.py] post_package(): [EITHER STATIC OR SHARED OF EACH LIB (KB-H076)] OK [HOOK - conan-center.py] post_package(): [PC-FILES (KB-H020)] OK [HOOK - conan-center.py] post_package(): [CMAKE-MODULES-CONFIG-FILES (KB-H016)] OK [HOOK - conan-center.py] post_package(): [PDB FILES NOT ALLOWED (KB-H017)] OK [HOOK - conan-center.py] post_package(): [LIBTOOL FILES PRESENCE (KB-H018)] OK [HOOK - conan-center.py] post_package(): [MS RUNTIME FILES (KB-H021)] OK [HOOK - conan-center.py] post_package(): [SHORT_PATHS USAGE (KB-H066)] OK ********************************************************************** ** Visual Studio 2019 Developer Command Prompt v16.11.26 ** Copyright (c) 2021 Microsoft Corporation ********************************************************************** [vcvarsall.bat] Environment initialized for: 'x64' [HOOK - conan-center.py] post_package(): [MISSING SYSTEM LIBS (KB-H043)] OK [HOOK - conan-center.py] post_package(): [APPLE RELOCATABLE SHARED LIBS (KB-H077)] OK llama-cpp/b3542 package(): Packaged 1 '.py' file: convert_hf_to_gguf.py llama-cpp/b3542 package(): Packaged 2 '.dll' files: ggml.dll, llama.dll llama-cpp/b3542 package(): Packaged 21 '.h' files llama-cpp/b3542 package(): Packaged 2 '.hpp' files: base64.hpp, json.hpp llama-cpp/b3542 package(): Packaged 4 '.lib' files: build_info.lib, common.lib, ggml.lib, llama.lib llama-cpp/b3542 package(): Packaged 1 '.cmake' file: llama-cpp-cuda-static.cmake llama-cpp/b3542 package(): Packaged 2 files: LICENSE, .editorconfig llama-cpp/b3542 package(): Packaged 16 '.gguf' files llama-cpp/b3542 package(): Packaged 13 '.inp' files llama-cpp/b3542 package(): Packaged 13 '.out' files llama-cpp/b3542: Package '4f1710918aa542fccb5a54d7bd712e4b0750b50d' created llama-cpp/b3542: Created package revision 386b32d9a1655a202a7b3d2c85652af7 [HOOK - conan-center.py] post_package_info(): [CMAKE FILE NOT IN BUILD FOLDERS (KB-H019)] OK [HOOK - conan-center.py] post_package_info(): [LIBRARY DOES NOT EXIST (KB-H054)] OK [HOOK - conan-center.py] post_package_info(): [INCLUDE PATH DOES NOT EXIST (KB-H071)] OK Aggregating env generators fatal: not a git repository (or any of the parent directories): .git fatal: not a git repository (or any of the parent directories): .git CMake Warning at common/CMakeLists.txt:30 (message): Git repository not found; to enable automatic generation of build info, make sure Git is installed and the project is a Git repository. WARN: *** Conan 1 is legacy and on a deprecation path *** WARN: *** Please upgrade to Conan 2 *** llama-cpp/b3542: WARN: Using the new toolchains and generators without specifying a build profile (e.g: -pr:b=default) is discouraged and might cause failures and unexpected behavior llama-cpp/b3542: WARN: Using the new toolchains and generators without specifying a build profile (e.g: -pr:b=default) is discouraged and might cause failures and unexpected behavior [HOOK - conan-center.py] post_package_info(): WARN: [CMAKE FILE NOT IN BUILD FOLDERS (KB-H019)] The *.cmake files have to be placed in a folder declared as `cpp_info.builddirs`. Currently folders declared: {'C:/J/workspace/prod-v1/bsr/84752/dabcc/.conan/data/llama-cpp/b3542/_/_/package/4f1710918aa542fccb5a54d7bd712e4b0750b50d/'} [HOOK - conan-center.py] post_package_info(): WARN: [CMAKE FILE NOT IN BUILD FOLDERS (KB-H019)] Found files: ./lib/cmake/llama-cpp-cuda-static.cmake