From d0fd9e5aa2b6cd5a029427f03d44c1e29f469e07 Mon Sep 17 00:00:00 2001 From: Gabe Goodhart Date: Tue, 24 Jun 2025 17:47:36 -0600 Subject: [PATCH] fix: Remove mtmd main cpp files Branch: GraniteFour Signed-off-by: Gabe Goodhart --- .../tools/mtmd/deprecation-warning.cpp | 22 - llama/llama.cpp/tools/mtmd/mtmd-cli.cpp | 386 ------------------ 2 files changed, 408 deletions(-) delete mode 100644 llama/llama.cpp/tools/mtmd/deprecation-warning.cpp delete mode 100644 llama/llama.cpp/tools/mtmd/mtmd-cli.cpp diff --git a/llama/llama.cpp/tools/mtmd/deprecation-warning.cpp b/llama/llama.cpp/tools/mtmd/deprecation-warning.cpp deleted file mode 100644 index dded0a56a..000000000 --- a/llama/llama.cpp/tools/mtmd/deprecation-warning.cpp +++ /dev/null @@ -1,22 +0,0 @@ -#include -#include - -int main(int argc, char** argv) { - std::string filename = "main"; - if (argc >= 1) { - filename = argv[0]; - } - - // Get only the program name from the full path - size_t pos = filename.find_last_of("/\\"); - if (pos != std::string::npos) { - filename = filename.substr(pos+1); - } - - fprintf(stdout, "\n"); - fprintf(stdout, "WARNING: The binary '%s' is deprecated.\n", filename.c_str()); - fprintf(stdout, "Please use 'llama-mtmd-cli' instead.\n"); - fprintf(stdout, "\n"); - - return EXIT_FAILURE; -} diff --git a/llama/llama.cpp/tools/mtmd/mtmd-cli.cpp b/llama/llama.cpp/tools/mtmd/mtmd-cli.cpp deleted file mode 100644 index 599e682e0..000000000 --- a/llama/llama.cpp/tools/mtmd/mtmd-cli.cpp +++ /dev/null @@ -1,386 +0,0 @@ -#include "arg.h" -#include "log.h" -#include "common.h" -#include "sampling.h" -#include "llama.h" -#include "ggml.h" -#include "console.h" -#include "chat.h" -#include "mtmd.h" -#include "mtmd-helper.h" - -#include -#include -#include - -#if defined (__unix__) || (defined (__APPLE__) && defined (__MACH__)) -#include -#include -#elif defined (_WIN32) -#define WIN32_LEAN_AND_MEAN -#ifndef NOMINMAX -#define NOMINMAX -#endif -#include -#include -#endif - -// volatile, because of signal being an interrupt -static volatile bool g_is_generating = false; -static volatile bool g_is_interrupted = false; - -/** - * Please note that this is NOT a production-ready stuff. - * It is a playground for trying multimodal support in llama.cpp. - * For contributors: please keep this code simple and easy to understand. - */ - -static void show_additional_info(int /*argc*/, char ** argv) { - LOG( - "Experimental CLI for multimodal\n\n" - "Usage: %s [options] -m --mmproj --image --audio