Skip to content

Eval bug: clip.cpp has no GPU support - a lot of work is at risk #11322

Closed
@cmp-nct

Description

@cmp-nct

Name and Version

all versions 2025

Operating systems

Linux

GGML backends

CUDA, Metal

Hardware

any

Models

No response

Problem description & steps to reproduce

In PR #10896 the GPU support of clip.cpp has been removed, it's basically just a few comments around fullly functional code.

Hundreds of hours went into CLIP and having it on GPU support was a major feat for the vision capabilities of llama.cpp, this also caused large SOTA models to be implemented in llama.cpp with people working dedicated on those patches.

I agree that the vision implementation was not great but we've had SOTA support by models like minicpm-v-2.6 and even the new minicpmv-o-2.6 has been implemented in llama.cpp at launch with dedicated people working on it with a current PR waiting for merge.

This change renders those models useless for anyone who is not aware on how to hack the llama.cpp code. It takes minutes instead of milliseconds now.

I strongly recommend to add the GPU support back into clip.cpp while it is still compatible (currently it is) with the core so people can use llama.cpp with vision capabilities again.
To prevent people posting unwelcome issues, add a warning message instead.
Right now we see issues being created that vision support is not working anymore, that GPU support is failing etc. Longterm it will cause developers to stop support llama.cpp as engine for vision.

First Bad Commit

#10896

Relevant log output

-

Metadata

Metadata

Assignees

No one assigned

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions