Skip to content

Conversation

@hmellor
Copy link
Member

@hmellor hmellor commented Dec 4, 2025

Phi 4 multimodal is currently compatible with vLLM in 2 ways:

Given that:

  • Microsoft appears to have no intention of merging the PR
  • The PR for the Transformers version is no longer compatible with vLLM on the latest Transformers (v5)

We have decided to drop support for the Transformers implementation in vLLM.

This means that:

  • Using Phi 4 multimodal in vLLM will always require --trust-remote-code
  • Phi 4 multimodal will eventually become incompatible with vLLM as its custom code rots
Signed-off-by: Harry Mellor <19981378+hmellor@users.noreply.github.com>
Signed-off-by: Harry Mellor <19981378+hmellor@users.noreply.github.com>
@mergify
Copy link

mergify bot commented Dec 4, 2025

@mergify mergify bot added documentation Improvements or additions to documentation multi-modality Related to multi-modality (#4194) new-model Requests to new models labels Dec 4, 2025
Copy link
Contributor

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request removes support for the Hugging Face Transformers version of the Phi 4 multimodal model (Phi4MultimodalForCausalLM). The changes are consistent and well-executed, covering the model implementation, its registration in various registries, associated tests, and documentation. The removal is clean and complete within the scope of the provided files, with no apparent dangling references. Overall, this is a solid cleanup effort.

Copy link
Member

@DarkLight1337 DarkLight1337 left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

@DarkLight1337 DarkLight1337 enabled auto-merge (squash) December 4, 2025 10:37
@hmellor
Copy link
Member Author

hmellor commented Dec 4, 2025

Sorry @Isotr0py, I know you put a lot of work into this implementation 😢

@DarkLight1337 DarkLight1337 added the ready ONLY add when PR is ready to merge/full CI is needed label Dec 4, 2025
Copy link
Member

@Isotr0py Isotr0py left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

BTW, I wonder if we can create a separated community repo like mistral-community/pixtral-12b/deepseek-community/deepseek-vl-7b-chat to hold HF format model instead of converting existing repo?

Signed-off-by: Harry Mellor <19981378+hmellor@users.noreply.github.com>
@hmellor
Copy link
Member Author

hmellor commented Dec 4, 2025

Good suggestion, we'll make a community repo so that at least the Transformers implementation is usable in Transformers without needing to specify a ref.

@DarkLight1337 DarkLight1337 merged commit 9998ea5 into vllm-project:main Dec 4, 2025
53 checks passed
@hmellor hmellor deleted the drop-hf-phi4 branch December 4, 2025 14:07
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

documentation Improvements or additions to documentation multi-modality Related to multi-modality (#4194) new-model Requests to new models ready ONLY add when PR is ready to merge/full CI is needed

3 participants