Skip to content

Commit cfcf20e

Browse files
committed
Fix some Transformers nightly tests
Signed-off-by: Harry Mellor <[email protected]>
1 parent 37593de commit cfcf20e

File tree

3 files changed

+29
-28
lines changed

3 files changed

+29
-28
lines changed

vllm/model_executor/models/jina_vl.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -29,7 +29,7 @@
2929
class JinaVLScorer(nn.Module):
3030
def __init__(self, model_config: "ModelConfig"):
3131
super().__init__()
32-
config = model_config.hf_config
32+
config = model_config.hf_config.get_text_config()
3333
head_dtype = model_config.head_dtype
3434
self.dense = ColumnParallelLinear(
3535
config.hidden_size, config.hidden_size, params_dtype=head_dtype, bias=True

vllm/model_executor/models/modernbert.py

Lines changed: 27 additions & 26 deletions
Original file line numberDiff line numberDiff line change
@@ -20,7 +20,7 @@
2020
PoolingParamsUpdate,
2121
PoolingType,
2222
)
23-
from vllm.model_executor.layers.rotary_embedding import RotaryEmbedding
23+
from vllm.model_executor.layers.rotary_embedding import get_rope
2424
from vllm.model_executor.layers.vocab_parallel_embedding import VocabParallelEmbedding
2525
from vllm.model_executor.model_loader.weight_utils import default_weight_loader
2626
from vllm.sequence import IntermediateTensors
@@ -62,19 +62,6 @@ def forward(
6262
return embeddings
6363

6464

65-
class ModernBertRotaryEmbedding(RotaryEmbedding):
66-
def __init__(self, config: ModernBertConfig, head_size: int, dim: int, base: float):
67-
super().__init__(
68-
head_size=head_size,
69-
rotary_dim=dim,
70-
max_position_embeddings=config.max_position_embeddings,
71-
base=base,
72-
is_neox_style=True,
73-
dtype=torch.float16,
74-
)
75-
self.config = config
76-
77-
7865
class ModernBertAttention(nn.Module):
7966
def __init__(self, config: ModernBertConfig, layer_id: int | None = None):
8067
super().__init__()
@@ -95,19 +82,33 @@ def __init__(self, config: ModernBertConfig, layer_id: int | None = None):
9582
bias=config.attention_bias,
9683
)
9784

98-
sliding_window = None
99-
if layer_id % config.global_attn_every_n_layers != 0:
100-
sliding_window = config.local_attention // 2
101-
rope_theta = (
102-
config.local_rope_theta
103-
if config.local_rope_theta is not None
104-
else config.global_rope_theta
105-
)
85+
if layer_types := getattr(config, "layer_types", None):
86+
# Transformers v5
87+
layer_type = layer_types[layer_id]
88+
rope_parameters = config.rope_parameters[layer_type]
89+
sliding_window: int | None = None
90+
if layer_type == "sliding_attention":
91+
sliding_window = config.local_attention // 2
10692
else:
107-
rope_theta = config.global_rope_theta
108-
109-
self.rotary_emb = ModernBertRotaryEmbedding(
110-
config=config, head_size=self.head_dim, dim=self.head_dim, base=rope_theta
93+
# Transformers v4
94+
sliding_window = None
95+
if layer_id % config.global_attn_every_n_layers != 0:
96+
sliding_window = config.local_attention // 2
97+
rope_theta = (
98+
config.local_rope_theta
99+
if config.local_rope_theta is not None
100+
else config.global_rope_theta
101+
)
102+
else:
103+
rope_theta = config.global_rope_theta
104+
rope_parameters = {"rope_type": "default", "rope_theta": rope_theta}
105+
106+
self.rotary_emb = get_rope(
107+
head_size=self.head_dim,
108+
rotary_dim=self.head_dim,
109+
max_position=config.max_position_embeddings,
110+
rope_parameters=rope_parameters,
111+
dtype=torch.float16,
111112
)
112113
self.attn = EncoderOnlyAttention(
113114
self.num_heads,

vllm/model_executor/models/qwen2.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -503,7 +503,7 @@ class Qwen2ForCausalLM(nn.Module, SupportsLoRA, SupportsPP, SupportsEagle3):
503503

504504
def __init__(self, *, vllm_config: VllmConfig, prefix: str = ""):
505505
super().__init__()
506-
config = vllm_config.model_config.hf_config
506+
config = vllm_config.model_config.hf_config.get_text_config()
507507
quant_config = vllm_config.quant_config
508508

509509
self.config = config

0 commit comments

Comments
 (0)