Skip to content

Conversation

quic-mamta
Copy link
Contributor

@quic-mamta quic-mamta commented Aug 19, 2025

Update Transformers to 4.55.0
Update PyTorch to 2.7.0+cpu
Torchvision to 0.22.0+cpu
and Python Requirement to >=3.9

Updated modeling files and Cache Utils for transformers 4.55.0

Updated models :

  1. codegen
  2. falcon
  3. gemma
  4. gemma2
  5. gptj
  6. gpt2
  7. granite
  8. granite_moe
  9. grok1
  10. llama
  11. llama_swiftkv
  12. mistral
  13. mixtral_moe
  14. mpt
  15. phi
  16. phi3
  17. qwen2
  18. starcoder2
  19. gpt_bigcode
  20. internvl
  21. llava
  22. llava_next
  23. whisper
  24. gemma3
  25. llama4
  26. mllama

@quic-mamta quic-mamta changed the title Tf version 4.55 upgrade Transformers version 4.55 upgrade Aug 19, 2025
@quic-mamta quic-mamta marked this pull request as draft August 19, 2025 19:42
@asmigosw asmigosw force-pushed the TF_version_4.55_upgrade branch from d36c124 to a514d36 Compare September 2, 2025 08:29
@quic-mamta quic-mamta force-pushed the TF_version_4.55_upgrade branch 2 times, most recently from e15d548 to 3643fee Compare September 23, 2025 08:45
@quic-mamta quic-mamta marked this pull request as ready for review September 24, 2025 05:27
@quic-mamta quic-mamta requested a review from vbaddi September 24, 2025 05:27
@quic-mamta quic-mamta marked this pull request as draft September 24, 2025 19:31
@quic-mamta quic-mamta force-pushed the TF_version_4.55_upgrade branch from 69ec2a4 to 6ad267b Compare September 25, 2025 07:46
@quic-mamta quic-mamta changed the title Transformers version 4.55 upgrade Transformers version 4.55 upgrade, Update PyTorch to 2.7.0+cpu, Torchvision to 0.22.0+cpu, and Python Requirement to >=3.9 Sep 25, 2025
@quic-mamta quic-mamta force-pushed the TF_version_4.55_upgrade branch 3 times, most recently from dd8b38e to 940dfcf Compare September 26, 2025 11:44
@quic-mamta quic-mamta marked this pull request as ready for review September 26, 2025 11:44
@quic-mamta quic-mamta force-pushed the TF_version_4.55_upgrade branch 2 times, most recently from 4f44dd4 to d8cf0a1 Compare September 28, 2025 13:10
# Apply the attention mask
attn_weights = torch.where(attention_mask, mask_value, attn_weights)

attn_weights = attn_weights / self.scale_attn
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Why it has been moved from line 51?

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

It was made equivalent to new TF code; they have moved it down since its placement whether at line 50 or 58, won't affect the performance. Should I move it back to 50?


EXTERNAL_MODELS = {
"hpcai-tech/grok-1",
"hpcai-tech/grok-1": {
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

nit: Do we need this?

@quic-mamta quic-mamta force-pushed the TF_version_4.55_upgrade branch 3 times, most recently from 8217cb5 to 7bf2298 Compare October 8, 2025 10:49
mamtsing and others added 3 commits October 10, 2025 08:30
Signed-off-by: Asmita Goswami <[email protected]>
Signed-off-by: Asmita Goswami <[email protected]>
Signed-off-by: Asmita Goswami <[email protected]>
asmigosw and others added 3 commits October 10, 2025 08:31
Signed-off-by: Asmita Goswami <[email protected]>
Signed-off-by: Asmita Goswami <[email protected]>
Signed-off-by: Mamta Singh <[email protected]>
@asmigosw asmigosw force-pushed the TF_version_4.55_upgrade branch from c9fd8ea to 178650f Compare October 10, 2025 08:32
Copy link
Contributor

@quic-hemagnih quic-hemagnih left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM

Signed-off-by: Asmita Goswami <[email protected]>
Comment on lines -555 to -558
# use local attention mask for ROPE layers
if self.use_chunked_attention:
attention_mask = chunk_causal_mask

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Why is this removed?

Comment on lines 722 to 724
logger.warning(
"Current version output doesn't match with HF output due to a bug in TF v_4.55. Switch to branch release/v_1.20 for TF match."
)
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Signed-off-by: Asmita Goswami <[email protected]>
Signed-off-by: Asmita Goswami <[email protected]>
Copy link
Contributor

@vbaddi vbaddi left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM, thanks 👍

@quic-rishinr quic-rishinr merged commit a9e404a into quic:main Oct 14, 2025
5 checks passed
quic-vargupt pushed a commit to quic-vargupt/efficient-transformers that referenced this pull request Oct 17, 2025
…vision to 0.22.0+cpu, and Python Requirement to >=3.9 (quic#542)

Update Transformers to 4.55.0
Update PyTorch to 2.7.0+cpu
Torchvision to 0.22.0+cpu
and Python Requirement to >=3.9

Updated modeling files and Cache Utils for transformers 4.55.0

Updated models :

1. codegen
2. falcon
3. gemma
4. gemma2
5. gptj
6. gpt2
7. granite
8. granite_moe
9. grok1
10. llama
11. llama_swiftkv
12. mistral
13. mixtral_moe
14. mpt
15. phi
16. phi3
17. qwen2
18. starcoder2
19. gpt_bigcode
20. internvl
21. llava
22. llava_next
23. whisper
24. gemma3
25. llama4
26. mllama

---------
Update Qeff Documentation to indicate vLLM Support in Validated Models Page

Signed-off-by: Asmita Goswami <[email protected]>
Signed-off-by: Mamta Singh <[email protected]>
Co-authored-by: Mamta Singh <[email protected]>
Co-authored-by: Asmita Goswami <[email protected]>
Signed-off-by: Varun Gupta <[email protected]>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

7 participants