Skip to content

Remove no exp usage from logical rule 1/N#3578

Draft
NuojCheng wants to merge 2 commits intomainfrom
chengnuojin-no-exp
Draft

Remove no exp usage from logical rule 1/N#3578
NuojCheng wants to merge 2 commits intomainfrom
chengnuojin-no-exp

Conversation

@NuojCheng
Copy link
Copy Markdown
Collaborator

@NuojCheng NuojCheng commented Apr 6, 2026

Description

This PR deprecates

  • activation_batch_no_exp
  • activation_length_no_exp

from logical names.

After this change

  • activation_batch always includes "expert" physical axis
  • activation_batch_moe does not include "expert"
  • activation_length does not include "expert"

Other logical names containing "_no_exp" will be deprecated in following PR.

Tests

Please describe how you tested this change, and include any instructions and/or
commands to reproduce.

Checklist

Before submitting this PR, please make sure (put X in square brackets):

  • I have performed a self-review of my code. For an optional AI review, add the gemini-review label.
  • I have necessary comments in my code, particularly in hard-to-understand areas.
  • I have run end-to-end tests tests and provided workload links above if applicable.
  • I have made or will make corresponding changes to the doc if needed, including adding new documentation pages to the relevant Table of Contents (toctree directive) as explained in our documentation.

@codecov
Copy link
Copy Markdown

codecov bot commented Apr 6, 2026

Codecov Report

❌ Patch coverage is 90.90909% with 2 lines in your changes missing coverage. Please review.

Files with missing lines Patch % Lines
src/maxtext/layers/decoders.py 66.66% 1 Missing ⚠️
src/maxtext/layers/nnx_decoders.py 50.00% 1 Missing ⚠️

📢 Thoughts on this report? Let us know!

@NuojCheng NuojCheng changed the title Remove no exp usage from logical rule Remove no exp usage from logical rule 1/N Apr 6, 2026
@RissyRan
Copy link
Copy Markdown
Collaborator

RissyRan commented Apr 6, 2026

Could you help check if this breaks inference workflow? Potentially this has been deprecated due to vLLM migration.

@NuojCheng
Copy link
Copy Markdown
Collaborator Author

Could you help check if this breaks inference workflow? Potentially this has been deprecated due to vLLM migration.

AFAIK inference.yml is deprecated. I will check inference path in https://github.com/AI-Hypercomputer/maxtext/blob/main/src/maxtext/configs/inference/vllm.yml.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants