Skip to content

feat: add GLM-4.7-Flash (glm4_moe_lite) model support#104

Open
Realmhang wants to merge 1 commit intoflagos-ai:mainfrom
Realmhang:feat/glm47_flash
Open

feat: add GLM-4.7-Flash (glm4_moe_lite) model support#104
Realmhang wants to merge 1 commit intoflagos-ai:mainfrom
Realmhang:feat/glm47_flash

Conversation

@Realmhang
Copy link
Copy Markdown

PR Category

Core

PR Type

New Features

Description

  • Add support for GLM-4.7-Flash (glm4_moe_lite) model, which combines MLA (Multi-head Latent Attention) from DeepSeek V2/V3 with MoE
    architecture from GLM-4 MoE
  • Add config bridge Glm4MoeLiteConfig extending Glm4MoeConfig with MLA fields, DSA Indexer fields, and MTP support
  • Add model implementation (Glm4MoeLiteForCausalLM) inheriting from both glm4_moe and deepseek_v2 components
  • Register model and config in plugin entry point

Changes

@github-actions github-actions bot added the core label Mar 24, 2026
@CLAassistant
Copy link
Copy Markdown

CLAassistant commented Mar 24, 2026

CLA assistant check
All committers have signed the CLA.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants