Skip to content

[Example] Add RoPE backward kernel with mask-based implementation#696

Open
liujaaa wants to merge 1 commit intotile-ai:ascendc_ptofrom
liujaaa:ascendc_pto
Open

[Example] Add RoPE backward kernel with mask-based implementation#696
liujaaa wants to merge 1 commit intotile-ai:ascendc_ptofrom
liujaaa:ascendc_pto

Conversation

@liujaaa
Copy link
Copy Markdown

@liujaaa liujaaa commented Mar 27, 2026

  • Implements RoPE forward and backward kernels
  • Supports TND [BS, N, D] and BSND [B, S, N, D] input formats
  • Supports float16, bfloat16, float dtypes with FP32 accumulation
  • Host-side pre-computed mask for gather operation
  • Autograd Function wrapper for PyTorch integration

@github-actions
Copy link
Copy Markdown

👋 Hi! Thank you for contributing to the TileLang project.

Please remember to run bash format.sh in the root directory of the project to ensure your changes are properly linted and formatted. This will help ensure your contribution passes the format check.

We appreciate you taking this step! Our team will review your contribution, and we look forward to your awesome work!

🚀

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant