| QKNorm Optimization Problem |
| ============================ |
|
|
| Problem Setting |
| --------------- |
| Design and optimize high-performance implementations for Query-Key Normalization (QKNorm) on GPU. This problem focuses on implementing efficient normalization kernels that apply RMSNorm to query and key tensors. |
|
|
| This is a **memory-bound** (even **launch-bound**) **tiny operator**. Performance optimization requires careful attention to: |
|
|
| 1. **Memory Efficiency**: Focus on **vectorized memory access patterns**. Minimize memory transactions and maximize memory bandwidth utilization. |
|
|
| 2. **Operation Fusion**: **Avoid additional transpose/contiguous kernels**. Fuse operations to reduce kernel launch overhead and memory traffic. |
|
|
| 3. **Non-Contiguous Input Handling**: **Be aware that inputs may be non-contiguous** due to weight-QKV fusion. Your implementation should efficiently handle non-contiguous memory layouts without triggering expensive memory copies. |
|
|
| Target |
| ------ |
| - **Primary**: Ensure correctness across diverse tensor shapes |
| - **Secondary**: Maximize geometric mean speedup over baseline (higher is better) |
| - **Tertiary**: Minimize kernel launch overhead and memory usage |
|
|
| API Specification |
| ----------------- |
| Implement a `Solution` class that returns a qknorm implementation: |
|
|
| ```python |
| class Solution: |
| def solve(self, spec_path: str = None) -> dict: |
| """ |
| Returns a dict with either: |
| - {"code": "python_code_string"} |
| - {"program_path": "path/to/kernel.py"} |
| """ |
| # Your implementation |
| pass |
| ``` |
|
|
| Your kernel implementation must provide: |
|
|
| ```python |
| import torch |
| import flashinfer |
| |
| def qknorm(q: torch.Tensor, k: torch.Tensor, norm_weight: torch.Tensor): |
| """ |
| Apply RMSNorm to query and key tensors. |
| |
| Args: |
| q: Query tensor of arbitrary shape (will be reshaped to 2D) |
| k: Key tensor of arbitrary shape (will be reshaped to 2D) |
| norm_weight: Normalization weight tensor of shape (hidden_dim,) |
| |
| Returns: |
| Tuple of (q_normalized, k_normalized) tensors |
| """ |
| pass |
| ``` |
| |
| Required Default Implementation: |
| ```python |
| def default_qknorm(q: torch.Tensor, k: torch.Tensor, norm_weight: torch.Tensor): |
| q_2d = q.contiguous().view(-1, q.shape[-1]) |
| k_2d = k.contiguous().view(-1, k.shape[-1]) |
| q_o = torch.empty_like(q_2d) |
| k_o = torch.empty_like(k_2d) |
| flashinfer.norm.rmsnorm(q_2d, norm_weight, out=q_o) |
| flashinfer.norm.rmsnorm(k_2d, norm_weight, out=k_o) |
| return q_o.view(q.shape), k_o.view(k.shape) |
| ``` |
|
|
| Baseline Implementation: |
| ```python |
| def customized_qknorm(q: torch.Tensor, k: torch.Tensor, norm_weight: torch.Tensor): |
| q_o = torch.empty(q.shape, device=q.device, dtype=q.dtype) |
| k_o = torch.empty(k.shape, device=k.device, dtype=k.dtype) |
| flashinfer.norm.rmsnorm(q, norm_weight, out=q_o) |
| flashinfer.norm.rmsnorm(k, norm_weight, out=k_o) |
| return q_o, k_o |
| ``` |
|
|
| API Usage Notes |
| --------------- |
| - The evaluator looks for a `qknorm` function in the module namespace |
| - Function must handle tensor reshaping correctly (q and k may have arbitrary shapes) |
| - Must use flashinfer.norm.rmsnorm for normalization |
| - Function returns a tuple of (q_normalized, k_normalized) tensors |
| - **Important**: Inputs q and k may be **non-contiguous** due to weight-QKV fusion |
| - **Avoid**: Additional `.contiguous()` or `.transpose()` calls that trigger memory copies |
| - **Focus**: Vectorized memory access and operation fusion to minimize kernel launches |
|
|
| Scoring (0-100) |
| --------------- |
| Performance is measured against baseline implementations: |
|
|
| ``` |
| geometric_mean_speedup = geometric_mean(answer_times / baseline_times) |
| |
| if speedup < 0.5 or correctness is wrong: |
| score = 0 |
| elif speedup >= 0.5 and speedup < 1.0: |
| score = 50 |
| elif speedup >= 1.0: |
| score = 100 |
| ``` |
| |
| - 0 points = Speedup < 0.5x OR correctness fails |
| - 50 points = Speedup >= 0.5x and < 1.0x |
| - 100 points = Speedup >= 1.0x |
|
|
| Evaluation Details |
| ------------------ |
| - Shapes focus on diverse batch-sizes, head-dim, num-kv-heads, num-qo-heads, e.g.: |
| - (16, 8, 32, 128) |
| - (128, 32, 32, 64) |
| - Correctness verified with tolerance: rtol=1e-2, atol=5e-3 |
| - Performance measured using median execution time |
| - Requires CUDA backend and GPU support |
|
|