ICLR 2025 Workshop on Sparsity in LLMs (SLLM)

Deep Dive into Mixture of Experts, Quantization, Hardware, and Inference