Add support for StableLM-2-12B on GPUs
aahouzi opened this issue · comments
Type of Change
Stability.ai open sourced StableLM-2-12B, which has a different architecture than its 1.6B & 3B counterparts. This issue is to ask for adding support for these models: stabilityai/stablelm-2-12b & stabilityai/stablelm-2-12b-chat to IPEX-LLM.
Description
- Uses GQA instead of MHA + Parallel MLP layer + per-head qk_normalization
- Model description: StableLM-2-12B