Pinned Repositories
AutoAWQ
Adds Interlm2 support to AutoAWQ
axolotl
Go ahead and axolotl questions
BTX
laserRMT
This is our own implementation of 'Layer Selective Rank Reduction'
LCKV
Layer-Condensed KV cache w/ 10 times larger batch size, fewer params and less computation. Dramatic speed up with better task performance. Accepted to ACL 2024.
MCT-Self-Refine-MCTSr-is-openai
A simple program of the MCT Self-Refine (MCTSr) algorithm, welcome to adapt or modify it
mergekit
Tools for merging pretrained large language models.
Mixtral-Deepseek
quiet-star-working
Code for Quiet-STaR
transformers
🤗 Transformers: Add GemMoE support
Crystalcareai's Repositories
Crystalcareai/BTX
Crystalcareai/mergekit
Tools for merging pretrained large language models.
Crystalcareai/transformers
🤗 Transformers: Add GemMoE support
Crystalcareai/quiet-star-working
Code for Quiet-STaR
Crystalcareai/AutoAWQ
Adds Interlm2 support to AutoAWQ
Crystalcareai/axolotl
Go ahead and axolotl questions
Crystalcareai/laserRMT
This is our own implementation of 'Layer Selective Rank Reduction'
Crystalcareai/LCKV
Layer-Condensed KV cache w/ 10 times larger batch size, fewer params and less computation. Dramatic speed up with better task performance. Accepted to ACL 2024.
Crystalcareai/MCT-Self-Refine-MCTSr-is-openai
A simple program of the MCT Self-Refine (MCTSr) algorithm, welcome to adapt or modify it
Crystalcareai/Mixtral-Deepseek
Crystalcareai/MoE-Any
Crystalcareai/quiet-star
Implementation of the Quiet-STAR paper (https://arxiv.org/pdf/2403.09629.pdf)
Crystalcareai/Quiet_STaR
This project aims to implements quiet_star algoithm
Crystalcareai/qwen2_moe_mergekit
根据Qwen2(Qwen1.5)模型生成qwen2 MoE模型的工具
Crystalcareai/RemoteYixtral
Crystalcareai/SystemChat