(논문 요약) BitNet
BitNet: Scaling 1-bit Transformers for Large Language Models The Era of 1-bit LLMs: All Large Language Models are in 1.58 Bits (repo)
핵심 내용
- {-1,0,1} weight matrix 로 muliplication 가속화.
- Weight quantization formula
(논문 요약) BitNet
BitNet: Scaling 1-bit Transformers for Large Language Models The Era of 1-bit LLMs: All Large Language Models are in 1.58 Bits (repo)