(논문 요약) BitNet

BitNet: Scaling 1-bit Transformers for Large Language Models The Era of 1-bit LLMs: All Large Language Models are in 1.58 Bits (repo)

핵심 내용

  • {-1,0,1} weight matrix 로 muliplication 가속화.
  • Weight quantization formula

실험 결과