« BackFine-Tuning LLMs to 1.58bithuggingface.coSubmitted by galeos 10 months ago
  • patleeman 10 months ago

    That's awesome. The original discussion of bitnet made it seem like you needed to train a model from scratch but its neat they were able to adapt an existing model. This is quite exciting.

    • cpldcpu 10 months ago

      the performance is still a bit degraded though.

    • amilios 10 months ago

      Very exciting, although it was a bit disappointing to see that they're hitting just llama1 7b performance by quantizing llama3. but i'm sure the performance gap will close over time!