Will a LLM trained with FP4 have competitive performance in 2 years time?
Plus
15
Ṁ1421Jan 21
25%
chance
1D
1W
1M
ALL
"Currently, the technology for 4-bit training does not exists, but research looks promising and I expect the first high performance FP4 Large Language Model (LLM) with competitive predictive performance to be trained in 1-2 years time." (see: https://timdettmers.com/2023/01/16/which-gpu-for-deep-learning/)
Granted, the model must be open source for us to know, so the market will resolve based on publicly available information.
This question is managed and resolved by Manifold.
Get
1,000
and3.00
Sort by:
This seems important @typedfemale
Will this resolve YES if scaling laws suggest a 4-bit model would be competitive if compute-matched to a SOTA 16-bit model?
@NoaNabeshima Yes, you need to be better than everything else, but be trained in 4-bit (to some extent)
Related questions
Related questions
Will an LLM break 1400 ELO on LMSys before February?
35% chance
Will an LLM improve its own ability along some important metric well beyond the best trained LLMs before 2026?
58% chance
Will a publicly-available LLM achieve gold on IMO before 2026?
54% chance
Will LLMs mostly overcome the Reversal Curse by the end of 2025?
67% chance
LLM Hallucination: Will an LLM score >90% on SimpleQA before 2026?
60% chance
6 months from now will I judge that LLMs had already peaked by Nov 2024?
16% chance
Will an LLM be able to solve the Self-Referential Aptitude Test before 2027?
66% chance
Will Apple release its own LLM on par with state of the art LLMs before 2026?
49% chance
Will LLM training costs fall 300x by 2028?
85% chance
Will one of the major LLMs be capable of continual lifelong learning (learning from inference runs) by EOY 2025?
50% chance