>>106181775 (OP)all of the innovations for better scaling are here, waiting to be implemented:
>1-bit quantization https://arxiv.org/abs/2310.11453v1>Kolmogorov-Arnold Networks https://arxiv.org/abs/2404.19756>diffusion llms https://arxiv.org/abs/2502.09992>ASI-ARCH/autonomous architecture discovery https://github.com/GAIR-NLP/ASI-Arch?tab=readme-ov-file>jepa https://arxiv.org/abs/2301.08243>sparsity https://arxiv.org/abs/2412.12178>synthetic data https://arxiv.org/abs/2503.14023>transformers squared https://arxiv.org/abs/2501.06252>titans https://arxiv.org/abs/2501.00663>context engineering https://arxiv.org/abs/2501.00663>MLE-STAR https://arxiv.org/abs/2506.15692>hierachical reasoning model https://arxiv.org/abs/2506.21734>graphRAG https://github.com/LHRLAB/Graph-R1The issue is that OpenAI has mistakenly positioned themselves as the new-big-model-every-year company (like they're releasing smartphones or something), so they're afraid to take risks on training (*and releasing*) smaller research models.
>>106181894maybe this is true for normal people and for like, 1-2 years, but we already have models that are undeniably useful. remember that google became a multi-trillion dollar company from a simple algorithm for ranking web pages. even the current models, with their hallucinations and obvious limitations, are basically like a search engine on steroids (and search engines also fucking suck nowadays).
i see massive divestment if openai continues to over-promise and under-deliver, but machine learning will be as relevant as computing in general forever now. it has already become a basically unavoidable part of life.