Thread 16711932 - /sci/ [Archived: 613 hours ago]

Anonymous
7/1/2025, 5:23:36 AM No.16711932
1234
1234
md5: 2790d8068dc944955001036b0e2fc8b6🔍
How exactly do you code LLMs?
Do you just keep a set of higher-level instructions and roll that shit back to an earlier version if it ever goes off the rails?
Replies: >>16711937 >>16711940 >>16711958 >>16712215
Anonymous
7/1/2025, 5:36:22 AM No.16711937
>>16711932 (OP)
Math
>That's not very specific
Words,
as vectors.
>That's really not specific
Weights
>That's reallly realllly not specific
Math
>Reeeeeeeeeee!!!!!
Anonymous
7/1/2025, 5:37:59 AM No.16711940
>>16711932 (OP)
You don't code an LLM, you code a transformer, then train it on huge corpora of language to make an LLM
Replies: >>16711941
Anonymous
7/1/2025, 5:39:15 AM No.16711941
>>16711940
>you code a transformer
Is that a "set of higher-level instructions"?
Is that the .exe?
Anonymous
7/1/2025, 5:39:48 AM No.16711942
just crunch numbers on the next most probable word
Anonymous
7/1/2025, 6:04:23 AM No.16711958
>>16711932 (OP)
It's basically just a recurrent neural network with words as input and the transformer thing is just some bullshit that allows the neural network to "pay attention" to certain words more than others. A neural network is just advanced curve fitting with billions of variables instead of a linear model that only has 1 variable. You could use a linear model to predict the next word but it would be shit.
Anonymous
7/1/2025, 11:38:07 AM No.16712215
>>16711932 (OP)
>How exactly do you code LLMs?
You download python scripts from github and dick around with them.