Remember how every scary AI in scifi stories eventually starts improving itself? Yeah that shit aint happening. A small inaccuracy now will only snowball into barely functional model in the future.
When you finally inherit that legacy codebase from a developer who didn't believe in comments.
// Increment the variable named i
i++; // Use semi colon to end the statement
Here have some comments.
If the company wants me to believe in comments, they should pay me and not threaten layoffs in such ways that make me not feel I need to protect my job
LLMs learning from insightful new data such as
"You're absolutely right!" and "Great point!"
I hope we become one of those programmers that programmed *before* Stack Overflow :)
I know it won't happen, though.
That's why they try to block LLM responses, it pre-cleans and humanizes the data so that they can sell it to third parties for AI training. Cha-ching!!!
The problem dtarts when LLMs use LLM data for learning
To be fair recent LLM perf improvements have been in large part due to synthetic data generation and data curation. A sign we're progressing in architecture should be the lack of necessity of new data (AlphaGo->AlphaZero). Doesn't make this any less true as a whole though.
How does synthetic data generation work? How is it possible that the output from model A can be used to train a model B so that it is better than A?
Human filters
Do you mean that humans choose which outputs go into the training pile? Is that basically like some sort of reinforcement learning then?
Or do the humans edit the generated outputs to make them better and then add them to the pile? That way it's basically human output
More reasoning-like data where it expands on earlier data. Re-mix and replay. Humans do this as well via imagination e.g. when you learn to ski you're taught to visualize the turn before doing it, or e.g. kids roleplaying all kinds of jobs to gain training data for tasks they can't do as often in real life.
missed stackoverflow circle so much even the guys downvoting everything
The onus will be on language/library/framework authors to provide good documentation that AI can understand.
Well, wouldn't this mean we would start having to use stack overflow again? (Or maybe even llms asking each other questions, dead stack overflow theory).
Do they ever response with "marked as duplicate, closed"?
https://www.reddit.com/r/ProgrammerHumor/s/OMS8vGbg6z
This meme clarifies it
The only time I asked a question on stackoverflow I got downvoted and they shamed me in comments that my questions "aren't clear".
Funny thing is it was about a poorly documented Azure service (at that time) and I was contacted by the team and they clearly understood my issue and they even added some new documentation based on my questions. It all went via e-mails.
Yet, I was downvoted on SO.
So after that, I always went straight to Azure support and it was much faster and convenient than being downvoted and shamed in comments for no real reason.
StackOverflow is so incredibly pedantic about things that didn't matter it just became useless. Questions constantly marked as duplicates even if they required different answers
Gotta get those points!