this post was submitted on 20 Feb 2024
1248 points (98.4% liked)
Programmer Humor
19572 readers
1696 users here now
Welcome to Programmer Humor!
This is a place where you can post jokes, memes, humor, etc. related to programming!
For sharing awful code theres also Programming Horror.
Rules
- Keep content in english
- No advertisements
- Posts must be related to programming or programmer topics
founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
The next generation of script kiddies is going to be iPad babies. It’ll be interesting to see, since the majority can’t use anything in tech unless it’s an app.
We built computer labs in schools, to teach kids how to use computers. Then we decided computers are ubiquitous enough that we didn’t need computer labs anymore. And now we have an entire generation that doesn’t know how to use computers, because they use their phones and tablets for everything instead.
I wonder who is going to write the apps in the future.
Chatgpt, of course..
Ugh. You’re probably right. Finally all those idiots who come up to me going “I’ve got a great idea for an app” will actually be able to release their great idea :)
I used to be able to say “ideas are easy, work is hard”. Now we won’t be.
I'm yet to hear anyone saying that chatGPT can navigate the complex series of design decisions needed to create a cohesive app (unless of course, it was trained on something exactly the same). Many people report spending an inordinate amount of time rectifying the mistakes these LLMs make. It sounds like a glorified autofill (I haven't used them yet). I shudder to think about the future of the software ecosystem if an entire generation is trained to rely entirely on them to create code.
LLM is great for writing code in small snippets. I’ve used it for quickly writing batch files, for instance. I couldn’t be bothered to look up how to format something obscure. So I use an LLM like ChatGPT to do the bulk work, then I just double check what it gave me.
I wouldn’t use it for anything over ~100 lines at a time. Just like with long conversations, it will have a tendency to “lose the plot” and start forgetting things that it said early on. Because as things get added to the conversation it has to parse more and more data. So it’ll start to drift off topic as conversations get longer.
It can also be handy for debugging sections of code. Because programming is just a form of language with strict grammar/diction/spelling rules. And a LLM will be really really good at spotting stupid grammar mistakes. It’ll instantly notice your missing semicolon and point it out to you, which can save you a ton of frustration.
Just like with any tool, how well it works is entirely up to the user. It will likely progress to the point of being able to manage longer code eventually. But right now it’s still incredibly useful as long as you accept its limitations and work within them.
I think you’re right at the minute. Whether you’ll be right in the future I’m less certain.
and they're going to be precisely as nonsensical as those AI articles are
sure, you can get good output from LLMs, but companies are absolutely not going to bother putting in the effort to do so, as not putting in effort is the entire point.
it's at least nice to know that corporations will enshittify themselves out of existence, while one guy living in a basement will silently release something they poured their soul into and it sells 5 billion copies in the hour