Almost as if it was made to simulate human output but without the ability to scrutinize itself.
160
mushroommunk @lemmy.today - 1day
To be fair most humans don't scrutinize themselves either.
(Fuck AI though. Planet burning trash)
79
atomicbocks @sh.itjust.works - 1day
The number of times I have received an un-proofread two sentence email is too damn high.
24
galaxy_nova @lemmy.world - 1day
And then the follow up email because they didn’t actually finish a complete thought
10
Sophienomenal - 15hr
I do this with texts/DMs, but I'd never do that with an email. I double or triple check everything, make sure my formatting is good, and that the email itself is complete. I'll DM someone 4 or 5 times in 30 seconds though, it feels like a completely different medium ¯\_(ツ)_/¯
4
PetteriPano @lemmy.world - 1day
It's like having a lightning-fast junior developer at your disposal. If you're vague, he'll go on shitty side-quests. If you overspecify he'll get overwhelmed. You need to break down tasks into manageable chunks. You'll need to ask follow-up questions about every corner case.
A real junior developer will have improved a lot in a year. Your AI agent won't have improved.
115
mcv @lemmy.zip - 20hr
This is the real thing. You can absolutely get good code out of AI, but it requires a lot of hand holding. It helps me speed some tasks, especially boring ones, but I don't see it ever replacing me. It makes far too many errors, and requires me to point them out, and to point in the direction of the solution.
They are great at churning out massive amounts of code. They're also great at completely missing the point. And the massive amount of code needs to be checked and reviewed. Personally I'd rather write the code and have the AI review it. That's a much more pleasant way to work, and that way it actually enhances quality.
25
Grimy @lemmy.world - 24hr
They are improving, and probably faster then junior devs. The models we had had 2 years ago would struggle with a simple black jack app. I don't think the ceiling has been hit.
10
lividweasel @lemmy.world - 23hr
Just a few trillion more dollars, bro. We’re almost there. Bro, if you give up a few showers, the AI datacenter will be able to work perfectly.
Bro.
61
Grimy @lemmy.world - 22hr
The cost of the improvement doesn't change the fact that it's happening. I guess we could all play pretend instead if it makes you feel better about it. Don't worry bro, the models are getting dumber!
12
underisk @lemmy.ml - 21hr
Don’t worry bro, the models are getting dumber!
That would be pretty impressive when they already lack any intelligence at all.
21
Eranziel @lemmy.world - 17hr
And I ask you - if those same trillions of dollars were instead spent on materially improving the lives of average people, how much more progress would we make as a society? This is an absolutely absurd sum of money were talking about here.
6
Grimy @lemmy.world - 17hr
It's beside the point. I'm simply saying that AI will improve in the next year. The cost to do so or all the others things that money could be spent on doesn't matter when it's clearly going to be spent on AI. I'm not in charge of monetary policies anywhere, I have no say in the matter. I'm just pushing back on the fantasies. I'm hoping the open source scene survives so we don't end up in some ugly dystopia where all AI is controlled by a handful of companies.
4
SabinStargem @lemmy.today - 2hr
I have the impression that anti-AI people don't understand that they are giving up agency for the sake of temporary feels. If they truly cared about ethical usage of AI, they would be wanting to have mastery that is at least equal to that of corporations and the 1%.
Making AI into a public good is key to a better future.
2
mcv @lemmy.zip - 20hr
They might. The amount of money they're pumping into this is absolutely staggering. I don't see how they're going to make all of that money back, unless they manage to replace nearly all employees.
Either way it's going to be a disaster: mass unemployment or the largest companies in the world collapsing.
6
SabinStargem @lemmy.today - 2hr
I dunno, the death of mega corporations would do the world a great deal of good. Healthier capitalism requires competition, and a handful of corporations of any given sector isn't going to seriously compete nor pay good wages.
1
mcv @lemmy.zip - 1hr
It's certainly the option I'm rooting for, but it would still be a massive drama and disrupt a lot of lives. Which is why they'll probably get bailed out with taxpayer money.
2
PetteriPano @lemmy.world - 16hr
My jr developer will eventually be familiar with the entire codebase and can make decisions with that in mind without me reminding them about details at every turn.
LLMs would need massive context windows and/or custom training to compete with that. I'm sure we'll get there eventually, but for now it seems far off. I think this bubble will have to burst and let hardware catch up with our ambitions. It'll take a couple of decades.
5
UnderpantsWeevil @lemmy.world - 1day
A computer is a machine that makes human errors at the speed of electricity.
61
MountingSuspicion @reddthat.com - 1day
I think one of the big issues is it often makes nonhuman errors. Sometimes I forget a semicolon or there's a typo, but I'm well equipped to handle that. In fact, most programs can actually catch that kind of issue already. AI is more likely to generate code that's hard to follow and therefore harder to check. It makes debugging more difficult.
28
UnderpantsWeevil @lemmy.world - 24hr
AI is more likely to generate code that’s hard to follow and therefore harder to check.
Sure. It's making the errors faster and at a far higher volume than any team of humans could do in twice the time. The technology behind inference is literally an iterative process of turning gibberish into something that resembles human text. So its sort of a speed run from baby babble into college level software design by trial, evaluation, and correction over and over and over again.
But because the baseline comparison code is, itself, full of errors, the estimation you get at the end of the process is going to be scattering errant semicolons (and far more esoteric coding errors) through the body of the program at a frequency equivalent to humans making similar errors over a much longer timeline.
4
5too @lemmy.world - 18hr
Also seems like it'd be a lot harder to modify or extend later
I actually believed somebody when they told me it was great at writing code, and asked it to write me the code for a very simple lua mod. It’s made several errors and ended up wasting my time because I had to rewrite it.
48
morto - 17hr
In a postgraduate class, everyone was praising ai, calling it nicknames and even their friend (yes, friend), and one day, the professor and a colleague were discussing some code when I approached, and they started their routine bullying on me for being dumb and not using ai. Then I looked at his code and asked to test his core algorithm that he converted from a fortran code and "enhanced" it. I ran it with some test data and compared to the original code and the result was different! They blindly trusted some ai code that deviated from their theoretical methodology, and are publishing papers with those results!
Even after showing the different result, they didn't convince themselves of anything and still bully me for not using ai. Seriously, this shit became some sort of cult at this point. People are becoming irrational. If people in other universities are behaving the same and publishing like this, I'm seriously concerned for the future of science and humanity itself. Maybe we should archive everything published up to 2022, to leave as a base for the survivors from our downfall.
18
MyMindIsLikeAnOcean - 14hr
The way it was described to me by some academics is that it’s useful…but only as a “research assistant” to bounce ideas off of and bring in arcane or tertiary concepts you might not have considered (after you vet them thoroughly, of course).
The danger, as described by the same academics, is that it can act as a “buddy” who confirms you biases. It can generate truly plausible bullshit to support deeply flawed hypotheses, for example. Their main concern is it “learning” to stroke the egos of the people using it so it creates a feedback loop and it’s own bubbles of bullshit.
4
tym @lemmy.world - 8hr
So, linkedin? What if the real artificial intelligence was the linkedin lunatics we met along the way?
1
Xenny @lemmy.world - 11hr
That's not a bad idea. I'm already downloading lots of human knowledge and media that I want backed up because I can't trust humanity anymore to have it available anymore
2
Deestan @lemmy.world - 1day
I've been coding for a while. I did an honest eager attempt at making a real functioning thing with all code written by AI. A breakout clone using SDL2 with music.
The game should look good, play good, have cool effects, and be balanced. It should have an attractor screen, scoring, a win state and a lose state.
I also required the code to be maintainable. Meaning I should be able to look at every single line and understand it enough to defend its existence.
I did make it work. And honestly Claude did better than expected. The game ran well and was fun.
But: The process was shit.
I spent 2 days and several hundred dollars to babysit the AI, to get something I could have done in 1 day including learning SDL2.
Everything that turned out well, turned out well because I brought years of skill to the table, and could see when Claude was coding itself into a corner and tell it to break up code in modules, collate globals, remove duplication, pull out abstractions, etc. I had to detect all that and instruct on how to fix it. Until I did it was adding and re-adding bugs because it had made so much shittily structured code it was confusing itself.
TLDR; LLM can write maintainable code if given full constant attention by a skilled coder, at 40% of the coder's speed.
39
thundermoose @lemmy.world - 1day
It depends on the subject area and your workflow. I am not an AI fanboy by any stretch of the imagination, but I have found the chatbot interface to be a better substitute for the "search for how to do X with library/language Y" loop. Even though it's wrong a lot, it gives me a better starting place faster than reading through years-old SO posts. Being able to talk to your search interface is great.
The agentic stuff is also really good when the subject is something that has been done a million times over. Most web UI areas are so well trodden that JS devs have already invented a thousand frameworks to do it. I'm not a UI dev, so being able to give the agent a prompt like, "make a configuration UI with a sidebar that uses the graphql API specified here" is quite nice.
AI is trash at anything it hasn't been trained on in my experience though. Do anything niche or domain-specific, and it feels like flipping a coin with a bash script. It just throws shit at the wall and runs tests until the tests pass (or it sneakily changes the tests because the error stacktrace repeatedly indicates the same test line as the problem).
17
Deestan @lemmy.world - 1day
Yeah what you say makes sense to me. Having it make a "wrong start" in something new is useful, as it gives you a lot of the typical structure, introduces the terminology, maybe something sorta moving that you can see working before messing with it, etc.
9
galaxy_nova @lemmy.world - 1day
It’s basically just for if you’re lazy and don’t want to write a bunch of boilerplate or hit your keyboard a bunch of times to move the cursor(s) around
7
mcv @lemmy.zip - 19hr
It is great for boilerplate code. It can also explain code for you, or help with an unfamiliar library. It's even helped me be productive when my brain wasn't ready to really engage with the code.
But here's the real danger: because I've got AI to do it for me, my brain doesn't have to engage fully with the code anymore. I don't really get into the flow where code just flows out of your hands like I used to. It's becoming a barrier between me and the real magic of coding. And that sucks, because that's what I love about this work. Instead, I'm becoming the AI's manager. I never asked for that.
7
galaxy_nova @lemmy.world - 16hr
I generally agree with what you’ve said for sure. I think I’ve honestly started to use it for helping me to go pinpoint where to go look for issues in the spaghetti code of new code bases. I’ve also mostly tried to avoid using it in my personal coding time but I feel like it’s gotten harder and harder to get legitimately good search results nowadays which I realize is also because of ai. Given the choice I’d happily just erase it from existence I think. Spending hours sifting through reddit and stack overflow was way more fulfilling + I feel like people used to be slightly less prickly about answering stuff because that was how you had to get answers. It seems like lemmy could replace that space at least, I’ve genuinely gotten helpful comments and I’ve always felt downvotes on here have been productive versus what Reddit is now.
2
Buckshot @programming.dev - 12hr
I've found the same thing. I've turned off the auto suggestions while tying because by the time I'm typing i already know what I'm going I'm to type and having mostly incorrect suggestions popping up every 2 seconds was distracting and counterproductive.
1
Deestan @lemmy.world - 1day
This was a very directed experiment at purely LLM written maintainable code.
Writing experiments and proof of concepts, even without skill, will give a different calculation and can make more sense.
Having it write a "starting point" and then take over, also is a different thing that can make more sense. This requires a coder with skill, you can't skip that.
5
justaman123 @lemmy.world - 1day
It would be really interesting to watch a video of this process. Though I'm certain it would be pretty difficult to pull off the editing.
2
Riskable - 1day
You want to see someone using say, VS Code to write something using say, Claude Code?
There's probably a thousand videos of that.
More interesting: I watched someone who was super cheap trying to use multiple AIs to code a project because he kept running out of free credits. Every now and again he'd switch accounts and use up those free credits.
That was an amazing dance, let me tell ya! Glorious!
I asked him which one he'd pay for if he had unlimited money and he said Claude Code. He has the $20/month plan but only uses it in special situations because he'll run out of credits too fast. $20 really doesn't get you much with Anthropic 🤷
That inspired me to try out all the code assist AIs and their respective plugins/CLI tools. He's right: Claude Code was the best by a HUGE margin.
Gemini 3.0 is supposed to be nearly as good but I haven't tried it yet so I dunno.
Now that I've said all that: I am severely disappointed in this article because it doesn't say which AI models were used. In fact, the study authors don't even know what AI models were used. So it's 430 pull requests of random origin, made at some point in 2025.
For all we know, half of those could've been made with the Copilot gpt5-mini that everyone gets for free when they install the Copilot extension in VS Code.
3
justaman123 @lemmy.world - 24hr
It's more I want to see the process of experienced coders explaining the coding mistakes that typical AI coding makes. I have very little experience and see it as a good learning experience. You're probably right about there being tons of videos like that.
3
Riskable - 21hr
The mistakes it makes depends on the model and the language. GPT5 models can make horrific mistakes though where it randomly removes huge swaths of code for no reason. Every time it happens I'm like, "what the actual fuck?" Undoing the last change and trying usually fixes it though 🤷
They all make horrific security mistakes quite often. Though, that's probably because they're trained on human code that is *also" chock full of security mistakes (former security consultant, so I'm super biased on that front haha).
3
Delusions @lemmy.dbzer0.com - 1day
Which is funny because you should be able to just copy and paste And combine from maybe two maybe three GitHub pages pretty easily and you learn just as much
2
pleaseletmein @lemmy.zip - 20hr
Water makes things wetter than fire does.
35
Goldholz - 10hr
Yeah no shit
30
Ledivin @lemmy.world - 17hr
Anyone blindly having AI write their code is an absolute moron.
Anyone with decent experience (5-10 years, maybe 10+?) can absolutely fucking skyrocket their output if they properly set up their environments and treat their agents as junior devs instead of competent programmers. You shouldn't trust generated code any more than you trust someone fresh out of college, but they produce code in seconds instead of weeks.
I have tripled my output while producing more secure code (based on my security audits), safer code (based on code coverage and security audits), and less error-prone code (based on production logs and our unchanged QA process).
Now, the ethical issues and environmental issues, I 100% can get behind. And I have no idea what companies are going to do in 10 years when they have to replace people like me and haven't been hiring or training replacements. But the productivity and quality debates are absolutely ridiculous, as long as a strong dev is behind the wheel and has been trained to use the tools.
People are very bad at judging their own productivity, and AI consistently makes devs feel like they are working faster, while in fact slowing them down.
I've experienced it myself - it feels fucking great to prompt a skeleton and have something brand new up and running in under an hour. The good chemicals come flooding in because I'm doing something new and interesting.
Then I need to take a scalpel to a hundred scattered lines to get CI to pass. Then I need to write tests that actually test functionality. Then I start extending things and realize the implementation is too rigid and I need to change the architecture.
It is as this point that I admit to myself that going in intentionally with a plan and building it myself the slow way would have saved all that pain and probably got the final product shipped sooner, even if the prototype was shipped later.
28
Ledivin @lemmy.world - 8hr
What about my comment made you believe I was using gut feelings to judge anything? My ticket completion rate, number of tickets, story points, and number of projects completed all point to massive productivity gains.
4
skibidi @lemmy.world - 5hr
The end of your comment was
But the productivity and quality debates are absolutely ridiculous
Which is a general statement and not dealing with your specific circumstance. If a tool works for you, by all means keep using it.
However, broadly across software that is not the case. So the "productivity and quality debates" are not ridiculous .. the data supports the sceptics.
5
Ledivin @lemmy.world - 3hr
Which is a general statement and not dealing with your specific circumstance. If a tool works for you, by all means keep using it.
Absolute nonsense. Do people talk shit about hammers because some people keep hitting their hands with them? Do people complain about how useless ladders are, as one of the single most dangerous items in any household?
I don't think we should be putting these tools in the hands of junior devs - as the studies show, it hinders their productivity and learning. But to generally claim that they are bad tools with no upsides is just as ridiculous as the strawman you set up.
2
setsubyou - 11hr
It depends on the task. As an extreme example, I can get AI to create a complete application in a language I don’t know. There’s no way that’s not more productive than me first learning the language to a point where I can make apps in it. Just have to pick something simple enough for the AI.
Of course the opposite extreme also exists. I’ve found that when I demand something impossible, AI will often just try to implement it anyway. It can easily get into an endless cycle where it keeps optimistically declaring that it identified the issue and fixed it with a small change, over and over again. This includes cases where there’s a bug in the underlying OS or similar. You can waste a huge amount of time going down an entirely wrong path if you don’t realize that an idea doesn’t work.
In my real work neither of these really happen. So the actual impact is much less. A lot of my work is not coding in the first place. And I’ve been writing code since I was a little kid, for almost 40 years now. So even the fast scaffolding I can do with AI is not that exciting. I can do that pretty quickly without AI too. When AI coding tools appeared my bosses started asking if I was fast because I was using one. No, I’m fast because some people ask for a new demo every week. Causes the same problems later too.
But I also do think that we all still need to learn how to use AI properly. This applies to all tools, but I think it’s more difficult than with other tools. If I try to use a hammer on something other than a nail, it will not enthusiastically tell me it can do it with just one more small change. AI tools absolutely will though, and it’s easy to just let them try because it’s just a few seconds to see what they come up with. But that’s a trap that leads to those productivity wasting spirals. Especially if the result actually somehow still works at first, so we have to fix it half a year later instead of right away.
At my work there are some other things that I feel limit the productivity potential of AI tools. First of all we’re only allowed to use a very limited number of tools, some of them made in-house. Then we’re not really allowed to integrate them into our workflows other than the part where we write code. E.g. I could trivially write an mcp server that interacts with our (custom in-house) ci system and actually increases my productivity because I could save a small number of seconds very often if I could tell an AI to find builds for me for integration or QA work. But it’s not allowed. We’re all being pushed to use AI but the company makes it really difficult at the same time.
So when I play around with AI on my spare time I do actually feel like I’m getting a huge boost. Not just because I can use a claude model instead of the ones I can use at work, but also just basic things like e.g. being able to turn on AI in Xcode at all when working on software for Apple platforms. On my work Macbook I can’t turn on any Apple AI features at all so even tab completion is worse. Or in other words, those realities of working on serious projects at a serious company with serious security policies can also kill any potential productivity boost from AI. They basically expect us to be productive with only those features the non-developer CEO likes, who also doesn’t have to follow any of our development processes…
Did they compare it to the code of that outsourced company that provided the lowest bid? My company hasn’t used AI to write code yet. They outcourse/offshore. The code is held together with hopes and dreams. They remove features that exist, only to have to release a hot fix to add it back. I wish I was making that up.
24
dustyData @lemmy.world - 5hr
Cool, the best AI has to offer is worse than the worst human code. Definitely worth burning the planet to a crisp for it.
6
coolmojo @lemmy.world - 5hr
And how do you know if the other company with the cheapest bid actually does not just vibe code it? With all that said it could be plain incompetence and ignorance as well.
5
JaddedFauceet @lemmy.world - 3hr
Because it has been like this before vibe coding existed...
5
kinther @lemmy.world - 4hr
That's a valid question, especially with AI coding being so prevalent.
3
SpicyTaint @lemmy.world - 1day
...is this supposed to be news?
20
UnderpantsWeevil @lemmy.world - 1day
Kinda. It's a novel technology and one that hasn't been well analyzed or exhaustively tested.
6
DudeImMacGyver - 1day
It's been tested a lot and the results are that it can't be trusted at all unless you are already an expert in the thing you're asking it to "help" you with so you can correct the many mistakes it will make, but it's slower and, again, is **guaranteed **to make mistakes (hallucinations are built into what techbros are insisting on labeling as "AI", no matter how many resources you throw at it).
All of this at great environmental and human cost too.
7
Davel23 - 1day
I think his point is that this is less "news", and more "well, duh".
4
WanderingThoughts @europe.pub - 22hr
And even worse, it doesn't realise it and can't fix the errors.
18
kalkulat @lemmy.world - 13hr
I'd never ask a friggin machine to do coding for me, that's MY blast.
That said, I've had good luck asking GPT specific questions about multiple obscure features of Javascript, and of various browsers. It'll often feed me a sample script using a feature it explains ... a lot more helpful than many of the wordy websites like MDN ... saving me shit-tons of time that I'd spend bouncing around a half-dozen 'help' pages.
15
Derpgon - 8hr
I've been using it to code a microservice as PoC for semantic search. As I've basically never coded Python (mainly PHP, but can do many langs) I've had to rely on AI (Kimi K2, or agentic Claude I think 4.5 or 4, can't remember) because I don't know the syntax, features, best practices, and tools to use for formatting, static analysis, and type checks.
Mind you, I've basically never coded in Python besides some shit in uni, which was 5-10 years ago. AI was a big help - albeit it didn't spit out fully working code, I have enough knowledge in this field to fix the issues. As I learn mainly by practice and not theory, AI is great because - same as many YouTubers and free tutorials - it spits out unoptimized and broken code.
I am usually not using it for my main line of work (PHP) besides some boiler plate (take this class, make a test, make it look the same as this other test = 300 lines I don't have to write myself).
4
azvasKvklenko @sh.itjust.works - 3hr
Oh, so my sceptical, uneducated guesses about AI are mostly spot on.
13
🍉 Albert 🍉 - 2hr
As a computer science experiment, making a program that can beat the Turing test is a monumental step in progress.
However as a productive tool it is useless in practically everything it is implemented on. It is incapable of performing the very basic "Sanity check" that is important in programming.
6
robobrain @programming.dev - 1hr
The Turing test says more about the side administering the test than the side trying to pass it
Just because something can mimic text sufficiently enough to trick someone else doesn't mean it is capable of anything more than that
2
🍉 Albert 🍉 - 1hr
We can argue about it's nuances. same with the Chinese room thought experiment.
However, we can't deny that it the Turing test, is no longer a thought exercise but a real test that can be passed under parameters most people would consider fair.
I thought a computer passing the Turing test would have more fanfare, about the morality if that problem, because the usual conclusion of that thought experiment was "if you cant tell the difference, is there one?", but now it has become "Shove it everywhere!!!".
2
M0oP0o @mander.xyz - 34min
Oh, I just realized that the whole ai bubble is just the whole "everything is a dildo if you are brave enough."
2
🍉 Albert 🍉 - 24min
yhea, and "everything is a nail if all you got is a hammer".
there are some uses for that kind of AI, but very limiting. less robotic voice assisants, content moderation, data analysis, quantification of text. the closest thing to Generative use should be to improve auto complete and spell checking (maybe, I'm still not sure on those ones)
2
M0oP0o @mander.xyz - 18min
I was wondering how they could make autocomplete worse, and now I know.
1
RememberTheApollo_ @lemmy.world - 2hr
The Turing Test has shown its weakness.
2
🍉 Albert 🍉 - 59min
Time for a Turing 2.0?
If you spend a lifetime with a bot wife and were unable to tell that she was AI, is there a difference?
1
RampantParanoia2365 @lemmy.world - 20hr
I'm not a programmer, but I've dabbled with Blender for 3D modeling, and it uses Node trees for a lot of different things, which is pretty much a programming GUI. I googled how to make a shader, and the AI gave me instructions. About half of it was complete nonsense, but I did make my shader.
12
fox2263 - 1day
You need to babysit and double check everything it does. You can’t just let it loose and trust everything it does.
10
Affidavit @lemmy.world - 5hr
I really, really, want to stop seeing posts about:
Musk
Trump
Israel
Microsoft
AI
I swear these are the only things that the entire Lemmy world wants to talk about.
Maybe I should just go back to Reddit... Fuck Spez, but at least there is some variety.
9
magz :3 - 5hr
your frontend of choice probably has some option to hide posts containing specific keywords
9
Affidavit @lemmy.world - 4hr
Yeah, good point. Was hoping to avoid downloading another random app, but at this stage, I guess It's something I should look into.
4
naticus - 2hr
Yes, please just hide these. We ignoring these issues at large is how we got to where we're at and it'll continue getting worse if we just stop talking about it. But you need to do what you can to take of yourself, first and foremost.
1
andallthat @lemmy.world - 3hr
Microsoft could write an AI agent to filter threads based on context you don't like.
Come to think of it, Megagenius Elon Musk already has one he wrote to censor anti-Israel posts on Trump's Truth Social.
There, I think I got them all.... Happy holidays!
3
Minizarbi @jlai.lu - 4hr
Not my code though. It contains a shit ton of bugs. When I am able to write some of course.
9
jj4211 @lemmy.world - 3hr
Nah, AI code gen bugs are weird. As a person used to doing human review even from wildly incompetent people, AI messes up things that my mind never even thought needed to be double checked.
8
🍉 Albert 🍉 - 2hr
Human bugs >>> AI bug slop
3
MonkderVierte @lemmy.zip - 9hr
This is news?
8
DudeImMacGyver - 1day
I'll go ahead and file this under "duh".
6
chunes @lemmy.world - 21hr
People expect perfection right out of the gate.
I mean damn, AI has only been able to write something resembling code for a few years now. The fact that this is even a headline is pretty amazing when you think about it.
6
5too @lemmy.world - 18hr
I don't mind imperfections while they work out the kinks. I dislike dismantling industries in favor of something that doesn't work yet.
7
Eranziel @lemmy.world - 17hr
A lot of LLM hype is wrapped up in how well it can write code. This hype is being used by corporations to justify pouring mind boggling amounts of money into the tech in the hopes that they can lay off all their staff.
I reserve the right to hate this state of affairs and enjoy seeing every headline that shows just how much of a pipe dream it is.
6
kent_eh @lemmy.ca - 1hr
AI-generated code produces 1.7x more issues than human code
AI doesn't generate its own code, humans using AI generate code. If a person uses AI to generate code and doesn't know good practices then of course the code is going to be worse.
3
VeryFrugal - 13hr
Shocker.
3
themurphy @lemmy.ml - 1day
Well, yeah. It also took 100x the time to write.
Vibe coding is only really useful for a coder. Because you can understand and correct it.
Do not ask a corpse for advice, the question is what are we going to do?
Boycott is a good first step, although I am not sure if it is better to boycott them or use their free tier to have the most deranged BS conversation that will consume their resources, eat at their scare cash reserves and when they use it in training, it will poison their data.
throws_lemy in technology @lemmy.world
AI-generated code contains more bugs and errors than human output
https://www.techradar.com/pro/security/ai-generated-code-contains-more-bugs-and-errors-than-human-outputAlmost as if it was made to simulate human output but without the ability to scrutinize itself.
To be fair most humans don't scrutinize themselves either.
(Fuck AI though. Planet burning trash)
The number of times I have received an un-proofread two sentence email is too damn high.
And then the follow up email because they didn’t actually finish a complete thought
I do this with texts/DMs, but I'd never do that with an email. I double or triple check everything, make sure my formatting is good, and that the email itself is complete. I'll DM someone 4 or 5 times in 30 seconds though, it feels like a completely different medium ¯\_(ツ)_/¯
It's like having a lightning-fast junior developer at your disposal. If you're vague, he'll go on shitty side-quests. If you overspecify he'll get overwhelmed. You need to break down tasks into manageable chunks. You'll need to ask follow-up questions about every corner case.
A real junior developer will have improved a lot in a year. Your AI agent won't have improved.
This is the real thing. You can absolutely get good code out of AI, but it requires a lot of hand holding. It helps me speed some tasks, especially boring ones, but I don't see it ever replacing me. It makes far too many errors, and requires me to point them out, and to point in the direction of the solution.
They are great at churning out massive amounts of code. They're also great at completely missing the point. And the massive amount of code needs to be checked and reviewed. Personally I'd rather write the code and have the AI review it. That's a much more pleasant way to work, and that way it actually enhances quality.
They are improving, and probably faster then junior devs. The models we had had 2 years ago would struggle with a simple black jack app. I don't think the ceiling has been hit.
Just a few trillion more dollars, bro. We’re almost there. Bro, if you give up a few showers, the AI datacenter will be able to work perfectly.
Bro.
The cost of the improvement doesn't change the fact that it's happening. I guess we could all play pretend instead if it makes you feel better about it. Don't worry bro, the models are getting dumber!
That would be pretty impressive when they already lack any intelligence at all.
And I ask you - if those same trillions of dollars were instead spent on materially improving the lives of average people, how much more progress would we make as a society? This is an absolutely absurd sum of money were talking about here.
It's beside the point. I'm simply saying that AI will improve in the next year. The cost to do so or all the others things that money could be spent on doesn't matter when it's clearly going to be spent on AI. I'm not in charge of monetary policies anywhere, I have no say in the matter. I'm just pushing back on the fantasies. I'm hoping the open source scene survives so we don't end up in some ugly dystopia where all AI is controlled by a handful of companies.
I have the impression that anti-AI people don't understand that they are giving up agency for the sake of temporary feels. If they truly cared about ethical usage of AI, they would be wanting to have mastery that is at least equal to that of corporations and the 1%.
Making AI into a public good is key to a better future.
They might. The amount of money they're pumping into this is absolutely staggering. I don't see how they're going to make all of that money back, unless they manage to replace nearly all employees.
Either way it's going to be a disaster: mass unemployment or the largest companies in the world collapsing.
I dunno, the death of mega corporations would do the world a great deal of good. Healthier capitalism requires competition, and a handful of corporations of any given sector isn't going to seriously compete nor pay good wages.
It's certainly the option I'm rooting for, but it would still be a massive drama and disrupt a lot of lives. Which is why they'll probably get bailed out with taxpayer money.
My jr developer will eventually be familiar with the entire codebase and can make decisions with that in mind without me reminding them about details at every turn.
LLMs would need massive context windows and/or custom training to compete with that. I'm sure we'll get there eventually, but for now it seems far off. I think this bubble will have to burst and let hardware catch up with our ambitions. It'll take a couple of decades.
A computer is a machine that makes human errors at the speed of electricity.
I think one of the big issues is it often makes nonhuman errors. Sometimes I forget a semicolon or there's a typo, but I'm well equipped to handle that. In fact, most programs can actually catch that kind of issue already. AI is more likely to generate code that's hard to follow and therefore harder to check. It makes debugging more difficult.
Sure. It's making the errors faster and at a far higher volume than any team of humans could do in twice the time. The technology behind inference is literally an iterative process of turning gibberish into something that resembles human text. So its sort of a speed run from baby babble into college level software design by trial, evaluation, and correction over and over and over again.
But because the baseline comparison code is, itself, full of errors, the estimation you get at the end of the process is going to be scattering errant semicolons (and far more esoteric coding errors) through the body of the program at a frequency equivalent to humans making similar errors over a much longer timeline.
Also seems like it'd be a lot harder to modify or extend later
https://wallpapercave.com/wp/wp5338276.jpg
No shit.
I actually believed somebody when they told me it was great at writing code, and asked it to write me the code for a very simple lua mod. It’s made several errors and ended up wasting my time because I had to rewrite it.
In a postgraduate class, everyone was praising ai, calling it nicknames and even their friend (yes, friend), and one day, the professor and a colleague were discussing some code when I approached, and they started their routine bullying on me for being dumb and not using ai. Then I looked at his code and asked to test his core algorithm that he converted from a fortran code and "enhanced" it. I ran it with some test data and compared to the original code and the result was different! They blindly trusted some ai code that deviated from their theoretical methodology, and are publishing papers with those results!
Even after showing the different result, they didn't convince themselves of anything and still bully me for not using ai. Seriously, this shit became some sort of cult at this point. People are becoming irrational. If people in other universities are behaving the same and publishing like this, I'm seriously concerned for the future of science and humanity itself. Maybe we should archive everything published up to 2022, to leave as a base for the survivors from our downfall.
The way it was described to me by some academics is that it’s useful…but only as a “research assistant” to bounce ideas off of and bring in arcane or tertiary concepts you might not have considered (after you vet them thoroughly, of course).
The danger, as described by the same academics, is that it can act as a “buddy” who confirms you biases. It can generate truly plausible bullshit to support deeply flawed hypotheses, for example. Their main concern is it “learning” to stroke the egos of the people using it so it creates a feedback loop and it’s own bubbles of bullshit.
So, linkedin? What if the real artificial intelligence was the linkedin lunatics we met along the way?
That's not a bad idea. I'm already downloading lots of human knowledge and media that I want backed up because I can't trust humanity anymore to have it available anymore
I've been coding for a while. I did an honest eager attempt at making a real functioning thing with all code written by AI. A breakout clone using SDL2 with music.
The game should look good, play good, have cool effects, and be balanced. It should have an attractor screen, scoring, a win state and a lose state.
I also required the code to be maintainable. Meaning I should be able to look at every single line and understand it enough to defend its existence.
I did make it work. And honestly Claude did better than expected. The game ran well and was fun.
But: The process was shit.
I spent 2 days and several hundred dollars to babysit the AI, to get something I could have done in 1 day including learning SDL2.
Everything that turned out well, turned out well because I brought years of skill to the table, and could see when Claude was coding itself into a corner and tell it to break up code in modules, collate globals, remove duplication, pull out abstractions, etc. I had to detect all that and instruct on how to fix it. Until I did it was adding and re-adding bugs because it had made so much shittily structured code it was confusing itself.
TLDR; LLM can write maintainable code if given full constant attention by a skilled coder, at 40% of the coder's speed.
It depends on the subject area and your workflow. I am not an AI fanboy by any stretch of the imagination, but I have found the chatbot interface to be a better substitute for the "search for how to do X with library/language Y" loop. Even though it's wrong a lot, it gives me a better starting place faster than reading through years-old SO posts. Being able to talk to your search interface is great.
The agentic stuff is also really good when the subject is something that has been done a million times over. Most web UI areas are so well trodden that JS devs have already invented a thousand frameworks to do it. I'm not a UI dev, so being able to give the agent a prompt like, "make a configuration UI with a sidebar that uses the graphql API specified here" is quite nice.
AI is trash at anything it hasn't been trained on in my experience though. Do anything niche or domain-specific, and it feels like flipping a coin with a bash script. It just throws shit at the wall and runs tests until the tests pass (or it sneakily changes the tests because the error stacktrace repeatedly indicates the same test line as the problem).
Yeah what you say makes sense to me. Having it make a "wrong start" in something new is useful, as it gives you a lot of the typical structure, introduces the terminology, maybe something sorta moving that you can see working before messing with it, etc.
It’s basically just for if you’re lazy and don’t want to write a bunch of boilerplate or hit your keyboard a bunch of times to move the cursor(s) around
It is great for boilerplate code. It can also explain code for you, or help with an unfamiliar library. It's even helped me be productive when my brain wasn't ready to really engage with the code.
But here's the real danger: because I've got AI to do it for me, my brain doesn't have to engage fully with the code anymore. I don't really get into the flow where code just flows out of your hands like I used to. It's becoming a barrier between me and the real magic of coding. And that sucks, because that's what I love about this work. Instead, I'm becoming the AI's manager. I never asked for that.
I generally agree with what you’ve said for sure. I think I’ve honestly started to use it for helping me to go pinpoint where to go look for issues in the spaghetti code of new code bases. I’ve also mostly tried to avoid using it in my personal coding time but I feel like it’s gotten harder and harder to get legitimately good search results nowadays which I realize is also because of ai. Given the choice I’d happily just erase it from existence I think. Spending hours sifting through reddit and stack overflow was way more fulfilling + I feel like people used to be slightly less prickly about answering stuff because that was how you had to get answers. It seems like lemmy could replace that space at least, I’ve genuinely gotten helpful comments and I’ve always felt downvotes on here have been productive versus what Reddit is now.
I've found the same thing. I've turned off the auto suggestions while tying because by the time I'm typing i already know what I'm going I'm to type and having mostly incorrect suggestions popping up every 2 seconds was distracting and counterproductive.
This was a very directed experiment at purely LLM written maintainable code.
Writing experiments and proof of concepts, even without skill, will give a different calculation and can make more sense.
Having it write a "starting point" and then take over, also is a different thing that can make more sense. This requires a coder with skill, you can't skip that.
It would be really interesting to watch a video of this process. Though I'm certain it would be pretty difficult to pull off the editing.
You want to see someone using say, VS Code to write something using say, Claude Code?
There's probably a thousand videos of that.
More interesting: I watched someone who was super cheap trying to use multiple AIs to code a project because he kept running out of free credits. Every now and again he'd switch accounts and use up those free credits.
That was an amazing dance, let me tell ya! Glorious!
I asked him which one he'd pay for if he had unlimited money and he said Claude Code. He has the $20/month plan but only uses it in special situations because he'll run out of credits too fast. $20 really doesn't get you much with Anthropic 🤷
That inspired me to try out all the code assist AIs and their respective plugins/CLI tools. He's right: Claude Code was the best by a HUGE margin.
Gemini 3.0 is supposed to be nearly as good but I haven't tried it yet so I dunno.
Now that I've said all that: I am severely disappointed in this article because it doesn't say which AI models were used. In fact, the study authors don't even know what AI models were used. So it's 430 pull requests of random origin, made at some point in 2025.
For all we know, half of those could've been made with the Copilot gpt5-mini that everyone gets for free when they install the Copilot extension in VS Code.
It's more I want to see the process of experienced coders explaining the coding mistakes that typical AI coding makes. I have very little experience and see it as a good learning experience. You're probably right about there being tons of videos like that.
The mistakes it makes depends on the model and the language. GPT5 models can make horrific mistakes though where it randomly removes huge swaths of code for no reason. Every time it happens I'm like, "what the actual fuck?" Undoing the last change and trying usually fixes it though 🤷
They all make horrific security mistakes quite often. Though, that's probably because they're trained on human code that is *also" chock full of security mistakes (former security consultant, so I'm super biased on that front haha).
Which is funny because you should be able to just copy and paste And combine from maybe two maybe three GitHub pages pretty easily and you learn just as much
Water makes things wetter than fire does.
Yeah no shit
Anyone blindly having AI write their code is an absolute moron.
Anyone with decent experience (5-10 years, maybe 10+?) can absolutely fucking skyrocket their output if they properly set up their environments and treat their agents as junior devs instead of competent programmers. You shouldn't trust generated code any more than you trust someone fresh out of college, but they produce code in seconds instead of weeks.
I have tripled my output while producing more secure code (based on my security audits), safer code (based on code coverage and security audits), and less error-prone code (based on production logs and our unchanged QA process).
Now, the ethical issues and environmental issues, I 100% can get behind. And I have no idea what companies are going to do in 10 years when they have to replace people like me and haven't been hiring or training replacements. But the productivity and quality debates are absolutely ridiculous, as long as a strong dev is behind the wheel and has been trained to use the tools.
Consider: the facts
People are very bad at judging their own productivity, and AI consistently makes devs feel like they are working faster, while in fact slowing them down.
I've experienced it myself - it feels fucking great to prompt a skeleton and have something brand new up and running in under an hour. The good chemicals come flooding in because I'm doing something new and interesting.
Then I need to take a scalpel to a hundred scattered lines to get CI to pass. Then I need to write tests that actually test functionality. Then I start extending things and realize the implementation is too rigid and I need to change the architecture.
It is as this point that I admit to myself that going in intentionally with a plan and building it myself the slow way would have saved all that pain and probably got the final product shipped sooner, even if the prototype was shipped later.
What about my comment made you believe I was using gut feelings to judge anything? My ticket completion rate, number of tickets, story points, and number of projects completed all point to massive productivity gains.
The end of your comment was
Which is a general statement and not dealing with your specific circumstance. If a tool works for you, by all means keep using it.
However, broadly across software that is not the case. So the "productivity and quality debates" are not ridiculous .. the data supports the sceptics.
Absolute nonsense. Do people talk shit about hammers because some people keep hitting their hands with them? Do people complain about how useless ladders are, as one of the single most dangerous items in any household?
I don't think we should be putting these tools in the hands of junior devs - as the studies show, it hinders their productivity and learning. But to generally claim that they are bad tools with no upsides is just as ridiculous as the strawman you set up.
It depends on the task. As an extreme example, I can get AI to create a complete application in a language I don’t know. There’s no way that’s not more productive than me first learning the language to a point where I can make apps in it. Just have to pick something simple enough for the AI.
Of course the opposite extreme also exists. I’ve found that when I demand something impossible, AI will often just try to implement it anyway. It can easily get into an endless cycle where it keeps optimistically declaring that it identified the issue and fixed it with a small change, over and over again. This includes cases where there’s a bug in the underlying OS or similar. You can waste a huge amount of time going down an entirely wrong path if you don’t realize that an idea doesn’t work.
In my real work neither of these really happen. So the actual impact is much less. A lot of my work is not coding in the first place. And I’ve been writing code since I was a little kid, for almost 40 years now. So even the fast scaffolding I can do with AI is not that exciting. I can do that pretty quickly without AI too. When AI coding tools appeared my bosses started asking if I was fast because I was using one. No, I’m fast because some people ask for a new demo every week. Causes the same problems later too.
But I also do think that we all still need to learn how to use AI properly. This applies to all tools, but I think it’s more difficult than with other tools. If I try to use a hammer on something other than a nail, it will not enthusiastically tell me it can do it with just one more small change. AI tools absolutely will though, and it’s easy to just let them try because it’s just a few seconds to see what they come up with. But that’s a trap that leads to those productivity wasting spirals. Especially if the result actually somehow still works at first, so we have to fix it half a year later instead of right away.
At my work there are some other things that I feel limit the productivity potential of AI tools. First of all we’re only allowed to use a very limited number of tools, some of them made in-house. Then we’re not really allowed to integrate them into our workflows other than the part where we write code. E.g. I could trivially write an mcp server that interacts with our (custom in-house) ci system and actually increases my productivity because I could save a small number of seconds very often if I could tell an AI to find builds for me for integration or QA work. But it’s not allowed. We’re all being pushed to use AI but the company makes it really difficult at the same time.
So when I play around with AI on my spare time I do actually feel like I’m getting a huge boost. Not just because I can use a claude model instead of the ones I can use at work, but also just basic things like e.g. being able to turn on AI in Xcode at all when working on software for Apple platforms. On my work Macbook I can’t turn on any Apple AI features at all so even tab completion is worse. Or in other words, those realities of working on serious projects at a serious company with serious security policies can also kill any potential productivity boost from AI. They basically expect us to be productive with only those features the non-developer CEO likes, who also doesn’t have to follow any of our development processes…
AI has made being OE insanely easy.
https://lemmy.dbzer0.com/pictrs/image/04024ce7-ab54-45b3-938e-38621ffa0cfb.webp
Did they compare it to the code of that outsourced company that provided the lowest bid? My company hasn’t used AI to write code yet. They outcourse/offshore. The code is held together with hopes and dreams. They remove features that exist, only to have to release a hot fix to add it back. I wish I was making that up.
Cool, the best AI has to offer is worse than the worst human code. Definitely worth burning the planet to a crisp for it.
And how do you know if the other company with the cheapest bid actually does not just vibe code it? With all that said it could be plain incompetence and ignorance as well.
Because it has been like this before vibe coding existed...
That's a valid question, especially with AI coding being so prevalent.
...is this supposed to be news?
Kinda. It's a novel technology and one that hasn't been well analyzed or exhaustively tested.
It's been tested a lot and the results are that it can't be trusted at all unless you are already an expert in the thing you're asking it to "help" you with so you can correct the many mistakes it will make, but it's slower and, again, is **guaranteed **to make mistakes (hallucinations are built into what techbros are insisting on labeling as "AI", no matter how many resources you throw at it).
All of this at great environmental and human cost too.
I think his point is that this is less "news", and more "well, duh".
And even worse, it doesn't realise it and can't fix the errors.
I'd never ask a friggin machine to do coding for me, that's MY blast.
That said, I've had good luck asking GPT specific questions about multiple obscure features of Javascript, and of various browsers. It'll often feed me a sample script using a feature it explains ... a lot more helpful than many of the wordy websites like MDN ... saving me shit-tons of time that I'd spend bouncing around a half-dozen 'help' pages.
I've been using it to code a microservice as PoC for semantic search. As I've basically never coded Python (mainly PHP, but can do many langs) I've had to rely on AI (Kimi K2, or agentic Claude I think 4.5 or 4, can't remember) because I don't know the syntax, features, best practices, and tools to use for formatting, static analysis, and type checks.
Mind you, I've basically never coded in Python besides some shit in uni, which was 5-10 years ago. AI was a big help - albeit it didn't spit out fully working code, I have enough knowledge in this field to fix the issues. As I learn mainly by practice and not theory, AI is great because - same as many YouTubers and free tutorials - it spits out unoptimized and broken code.
I am usually not using it for my main line of work (PHP) besides some boiler plate (take this class, make a test, make it look the same as this other test = 300 lines I don't have to write myself).
Oh, so my sceptical, uneducated guesses about AI are mostly spot on.
As a computer science experiment, making a program that can beat the Turing test is a monumental step in progress.
However as a productive tool it is useless in practically everything it is implemented on. It is incapable of performing the very basic "Sanity check" that is important in programming.
The Turing test says more about the side administering the test than the side trying to pass it
Just because something can mimic text sufficiently enough to trick someone else doesn't mean it is capable of anything more than that
We can argue about it's nuances. same with the Chinese room thought experiment.
However, we can't deny that it the Turing test, is no longer a thought exercise but a real test that can be passed under parameters most people would consider fair.
I thought a computer passing the Turing test would have more fanfare, about the morality if that problem, because the usual conclusion of that thought experiment was "if you cant tell the difference, is there one?", but now it has become "Shove it everywhere!!!".
Oh, I just realized that the whole ai bubble is just the whole "everything is a dildo if you are brave enough."
yhea, and "everything is a nail if all you got is a hammer".
there are some uses for that kind of AI, but very limiting. less robotic voice assisants, content moderation, data analysis, quantification of text. the closest thing to Generative use should be to improve auto complete and spell checking (maybe, I'm still not sure on those ones)
I was wondering how they could make autocomplete worse, and now I know.
The Turing Test has shown its weakness.
Time for a Turing 2.0?
If you spend a lifetime with a bot wife and were unable to tell that she was AI, is there a difference?
I'm not a programmer, but I've dabbled with Blender for 3D modeling, and it uses Node trees for a lot of different things, which is pretty much a programming GUI. I googled how to make a shader, and the AI gave me instructions. About half of it was complete nonsense, but I did make my shader.
You need to babysit and double check everything it does. You can’t just let it loose and trust everything it does.
I really, really, want to stop seeing posts about:
I swear these are the only things that the entire Lemmy world wants to talk about.
Maybe I should just go back to Reddit... Fuck Spez, but at least there is some variety.
your frontend of choice probably has some option to hide posts containing specific keywords
Yeah, good point. Was hoping to avoid downloading another random app, but at this stage, I guess It's something I should look into.
Yes, please just hide these. We ignoring these issues at large is how we got to where we're at and it'll continue getting worse if we just stop talking about it. But you need to do what you can to take of yourself, first and foremost.
Microsoft could write an AI agent to filter threads based on context you don't like. Come to think of it, Megagenius Elon Musk already has one he wrote to censor anti-Israel posts on Trump's Truth Social. There, I think I got them all.... Happy holidays!
Not my code though. It contains a shit ton of bugs. When I am able to write some of course.
Nah, AI code gen bugs are weird. As a person used to doing human review even from wildly incompetent people, AI messes up things that my mind never even thought needed to be double checked.
Human bugs >>> AI bug slop
This is news?
I'll go ahead and file this under "duh".
People expect perfection right out of the gate.
I mean damn, AI has only been able to write something resembling code for a few years now. The fact that this is even a headline is pretty amazing when you think about it.
I don't mind imperfections while they work out the kinks. I dislike dismantling industries in favor of something that doesn't work yet.
A lot of LLM hype is wrapped up in how well it can write code. This hype is being used by corporations to justify pouring mind boggling amounts of money into the tech in the hopes that they can lay off all their staff.
I reserve the right to hate this state of affairs and enjoy seeing every headline that shows just how much of a pipe dream it is.
As expected
AI doesn't generate its own code, humans using AI generate code. If a person uses AI to generate code and doesn't know good practices then of course the code is going to be worse.
Shocker.
Well, yeah. It also took 100x the time to write.
Vibe coding is only really useful for a coder. Because you can understand and correct it.
Not if Yandev has anything to say about it.
Depends on who the human is haha
Yeah, you've clearly never seen my code!
https://feddit.org/pictrs/image/caff1e5f-b252-43ea-8636-d28203747026.gif
what would marx do?
Do not ask a corpse for advice, the question is what are we going to do?
Boycott is a good first step, although I am not sure if it is better to boycott them or use their free tier to have the most deranged BS conversation that will consume their resources, eat at their scare cash reserves and when they use it in training, it will poison their data.