Generative AI is great for loads of programming tasks like helping create regular expressions or syntax conversions between languages. The main issue I’ve seen in codebases that rely heavily on generative AI is that the “solutions” often fix today’s bug while making future debugging more difficult. Generative AI makes it easy to go fast in the wrong direction. Used right it’s a useful tool.
Yep, by definition generative AI gets worse the more specific you get. If you need common templates though, it’s almost as good as today’s google.
lol Uplevel’s “”“full report”“” saying devs using Copilot create 41% more bugs has 2 pages and reads like a promotional material.
you can download it with a 10 minute email if you really want to see for yourself.
just some meaningless numbers.
For me, it is a glorified auto-complete function. Could definitely live without it.
Same for me, but that glorified auto complete helps a lot.
Hell yea. Our unit test coverage went way up because you can blow through test creation in second. I had a large complicated migration from one data set to another with specific mutations based on weird rules and GPT got me 80% of the way there and with a little nudging basically got it perfect. Code that would’ve taken a few hours took about 6 prompts. If I’m curious about a new library I can get a working example right away to see how everything fits together. When these articles say there’s no benefit I feel people aren’t using these tools or don’t know how to use them effectively.
Good devs gain little.
I gain a lot.
Its basically a template generator, which is really helpful when you’re generating boilerplate. It doesn’t save me much if any time to refactor/fill in that template, but it does save some mental fatigue that I can then spend on much more interesting problems.
It’s a niche tool, but occasionally quite handy. Without leaps forward technically though, it’s never going to become more than that.
Feel the same way!
My main use is skipping the blank page problem when writing a new suite of tests—which after about 10 mins of refactoring are often a good starting point
And yet, higher ups continue to lay off more devs because AI “is the future”.
I truly don’t understand the tendency of people to hate these kinds of tools. Honestly seems like an ego thing to me.
Having to deal with pull requests defecated by “developers” who blindly copy code from chatgpt is a particularly annoying and depressing waste of time.
At least back when they blindly copied code from stack overflow they had to read through the answers and comments and try to figure out which one fit their use case better and why, and maybe learn something… now they just assume the LLM is right (despite the fact that they asked the wrong question and even if they had asked the right one it’d’ve given the wrong answer) and call it a day; no brain activity or learning whatsoever.
Carbon footprint. Techbro arrogance. Not sure what’s hard to understand about it.
Yeah I’m sure you are concerned about the carbon footprint of it and that some dude you talked to once was arrogant about this technology.
Of course you know me better than myself.
I guess you wanted an answer but decided upfront you weren’t gonna like it no matter what? Not much I can do about that.
Also, when a tool increases your productivity but your salary and paid time off don’t increase, it’s a tool that only benefits the overlords and as such deserves to be hated.
Some people feel proud that their work got done quicker and also aren’t micromanaged so if they choose, yes actually they can have more time for their personal lives. Not everyone’s job is purely a transaction in which they do the absolute minimum they can do without being fired.
I hope you feel better soon, because you’re clearly bitter and lashing out at whatever you can lash at.
Places GPT-based “AI” next to flying cars
Flying cars exist, they’re just not cost effective. AFAICT there’s no GPT that is proficient at coding yet.
It’s a lot easier to access ChatGPT than it is to access a flying car
The more people using chatgpt to generate low quality code they don’t understand, the more job safety and greater salary I get.
I honestly stopped using it after a week
I like to use suggestions to feel superior when trash talking the generated code
I’m shocked. There must be an error in this analysis. /s
Maybe engage an AI coding assistant to massage the data analysis lol
Devs that are punching above their class, however, probably get great benefit from it. I would think it’s also an OK learning tool, except for how inaccurate it can be sometimes.