ChatGPT is hilariously incompetent… but on a serious note, I still firmly reject tools like copilot outside demos and the like because they drastically reduce code quality for short term acceleration. That’s a terrible trade-off in terms of cost.
As a software engineer, the number of people I encounter in a given week who either refuse to or are incapable of understanding that distinction baffles and concerns me.
The problem I have with it is that all the time it saves me I have to use on reading the code. I probably spend more time on that as once in a while the code it produces is broken in a subtle way.
I see some people swearing by it, which is the opposite of my experience. I suspect that if your coding was copying code from stack overflow then it indeed improved your experience as now this process is streamlined.
I don’t know if it does yet, but if ChatGPT starts providing source for every information, then it would make it much faster to find the relevant information and check their sources, rather than clicking websites one by one.
It used to have a problem with making a claims that were not relevant to or contradicted its own sources, but I don’t recall encountering that problem recently.
Biggest problem with it is that it lies with the exact same confidence it tells the truth.
Or, put another way, it’s confidently incorrect as often as it is confidently correct - and there’s no way to tell the difference unless you already know the answer.
it’s kinda hilarious to me because one of the FIRST things ai researchers did was get models to identify things and output answers together with the confidence of each potential ID, and now we’ve somehow regressed back from that point
i mean giving a confidence for recognizing a certain object in a picture is relatively straightforward.
But LLMs put together words by their likeliness of belonging together under your input (terribly oversimplified).the confidence behind that has no direct relation to how likely the statements made are true. I remember an example where someone made chatgpt say that 2+2 equals 5 because his wife said so. So chatgpt was confident that something is right when the wife says it, simply because it thinks these words to belong together.
But LinkedIn bros and corporate people are gonna gobble it up anyways because it has the right buzzwords (including “AI”) and they can squeeze more (low quality) work from devs to brag about how many things they (the corporate owners) are doing.
It’s just a fad. There’s just a small bit that will stay after the hype is gone. You know, like blockchain, AR, metaverse, NFT and whatever it was before that. In a few years there will be another breakthrough with it and we’ll hear from it again for a short while, but for now it’s just a one trick pony.
The hilarious thing about blockchain is that the core concept is actively making the whole thing worse. The matrix protocol is sort of essentially blockchain without the decentralized ledger part, and it’s so vastly superior in every single way.
NFTs just show how fundamentally dumb blockchains are, if you skip the decentralized ledger bit then you never need to invent NFT functionality in the first place…
Well, if you stop listening to people who think it’s a way to get really rich really fast (which it obviously isn’t), cryptocurrencies are quite useful. International transfers are so much cheaper and easier with them.
Yes, you could, for example, use it to manage who is allowed to park in a garage, anonymously. The owner of a parking spot NFT can unlock the door from the outside. Stuff like that.
However, it’s also possible to do that with a small web application. Just payments and transfer of the parking spots are less free and it’s not decentralized.
I disagree, because unlike those things, ai actually has a use case. It needs a human supervisor and it isn’t always faster, but chat gpt has been the best educational resource I’ve ever had next to YouTube. It’s also decent at pumping out a lot of lazy work and writing so i don’t have to, or helping me break down a task into smaller parts. As long as you’re not expecting it to solve all your problems for you, it’s an amazing tool.
People said the same things about 3d printing and yeah, while it can’t create literally everything at industrial scale, and it’s not going to see much consumer use, it has found a place in certain applications like prototyping and small scale production.
ChatGPT is OK at summarizing popular, low specificity topics that tons of people have already written a ton about, but it’s terrible at anything else. When I tested its knowledge about the process of a niche interest of mine (fabric dyeing) it skipped completely over certain important pieces of information, and when I prompted it to include them it basically just mirrored my prompt back at me.
Which has pretty much summed up my ChatGPT experience: it just regurgitates stuff I can find myself, but removes the ability to determine if the source is reliable. And if it’s something I’m already having trouble finding detailed information about it usually doesn’t help.
i’d argue that chat gpt is mostly great at taking human bullshit tasks from humans, who dontwant to dothe bullshit, like regurgitating the text from a textbook in different words, writing cover letters for job applications, that are often machine analyzed for buzzwords anyways.
So its use case only exist in the domain of bullshit tasks that only exist to occupy two people without any added value.
Yeah, that is one thing it can do, but it’s not the only thing it does. I’m not sure how to get my point across well but, just because it gives you the wrong answer 25% of the time doesn’t mean it’s useless. In whatever you ask it to do, it often gets you most of the way there as long as you know how to correct it or check its work. The ability to ask specific or follow-up questions when learning something makes it invaluable as a learning tool (if you’re teaching yourself that is (ie. If you’re a university student)). It’s also very useful when brainstorming ideas or helping you approach a problem from a different angle. I can also ask it questions that are far more specific than what a search engine would get me.
It really comes down to if the human operator knows how (and when) to use it properly.
i never trust that information id rather learn with a book videos or just from websites if i ever use it i always fact check it. never blindly trust a computer
That was an integral part to this whole thing, you always fact check the ai. I said this in both of my comments.
I totally get preferring textbooks or videos though. I just find that the ai saves me time since i can ask specific questions about things, and it often gives me concise information that i understand more quickly.
Sort of. Nobody’s cutting corners on aviation structural components, for example. We’ve been pretty good at maximizing general value output, and usually that means lower quality, but not always.
I’ve asked ChatGPT to fix simple bugs that were in the code it’s given me. It usually doesn’t understand that there’s a problem and just rewrites it with different variables names.
It’s helped me a bit with resolving weird tomcat/Java issues when upgrading to RHEL8, though. It didn’t give me an answer, but it gave me ideas on where to look (in my case I didn’t realize fapolicyd replaced selinux)
That’s the point - you have the expertise to make proper sense of whatever it outputs. The people pushing for “AI” the most want to rely on it without any necessary expertise or just minimal efforts, like feeding it some of your financial reports and have generate a 5-year strategy only to fail miserably and have no one to blame this time (will still blame anyone else but themselves btw).
It’s not the most useless tool in the world by any means, but the mainstream talk is completely out of touch with reality on the matter, and so are mainstream actions (i.e. overrelying on it and putting way too much faith into it).
It helps a complete newbie like me get started and even learn while I do. Due to its restrictions and shortcoming, I’ve been having to learn how to structure and plan a project more carefully and thoughtfully, even creating design specs for programs and individual functions, all in order to provide useful prompts for ChatGPT to act on. I learn best by trial and error, with the ability to ask why things happened or are the way they are.
So, as a secondary teaching assistant, I think it’s very useful. But trying to use the API for ChatGPT 4 is…not worth it. I can easily blow through $20 in a few hours. So, I got a day and a half of use out of it before I gave up. :|
An unpopular opinion, I am sure, but if you’re a beginner with something - a new language, a new framework - and hate reading the docs, it’s a great way of just jumping into a new project. Like, I’ve been hacking away on a django web server for a personal project and it saved me a huge amount of time with understanding how apps are structured, how to interact with its settings, registering urls, creating views, the general development lifecycle of the project and the basic commands I need to do what I’m trying to do. God knows Google is a shitshow now and while Stackoverflow is fine and dandy (when it isn’t remarkably toxic and judgmental), the fact is that it cuts down on hours of fruitless research, assuming you’re not asking it to do anything genuinely novel or hyper-specific.
ChatGPT is hilariously incompetent… but on a serious note, I still firmly reject tools like copilot outside demos and the like because they drastically reduce code quality for short term acceleration. That’s a terrible trade-off in terms of cost.
I enjoy using copilot, but it is not made to think for you. It’s a better autocomplete, but don’t ever let it do more than a line at once.
Yup, AI is a tool, not a complete solution.
As a software engineer, the number of people I encounter in a given week who either refuse to or are incapable of understanding that distinction baffles and concerns me.
The problem I have with it is that all the time it saves me I have to use on reading the code. I probably spend more time on that as once in a while the code it produces is broken in a subtle way.
I see some people swearing by it, which is the opposite of my experience. I suspect that if your coding was copying code from stack overflow then it indeed improved your experience as now this process is streamlined.
deleted by creator
Same as ChatGPT is better web search.
id rather search the web than chatgpt because i fact check it anyway
I don’t know if it does yet, but if ChatGPT starts providing source for every information, then it would make it much faster to find the relevant information and check their sources, rather than clicking websites one by one.
Yep, ChatGPT4 allows optional calls to Bing now.
It used to have a problem with making a claims that were not relevant to or contradicted its own sources, but I don’t recall encountering that problem recently.
Biggest problem with it is that it lies with the exact same confidence it tells the truth. Or, put another way, it’s confidently incorrect as often as it is confidently correct - and there’s no way to tell the difference unless you already know the answer.
it’s kinda hilarious to me because one of the FIRST things ai researchers did was get models to identify things and output answers together with the confidence of each potential ID, and now we’ve somehow regressed back from that point
did we really regress back from that?
i mean giving a confidence for recognizing a certain object in a picture is relatively straightforward.
But LLMs put together words by their likeliness of belonging together under your input (terribly oversimplified).the confidence behind that has no direct relation to how likely the statements made are true. I remember an example where someone made chatgpt say that 2+2 equals 5 because his wife said so. So chatgpt was confident that something is right when the wife says it, simply because it thinks these words to belong together.
ChatGPT what is the Gödel number for the proof of 2+2=5?
But LinkedIn bros and corporate people are gonna gobble it up anyways because it has the right buzzwords (including “AI”) and they can squeeze more (low quality) work from devs to brag about how many things they (the corporate owners) are doing.
It’s just a fad. There’s just a small bit that will stay after the hype is gone. You know, like blockchain, AR, metaverse, NFT and whatever it was before that. In a few years there will be another breakthrough with it and we’ll hear from it again for a short while, but for now it’s just a one trick pony.
I always forget Metaverse is a thing.
deleted by creator
It was never a thing.
deleted by creator
Is there really any utility for blockchain and NFTs?
The hilarious thing about blockchain is that the core concept is actively making the whole thing worse. The matrix protocol is sort of essentially blockchain without the decentralized ledger part, and it’s so vastly superior in every single way.
NFTs just show how fundamentally dumb blockchains are, if you skip the decentralized ledger bit then you never need to invent NFT functionality in the first place…
Well, if you stop listening to people who think it’s a way to get really rich really fast (which it obviously isn’t), cryptocurrencies are quite useful. International transfers are so much cheaper and easier with them.
Yes, you could, for example, use it to manage who is allowed to park in a garage, anonymously. The owner of a parking spot NFT can unlock the door from the outside. Stuff like that.
However, it’s also possible to do that with a small web application. Just payments and transfer of the parking spots are less free and it’s not decentralized.
I disagree, because unlike those things, ai actually has a use case. It needs a human supervisor and it isn’t always faster, but chat gpt has been the best educational resource I’ve ever had next to YouTube. It’s also decent at pumping out a lot of lazy work and writing so i don’t have to, or helping me break down a task into smaller parts. As long as you’re not expecting it to solve all your problems for you, it’s an amazing tool.
People said the same things about 3d printing and yeah, while it can’t create literally everything at industrial scale, and it’s not going to see much consumer use, it has found a place in certain applications like prototyping and small scale production.
ChatGPT is OK at summarizing popular, low specificity topics that tons of people have already written a ton about, but it’s terrible at anything else. When I tested its knowledge about the process of a niche interest of mine (fabric dyeing) it skipped completely over certain important pieces of information, and when I prompted it to include them it basically just mirrored my prompt back at me.
Which has pretty much summed up my ChatGPT experience: it just regurgitates stuff I can find myself, but removes the ability to determine if the source is reliable. And if it’s something I’m already having trouble finding detailed information about it usually doesn’t help.
i’d argue that chat gpt is mostly great at taking human bullshit tasks from humans, who dontwant to dothe bullshit, like regurgitating the text from a textbook in different words, writing cover letters for job applications, that are often machine analyzed for buzzwords anyways.
So its use case only exist in the domain of bullshit tasks that only exist to occupy two people without any added value.
Yeah, that is one thing it can do, but it’s not the only thing it does. I’m not sure how to get my point across well but, just because it gives you the wrong answer 25% of the time doesn’t mean it’s useless. In whatever you ask it to do, it often gets you most of the way there as long as you know how to correct it or check its work. The ability to ask specific or follow-up questions when learning something makes it invaluable as a learning tool (if you’re teaching yourself that is (ie. If you’re a university student)). It’s also very useful when brainstorming ideas or helping you approach a problem from a different angle. I can also ask it questions that are far more specific than what a search engine would get me.
It really comes down to if the human operator knows how (and when) to use it properly.
i never trust that information id rather learn with a book videos or just from websites if i ever use it i always fact check it. never blindly trust a computer
That was an integral part to this whole thing, you always fact check the ai. I said this in both of my comments.
I totally get preferring textbooks or videos though. I just find that the ai saves me time since i can ask specific questions about things, and it often gives me concise information that i understand more quickly.
Yeah, they think it can turn a beginner dev into an advanced dev, but really it’s more like having a team of beginner devs.
It’s alright for translation. As an intermediate dev, being able to translate knowledge into languages I’m not as familiar with is nice.
Oh boy do I have news for you, that’s basically the only thing middle managers care about, short tem acceleration
I’m still convinced that GitHub copilot is actively violating copyleft violences. If not in word, then in the spirit.
Removed by mod
Western society is built on this principle
Tell me about it…
I left my more mature company for a startup.
I feel like Tyler Durden sometimes.
How you liking it? How many years have you aged in the months working at your startup?
Sort of. Nobody’s cutting corners on aviation structural components, for example. We’ve been pretty good at maximizing general value output, and usually that means lower quality, but not always.
I’ve asked ChatGPT to fix simple bugs that were in the code it’s given me. It usually doesn’t understand that there’s a problem and just rewrites it with different variables names.
It’s helped me a bit with resolving weird tomcat/Java issues when upgrading to RHEL8, though. It didn’t give me an answer, but it gave me ideas on where to look (in my case I didn’t realize fapolicyd replaced selinux)
That’s the point - you have the expertise to make proper sense of whatever it outputs. The people pushing for “AI” the most want to rely on it without any necessary expertise or just minimal efforts, like feeding it some of your financial reports and have generate a 5-year strategy only to fail miserably and have no one to blame this time (will still blame anyone else but themselves btw).
It’s not the most useless tool in the world by any means, but the mainstream talk is completely out of touch with reality on the matter, and so are mainstream actions (i.e. overrelying on it and putting way too much faith into it).
Yeah, it can speed up the process but you still have to know how to do it yourself when it inevitably messes something up.
It helps a complete newbie like me get started and even learn while I do. Due to its restrictions and shortcoming, I’ve been having to learn how to structure and plan a project more carefully and thoughtfully, even creating design specs for programs and individual functions, all in order to provide useful prompts for ChatGPT to act on. I learn best by trial and error, with the ability to ask why things happened or are the way they are.
So, as a secondary teaching assistant, I think it’s very useful. But trying to use the API for ChatGPT 4 is…not worth it. I can easily blow through $20 in a few hours. So, I got a day and a half of use out of it before I gave up. :|
An unpopular opinion, I am sure, but if you’re a beginner with something - a new language, a new framework - and hate reading the docs, it’s a great way of just jumping into a new project. Like, I’ve been hacking away on a django web server for a personal project and it saved me a huge amount of time with understanding how apps are structured, how to interact with its settings, registering urls, creating views, the general development lifecycle of the project and the basic commands I need to do what I’m trying to do. God knows Google is a shitshow now and while Stackoverflow is fine and dandy (when it isn’t remarkably toxic and judgmental), the fact is that it cuts down on hours of fruitless research, assuming you’re not asking it to do anything genuinely novel or hyper-specific.