I use ChatGPT just for programming and it gives wrong answers half of the time.
i’m studying mechanical engineering and there’s a guy in our class who’s obsessed with chatgpt. he’s always trying to solve all of the tasks using chatgpt and he’s always the first to share the solution in zoom. so far it’s never been correct but he just sticks with it…
I am a mechanical engineer. I was able to get special permission from my IT department to use LLMs as part of my workflow as a genie pig for the department. It is completely useless.
One the most valuable skill an engineer can have is being able to communicate technical information effectively to different audiences. GPT is on overly polite meat grinder, spitting out half chewed technical slop.
AI will have a better sense of something like mechanical engineering when it’s inhabited a body for a while.
If I’m recalling right, I asked Chatgpt re banked turn with friction. Didn’t give the answer I was looking for.
I asked Chatgpt re the best big phones of 2022. 1 of the phones it cited was released in 2021.
Yeah I also do and it is indeed frequently incorrect. It is good when you have like no idea about what you’re doing. It can help you get on track and then you can research by yourself.
I use chatGPT for any topic I’m curious about, and like half the time when i double check the answers it turns out they’re wrong.
For example i asked for a list of phones with screens that don’t use PWM, and when i looked up the specs of the phones it recommended it turned out they all had PWM, even though in the chatGPT answer it explicitly stated that each of these phones don’t use PWM. Why does it straight up lie?!
It’s not lying. It has no concept of context or truth. Auto complete on steroids.
Not picking fights. Just curious.
Is this an improvement or a decline in your overall code programming success?
I am a hobbyist (and not very good) programmer, and while ChatGPT (free version) often gives me wrong answers, it still gives me some insight on how some stuff could be done (intentionally or not) or how something works and is actually somewhat helpful in learning stuff, but I guess this could be double-edged sword even in that regard.
It is also pretty good at detecting simple code errors, from what I have seen.
Overall more positive than negative, but I wouldn’t recommend to use it blindly.
Huge improvement in work flow.
Don’t get it to write your code for you, it’s not gonna work 3/10 times. Instead use it to review your code, help remove any code smells for refactoring.
I don’t use chatGPT, but work with colleagues who do. They’re productivity visibly drops and half the time I gotta fix their shitty code.
The other half though is sweet times.
You should give phind a try
phind
phind
People on ELI5 ask questions that can be answered with a single google search. Yet they do not do the google search. What makes yoi think they will do the bard or chatgpt?
Because if the second worst option is asking ELI5 something basic, then the worst thing is asking Al the same question and then getting the wrong answer. So they choose Al
During the early newspaper era, they would write the editor with basic questions like they were google.
Yeah I read a 30 year old newspaper a while back and it was like super high latency internet. Message board, posts, replies to posts, personals etc. None of that stuff makes it into newspapers anymore…
Isn’t it still a thing.
It is. I’ve seen ‘Write the Editor’ sections often in the magazines I check out from the library from time to time. IIRC: Popular Mechanics, Popular Science, and The New Yorker have one.
Newspapers? Depends who you ask
Jokes aside, I am not talking about the “write the editor” sections we see now. I am saying they’d use it like GOOGLE. You’re not going to see someone ask “what’s 32,344 divided by 7?” Or “who is the senator of idaho” In the new yorker.
I don’t get why people trust their answers so much. They lie. Confidently. Constantly.
I usually ask the GPT, then look up the topic myself based on terms and keywords that were mentioned
No! You’re meant to blindly trust every source of information you read, AI or not.
I don’t get why people completely disregard their usefulness because of that. Just don’t trust anything they say until you verify it. It’s still useful for exploration or to get enough of a grasp of something that you can figure it out on your own.
Interestingly, as ChatGPT might be trained on these ELI5 questions and as a result they are asked more infrequently, it might get worse over time or out of date on these types of questions by its own doing. I especially wonder how bad this influence will get on subjects that you’d normally search stackoverflow for.
What are some other communities that are less used now? WritingPrompts and PhotoshopBattles come to mind for me
Oh man. I forgot that PhotoshopBattles existed.
That’s a good question. I would guess that it’s lessened some. But both are for creative tasks vs explaining a proven topic, item, or thing.
Quota is in shambles
So, no change then?
All they need is a couple bots that rely on GPT for outputs and it’d be like nothing changed.
ELI5 how to use ChatGPT and Bard