I’ve said it before and I’ll say it again: Large Language Models (what has been passing for AI recently) are incredibly stupid. They do not know anything. All they can do is look to see which words frequently appear near each other and stick them into a programmed capacity for grammar to imitate thought.
Obviously countering Drax isn’t a real problem you’re having, but just saying this for everybody’s benefit: DO NOT USE CHATGPT OR ANY OTHER AI SYSTEM TO ANSWER YOUR QUESTIONS.
I’ve said it before and I’ll say it again: Large Language Models (what has been passing for AI recently) are incredibly stupid. They do not know anything. All they can do is look to see which words frequently appear near each other and stick them into a programmed capacity for grammar to imitate thought.
Obviously countering Drax isn’t a real problem you’re having, but just saying this for everybody’s benefit: DO NOT USE CHATGPT OR ANY OTHER AI SYSTEM TO ANSWER YOUR QUESTIONS.
That’s what resonated with me. People ask these systems a question and assume the response is valid. In this case, we can see how wrong it is but if you asked about a topic you knew nothing about you could get misled very easily.
I’ve said it before and I’ll say it again: Large Language Models (what has been passing for AI recently) are incredibly stupid. They do not know anything. All they can do is look to see which words frequently appear near each other and stick them into a programmed capacity for grammar to imitate thought.
Obviously countering Drax isn’t a real problem you’re having, but just saying this for everybody’s benefit: DO NOT USE CHATGPT OR ANY OTHER AI SYSTEM TO ANSWER YOUR QUESTIONS.
I mean, chatgpt is really good for academics, and definitely good for coding if you atleast have basic idea bout the language/framework you're working on. It can't answer niche topics like a specific mobile game (Mcoc here) since they train on the data user gives (I believe GPT4 can browse the net but still the niche info is not available on the internet either).
I’ve said it before and I’ll say it again: Large Language Models (what has been passing for AI recently) are incredibly stupid. They do not know anything. All they can do is look to see which words frequently appear near each other and stick them into a programmed capacity for grammar to imitate thought.
Obviously countering Drax isn’t a real problem you’re having, but just saying this for everybody’s benefit: DO NOT USE CHATGPT OR ANY OTHER AI SYSTEM TO ANSWER YOUR QUESTIONS.
I would argue they do know things, but what they know is subtle. When you ask them a question, what they know is "how might a person answer the question: fill in the blank." What that means exactly is an unresolved debate.
The real problem is that the data they are trained on does not, in general, contain the null result. In other words, if you were to ask a million people to answer the question "what is the capital of Nebraska" you'd get a lot of right answers, a bunch of wrong answers, and a bunch of people who would simply not answer because they don't know.
LLMs cannot be trained on the null response. They never learn that the correct answer is ever: I don't know.
That's why they hallucinate. They are essentially mimicking the behavior of people who always answer something, even when they don't know or are wrong. So of course, when they don't know they make up an answer, because their training says there is always an answer.
You could also argue that this is not the fault of the LLMs directly, but rather a design decision built into the way the system generates results. LLMs invented the temperature parameter to tune the output to be less robotic, but no one has seriously tried to invent a parameter or set of parameters to quantify when the LLM "doesn't know." They are designed to be conversational, not cautious.
I’ve said it before and I’ll say it again: Large Language Models (what has been passing for AI recently) are incredibly stupid. They do not know anything. All they can do is look to see which words frequently appear near each other and stick them into a programmed capacity for grammar to imitate thought.
Obviously countering Drax isn’t a real problem you’re having, but just saying this for everybody’s benefit: DO NOT USE CHATGPT OR ANY OTHER AI SYSTEM TO ANSWER YOUR QUESTIONS.
I would argue they do know things, but what they know is subtle. When you ask them a question, what they know is "how might a person answer the question: fill in the blank." What that means exactly is an unresolved debate.
The real problem is that the data they are trained on does not, in general, contain the null result. In other words, if you were to ask a million people to answer the question "what is the capital of Nebraska" you'd get a lot of right answers, a bunch of wrong answers, and a bunch of people who would simply not answer because they don't know.
LLMs cannot be trained on the null response. They never learn that the correct answer is ever: I don't know.
That's why they hallucinate. They are essentially mimicking the behavior of people who always answer something, even when they don't know or are wrong. So of course, when they don't know they make up an answer, because their training says there is always an answer.
You could also argue that this is not the fault of the LLMs directly, but rather a design decision built into the way the system generates results. LLMs invented the temperature parameter to tune the output to be less robotic, but no one has seriously tried to invent a parameter or set of parameters to quantify when the LLM "doesn't know." They are designed to be conversational, not cautious.
Meanwhile, computer science teachers i know: ah yes, so a cpu is where all the electrical stuff is innit
I’ve said it before and I’ll say it again: Large Language Models (what has been passing for AI recently) are incredibly stupid. They do not know anything. All they can do is look to see which words frequently appear near each other and stick them into a programmed capacity for grammar to imitate thought.
Obviously countering Drax isn’t a real problem you’re having, but just saying this for everybody’s benefit: DO NOT USE CHATGPT OR ANY OTHER AI SYSTEM TO ANSWER YOUR QUESTIONS.
I would argue they do know things, but what they know is subtle. When you ask them a question, what they know is "how might a person answer the question: fill in the blank." What that means exactly is an unresolved debate.
The real problem is that the data they are trained on does not, in general, contain the null result. In other words, if you were to ask a million people to answer the question "what is the capital of Nebraska" you'd get a lot of right answers, a bunch of wrong answers, and a bunch of people who would simply not answer because they don't know.
LLMs cannot be trained on the null response. They never learn that the correct answer is ever: I don't know.
That's why they hallucinate. They are essentially mimicking the behavior of people who always answer something, even when they don't know or are wrong. So of course, when they don't know they make up an answer, because their training says there is always an answer.
You could also argue that this is not the fault of the LLMs directly, but rather a design decision built into the way the system generates results. LLMs invented the temperature parameter to tune the output to be less robotic, but no one has seriously tried to invent a parameter or set of parameters to quantify when the LLM "doesn't know." They are designed to be conversational, not cautious.
Meanwhile, computer science teachers i know: ah yes, so a cpu is where all the electrical stuff is innit
To be fair, LLMs are 800 pounds of math with two coats of computer science paint. If your job title is "computer science teacher" LLMs might as well be voodoo magic.
Wonder if you asked specifically for a “Champ Released in 2024”, whether it would have given more precise (year-release) candidates.
Instead of it interpreting the question as “in 2024, what is a good champ….” or “what character in 2024”, which neither specifically say you only want one that was released in 2024.
I’ve said it before and I’ll say it again: Large Language Models (what has been passing for AI recently) are incredibly stupid. They do not know anything. All they can do is look to see which words frequently appear near each other and stick them into a programmed capacity for grammar to imitate thought.
Obviously countering Drax isn’t a real problem you’re having, but just saying this for everybody’s benefit: DO NOT USE CHATGPT OR ANY OTHER AI SYSTEM TO ANSWER YOUR QUESTIONS.
I mean, chatgpt is really good for academics, and definitely good for coding if you atleast have basic idea bout the language/framework you're working on. It can't answer niche topics like a specific mobile game (Mcoc here) since they train on the data user gives (I believe GPT4 can browse the net but still the niche info is not available on the internet either).
I refuse to believe that people think that GPT doesn’t work for academics. It may not handle expert or complex questions but it can do enough that helps with a lot of stuff
I’ve said it before and I’ll say it again: Large Language Models (what has been passing for AI recently) are incredibly stupid. They do not know anything. All they can do is look to see which words frequently appear near each other and stick them into a programmed capacity for grammar to imitate thought.
Obviously countering Drax isn’t a real problem you’re having, but just saying this for everybody’s benefit: DO NOT USE CHATGPT OR ANY OTHER AI SYSTEM TO ANSWER YOUR QUESTIONS.
I mean, chatgpt is really good for academics, and definitely good for coding if you atleast have basic idea bout the language/framework you're working on. It can't answer niche topics like a specific mobile game (Mcoc here) since they train on the data user gives (I believe GPT4 can browse the net but still the niche info is not available on the internet either).
I refuse to believe that people think that GPT doesn’t work for academics. It may not handle expert or complex questions but it can do enough that helps with a lot of stuff
True. I wrote 2 of my project papers and studied for all of my Undergraduate exams via chatgpt. I am currently using chatgpt to do most of my work in my internship. It's all depends on how you interact it with it, how you train it smartly to give the answers you want.
Wonder if you asked specifically for a “Champ Released in 2024”, whether it would have given more precise (year-release) candidates.
Instead of it interpreting the question as “in 2024, what is a good champ….” or “what character in 2024”, which neither specifically say you only want one that was released in 2024.
Like I said before, people out here expecting something like JARVIS or KITT (eighties kids represent) and what they're getting is something like JARVIS or KITT after a lobotomy.
A lot of this AI stuff is a load of garbage and it's bad for the environment because it needs loads of power which is wasting resources, it means computers run for longer, meaning they wear out quicker meaning more waste when they're junked. I know there's been some good come of it, but this chatGPT stuff is just absolute pony and trap.
Like I said before, people out here expecting something like JARVIS or KITT (eighties kids represent) and what they're getting is something like JARVIS or KITT after a lobotomy.
A lot of this AI stuff is a load of garbage and it's bad for the environment because it needs loads of power which is wasting resources, it means computers run for longer, meaning they wear out quicker meaning more waste when they're junked. I know there's been some good come of it, but this chatGPT stuff is just absolute pony and trap.
I’ve said it before and I’ll say it again: Large Language Models (what has been passing for AI recently) are incredibly stupid. They do not know anything. All they can do is look to see which words frequently appear near each other and stick them into a programmed capacity for grammar to imitate thought.
Obviously countering Drax isn’t a real problem you’re having, but just saying this for everybody’s benefit: DO NOT USE CHATGPT OR ANY OTHER AI SYSTEM TO ANSWER YOUR QUESTIONS.
I mean, chatgpt is really good for academics, and definitely good for coding if you atleast have basic idea bout the language/framework you're working on. It can't answer niche topics like a specific mobile game (Mcoc here) since they train on the data user gives (I believe GPT4 can browse the net but still the niche info is not available on the internet either).
I refuse to believe that people think that GPT doesn’t work for academics. It may not handle expert or complex questions but it can do enough that helps with a lot of stuff
Agree. I use it all the time for my studies, especially for biology and chemistry.
I actually never use chat gpt, in what way do yall use it for academics? Exam questions? Personally I just use websites that give past exams or past exam questions by topic and difficulty
Comments
Obviously countering Drax isn’t a real problem you’re having, but just saying this for everybody’s benefit: DO NOT USE CHATGPT OR ANY OTHER AI SYSTEM TO ANSWER YOUR QUESTIONS.
The real problem is that the data they are trained on does not, in general, contain the null result. In other words, if you were to ask a million people to answer the question "what is the capital of Nebraska" you'd get a lot of right answers, a bunch of wrong answers, and a bunch of people who would simply not answer because they don't know.
LLMs cannot be trained on the null response. They never learn that the correct answer is ever: I don't know.
That's why they hallucinate. They are essentially mimicking the behavior of people who always answer something, even when they don't know or are wrong. So of course, when they don't know they make up an answer, because their training says there is always an answer.
You could also argue that this is not the fault of the LLMs directly, but rather a design decision built into the way the system generates results. LLMs invented the temperature parameter to tune the output to be less robotic, but no one has seriously tried to invent a parameter or set of parameters to quantify when the LLM "doesn't know." They are designed to be conversational, not cautious.
A.I. only have data before some date.
And as @DNA3000 said sometimes is at random...
Instead of it interpreting the question as “in 2024, what is a good champ….” or “what character in 2024”, which neither specifically say you only want one that was released in 2024.
A lot of this AI stuff is a load of garbage and it's bad for the environment because it needs loads of power which is wasting resources, it means computers run for longer, meaning they wear out quicker meaning more waste when they're junked. I know there's been some good come of it, but this chatGPT stuff is just absolute pony and trap.