I’ve said it before and I’ll say it again: Large Language Models (what has been passing for AI recently) are incredibly stupid. They do not know anything. All they can do is look to see which words frequently appear near each other and stick them into a programmed capacity for grammar to imitate thought. Obviously countering Drax isn’t a real problem you’re having, but just saying this for everybody’s benefit: DO NOT USE CHATGPT OR ANY OTHER AI SYSTEM TO ANSWER YOUR QUESTIONS.
Gotta rely on your fellow human community members instead for advice it looks like haha
I’ve said it before and I’ll say it again: Large Language Models (what has been passing for AI recently) are incredibly stupid. They do not know anything. All they can do is look to see which words frequently appear near each other and stick them into a programmed capacity for grammar to imitate thought. Obviously countering Drax isn’t a real problem you’re having, but just saying this for everybody’s benefit: DO NOT USE CHATGPT OR ANY OTHER AI SYSTEM TO ANSWER YOUR QUESTIONS. I would argue they do know things, but what they know is subtle. When you ask them a question, what they know is "how might a person answer the question: fill in the blank." What that means exactly is an unresolved debate.The real problem is that the data they are trained on does not, in general, contain the null result. In other words, if you were to ask a million people to answer the question "what is the capital of Nebraska" you'd get a lot of right answers, a bunch of wrong answers, and a bunch of people who would simply not answer because they don't know.LLMs cannot be trained on the null response. They never learn that the correct answer is ever: I don't know.That's why they hallucinate. They are essentially mimicking the behavior of people who always answer something, even when they don't know or are wrong. So of course, when they don't know they make up an answer, because their training says there is always an answer.You could also argue that this is not the fault of the LLMs directly, but rather a design decision built into the way the system generates results. LLMs invented the temperature parameter to tune the output to be less robotic, but no one has seriously tried to invent a parameter or set of parameters to quantify when the LLM "doesn't know." They are designed to be conversational, not cautious.
I’ve said it before and I’ll say it again: Large Language Models (what has been passing for AI recently) are incredibly stupid. They do not know anything. All they can do is look to see which words frequently appear near each other and stick them into a programmed capacity for grammar to imitate thought. Obviously countering Drax isn’t a real problem you’re having, but just saying this for everybody’s benefit: DO NOT USE CHATGPT OR ANY OTHER AI SYSTEM TO ANSWER YOUR QUESTIONS. I would argue they do know things, but what they know is subtle. When you ask them a question, what they know is "how might a person answer the question: fill in the blank." What that means exactly is an unresolved debate.The real problem is that the data they are trained on does not, in general, contain the null result. In other words, if you were to ask a million people to answer the question "what is the capital of Nebraska" you'd get a lot of right answers, a bunch of wrong answers, and a bunch of people who would simply not answer because they don't know.LLMs cannot be trained on the null response. They never learn that the correct answer is ever: I don't know.That's why they hallucinate. They are essentially mimicking the behavior of people who always answer something, even when they don't know or are wrong. So of course, when they don't know they make up an answer, because their training says there is always an answer.You could also argue that this is not the fault of the LLMs directly, but rather a design decision built into the way the system generates results. LLMs invented the temperature parameter to tune the output to be less robotic, but no one has seriously tried to invent a parameter or set of parameters to quantify when the LLM "doesn't know." They are designed to be conversational, not cautious. Meanwhile, computer science teachers i know: ah yes, so a cpu is where all the electrical stuff is innit
I’ve said it before and I’ll say it again: Large Language Models (what has been passing for AI recently) are incredibly stupid. They do not know anything. All they can do is look to see which words frequently appear near each other and stick them into a programmed capacity for grammar to imitate thought. Obviously countering Drax isn’t a real problem you’re having, but just saying this for everybody’s benefit: DO NOT USE CHATGPT OR ANY OTHER AI SYSTEM TO ANSWER YOUR QUESTIONS. I mean, chatgpt is really good for academics, and definitely good for coding if you atleast have basic idea bout the language/framework you're working on. It can't answer niche topics like a specific mobile game (Mcoc here) since they train on the data user gives (I believe GPT4 can browse the net but still the niche info is not available on the internet either).
I’ve said it before and I’ll say it again: Large Language Models (what has been passing for AI recently) are incredibly stupid. They do not know anything. All they can do is look to see which words frequently appear near each other and stick them into a programmed capacity for grammar to imitate thought. Obviously countering Drax isn’t a real problem you’re having, but just saying this for everybody’s benefit: DO NOT USE CHATGPT OR ANY OTHER AI SYSTEM TO ANSWER YOUR QUESTIONS. I mean, chatgpt is really good for academics, and definitely good for coding if you atleast have basic idea bout the language/framework you're working on. It can't answer niche topics like a specific mobile game (Mcoc here) since they train on the data user gives (I believe GPT4 can browse the net but still the niche info is not available on the internet either). I refuse to believe that people think that GPT doesn’t work for academics. It may not handle expert or complex questions but it can do enough that helps with a lot of stuff
Wonder if you asked specifically for a “Champ Released in 2024”, whether it would have given more precise (year-release) candidates.Instead of it interpreting the question as “in 2024, what is a good champ….” or “what character in 2024”, which neither specifically say you only want one that was released in 2024.
Like I said before, people out here expecting something like JARVIS or KITT (eighties kids represent) and what they're getting is something like JARVIS or KITT after a lobotomy. A lot of this AI stuff is a load of garbage and it's bad for the environment because it needs loads of power which is wasting resources, it means computers run for longer, meaning they wear out quicker meaning more waste when they're junked. I know there's been some good come of it, but this chatGPT stuff is just absolute pony and trap.
OP, any chance you are between 14-17 years old?