Interesting spreadsheet example. I have the opposite problem, Gemini insists to bring up the option of placing stuff in my google workspace (most of the time spreadsheets) although I've never told it to.
> working with [chatbots] feels like groping through a cave in the dark – a horrible game I call "PromptQuest" – while being told this is improving my productivity.
It’s The Register. They’re always more about the ‘tude than substance.
This seems to be someone who has no idea how to use LLMs yelling at clouds. Or maybe just someone pretending to have no idea because it makes for good cloud-yelling.
I make no judgement, but I definitely have had the opposite experience of the author, therefore the article doesn't resonate and I don't even understand it.
The way I see it, when LLMs work, they're almost magical. When they don't, oh well, it didn't take that long anyway, and I didn't have them until recently, so I can just do things the old boring way if the magic fails.
The problem with zork is that you don’t have a list of all the options in front of you so you have to guess. You could have a menu that lists all the valid options, but that changes the game. It doesn’t require you to use imagination and open-ended thinking, it becomes more of a point’n’click storybook.
But for tools, we should have a clear up front list of capabilities and menu options. Photoshop and VScode give you menu after menu of options with explicit well defined behaviors because they are tools used to achieve a specific aim and not toys for open ended exploration.
An llm doesn’t give you a menu because the llm doesn’t even know what it’s capable of. And that’s why I think we can see such polarized responses - some people want an LLM that’s a supercharged version of a tool, others want a toy for exploration.
The only time it ever seems like magic is when you don't really care about the problem or how it gets "solved" and are willing to ignore all the little things it got wrong.
Generative AI is neither magic, nor does it really solve any problems. The illusion of productivity is all in your head.
This article is garbage. I was half expecting or hoping for a nuanced analysis of regressions manifested in a specific leading model as a result of purported "upgrades" but instead found an idiot who doesn't understand how LLMs work or seem to even care, really.
Idiots like this seem to want a robot that does things for them instead of a raw tool that builds sometimes useful context, and the LLM peddlers are destroying their creations to oblige this insatiable contingent.
A "robot that does things" is the overpromise that doesn't deliver.
I actually agree with the article that non-determinism is why generative AI is the wrong tool in most cases.
In the past, the non-determinism came from the user's inconsistent grammar and the game's poor documentation of its rigid rules. Now the non-determinism comes 100% from the AI no matter what the user does. This is objectively worse!
Interesting spreadsheet example. I have the opposite problem, Gemini insists to bring up the option of placing stuff in my google workspace (most of the time spreadsheets) although I've never told it to.
Gen X translator here. This is a user story complaint that product output is nondeterministic.
Anyone else completely confused about what this article is even about?
I actually read the article, so I'm not confused about what it is about.
I also understood that it's about Copilot not doing the thing the author wanted.
According to the article:
> working with [chatbots] feels like groping through a cave in the dark – a horrible game I call "PromptQuest" – while being told this is improving my productivity.
AI bad, AI bad, AI bad. bad bad bad, AI-bad.
It’s The Register. They’re always more about the ‘tude than substance.
This seems to be someone who has no idea how to use LLMs yelling at clouds. Or maybe just someone pretending to have no idea because it makes for good cloud-yelling.
From the reactions here, we can already infer we're dealing with user error.
I make no judgement, but I definitely have had the opposite experience of the author, therefore the article doesn't resonate and I don't even understand it.
Seems kinda like a first world problem to me.
The way I see it, when LLMs work, they're almost magical. When they don't, oh well, it didn't take that long anyway, and I didn't have them until recently, so I can just do things the old boring way if the magic fails.
The problem with zork is that you don’t have a list of all the options in front of you so you have to guess. You could have a menu that lists all the valid options, but that changes the game. It doesn’t require you to use imagination and open-ended thinking, it becomes more of a point’n’click storybook.
But for tools, we should have a clear up front list of capabilities and menu options. Photoshop and VScode give you menu after menu of options with explicit well defined behaviors because they are tools used to achieve a specific aim and not toys for open ended exploration.
An llm doesn’t give you a menu because the llm doesn’t even know what it’s capable of. And that’s why I think we can see such polarized responses - some people want an LLM that’s a supercharged version of a tool, others want a toy for exploration.
The only time it ever seems like magic is when you don't really care about the problem or how it gets "solved" and are willing to ignore all the little things it got wrong.
Generative AI is neither magic, nor does it really solve any problems. The illusion of productivity is all in your head.
This article is garbage. I was half expecting or hoping for a nuanced analysis of regressions manifested in a specific leading model as a result of purported "upgrades" but instead found an idiot who doesn't understand how LLMs work or seem to even care, really.
Idiots like this seem to want a robot that does things for them instead of a raw tool that builds sometimes useful context, and the LLM peddlers are destroying their creations to oblige this insatiable contingent.
A "robot that does things" is the overpromise that doesn't deliver.
I actually agree with the article that non-determinism is why generative AI is the wrong tool in most cases.
In the past, the non-determinism came from the user's inconsistent grammar and the game's poor documentation of its rigid rules. Now the non-determinism comes 100% from the AI no matter what the user does. This is objectively worse!