by greybox on 3/26/25, 4:54 PM with 138 comments
by Maro on 3/26/25, 6:30 PM
Ie. most managers can't help their team find a hard bug that is causing a massive outage.
Note: I'm a manager, and I spend a lot of time pondering how to spend my time, how to be useful, how to remain relevant, especially in this age of AI.
by vunderba on 3/26/25, 6:04 PM
The seductive promise of solving all your problems is the issue. By reaching for it to solve any problem at an almost instinctual level you are completely failing to cultivate an intrinsically valuable skill - that of critical reasoning.
That act of manipulating the problem in your head—critical thinking—is ultimately a craft. And the only way to become better at it is by practicing it in a deliberate, disciplined fashion.
This is why it's pretty baffling to me when I see attempts at comparing LLMs to the invention of the calculator. A calculator is still used IN SERVICE of a larger problem you are trying to solve.
by jonahx on 3/26/25, 7:21 PM
> In this paper, we aim to address this gap by conducting a survey of a professionally diverse set of knowledge workers ( = 319), eliciting detailed real-world examples of tasks (936) for which they use GenAI, and directly measuring their perceptions of critical thinking during these tasks
So, they asked people to remember times they used AI, and then asked them about their own perceptions about their critical thinking when they did.
How are we even pretending there is serious scientific discussion to be had about these "results"?
by oneofyourtoys on 3/26/25, 7:04 PM
Common activities provided by these gyms include fixing misconfigured printers, telling a virtual support customer to turn their PC off and back on again, and troubleshooting mysterious NVIDIA driver issues (the company has gone bankrupt 5 years ago, but their hardware is still in great demand for frustration tolerance training).
by sitkack on 3/26/25, 6:29 PM
> Abstract The rise of Generative AI (GenAI) in knowledge workflows raises questions about its impact on critical thinking skills and practices. We survey 319 knowledge workers to investigate 1) when and how they perceive the enaction of critical thinking when using GenAI, and 2) when and why GenAI affects their effort to do so. Participants shared 936 first-hand examples of using GenAI in work tasks. Quantitatively, when considering both task- and user-specific factors, a user’s task-specific self-confidence and confidence in GenAI are predictive of whether critical thinking is enacted and the effort of doing so in GenAI-assisted tasks. Specifically, higher confidence in GenAI is associated with less critical thinking, while higher self-confidence is associated with more critical thinking. Qualitatively, GenAI shifts the nature of critical thinking toward information verification, response integration, and task stewardship. Our insights reveal new design challenges and opportunities for developing GenAI tools for knowledge work.
It is be presented at CHI Conference https://chi2025.acm.org/
https://en.wikipedia.org/wiki/Conference_on_Human_Factors_in...
by lenerdenator on 3/26/25, 5:30 PM
Fortunately, my ADHD-addled brain doesn't need some fancy AI to make its cognition "Atrophied and Unprepared"; I can do that all on my own, thank you very much.
by greybox on 3/26/25, 4:54 PM
“[A] key irony of automation is that by mechanising routine tasks and leaving exception-handling to the human user, you deprive the user of the routine opportunities to practice their judgement and strengthen their cognitive musculature, leaving them atrophied and unprepared when the exceptions do arise,” the researchers wrote.
by pseudocomposer on 3/26/25, 6:38 PM
Basically, we might need to standardize 10-20% of work time being used to “keep up” automatable skills that once took up 80+% of work time in fields where AI-based automation is making things more efficient.
This could even be done within automation platforms themselves, and sold to their customers as an additional feature. I suspect/hope that most employers do not want to see these automatable skills atrophy in their employees, for the sake of long-term efficiency, even if that means a small reduction in short-term efficiency gains from automation.
by nopelynopington on 3/26/25, 6:54 PM
I've also turned to AI in side projects, and it's allowed me to create some very fast MVPs, but the code is worse than spaghetti - it's spaghetti mixed with the hair from the shower drain.
None of the things I've built are beyond my understanding, but I'm lazy and it doesn't seem worth the effort to use my brain to code.
Probably the most use my brain gets every day is wordle
by bentt on 3/26/25, 6:48 PM
by sollewitt on 3/26/25, 5:41 PM
Whereas, you can ask an LLM to speak to you in e.g. Spanish, about whatever topic you're interested in, and be able to stop and ask it to explain any idioms or vocabulary or grammar in English at any time.
I found this to be more like a "cognitive gym". Maybe we're just not using the tools beneficially.
by divtiwari on 3/27/25, 1:52 AM
by masfuerte on 3/26/25, 5:43 PM
by rraghur on 3/26/25, 6:41 PM
by tunesmith on 3/26/25, 8:30 PM
With using GenAI (and/or "being a manager") aren't they somewhat inversely related?
I find implementation level programmers to generally be poor at stating specifications. They often phrase problems in terms of lacking their desired solutions. They jump straight to implementation.
But a manager has to get skilled at giving specification: being clear about what they expect, without stating how to do it. And that's a skill that needs to be quickly developed to use GenAI well as well. I think getting good at specifying is definitely worthwhile, and I think GenAI is helping a lot of people get better at that quickly.
Overall, it seems that should very much be considered part of "critical thinking".
by piltdownman on 3/26/25, 5:35 PM
While much is made of the 'diminished skill for independent problem-solving' caused by over-reliance, is there a more salient KPI than some iteration of this 'Synthetic Thinking Effort' by which to baseline and optimise the cost/benefit of AI usage versus traditional cognition?
by labrador on 3/26/25, 7:18 PM
I don't depend on AI for anything. I am not doing corporate work. Could it be that what people are experiencing is that they are becoming less suitable for corporate work as AI and robots replace them? Isn't this a good thing? Shouldn't the focus be on using AI to bring out the innate talents of humans that aren't profit driven?
by tsumnia on 3/26/25, 6:57 PM
That said, the counter to my own counter is "do I really need to memorize that?" Yes yes no internet and I'm screwed... but that's such a rare edge case. I am able to quickly find the command and knowing that it is stored somewhere else may be enough knowledge for me rather than memorization. I can see Gen AI falling into a similar design, I don't need to know explicitly how to do something, just that that task can be resolved through an LLM prompt.
Granted, we're still trying to figure out how to communicate with LLMs and we only really have 3 years of experience. Most of our insights have come from blog posts and a handful of research articles. I agree that Gen AI laziness is a growing issue, but I don't think it needs to go full Idiocracy sensationalist headline.
by Qem on 3/26/25, 5:06 PM
by 1vuio0pswjnm7 on 3/27/25, 3:19 PM
"Impact of Gen AI on Critical Thinking: Reduction in Cognitive Effort, Confidence"
"Impact of AI on Critical Thinking: Self-Reported Reductions in Cognitive Effort"
"The Impact of Generative AI on Critical Thinking: Reductions in Cognitive Effort"
Actual title of the paper:
"The Impact of Generative AI on Critical Thinking: Self-Reported Reductions in Cognitive Effort and Confidence Effects From a Survey of Knowledge Workers"
Previous discussion:
10 Feb 2025 17:01:08 UTC https://news.ycombinator.com/item?id=43002458 (1 comment)
10 Feb 2025 22:31:05 UTC https://news.ycombinator.com/item?id=43006140 (0 comments)
11 Feb 2025 11:14:06 UTC https://news.ycombinator.com/item?id=43011483 (0 comments) [dead]
11 Feb 2025 14:13:36 UTC https://news.ycombinator.com/item?id=43012911 (1 comment)
12 Feb 2025 01:47:16 UTC https://news.ycombinator.com/item?id=43020846 (0 comments) [flagged] [dead]
14 Feb 2025 15:54:57 UTC https://news.ycombinator.com/item?id=43049676 (1 comment)
15 Feb 2025 12:06:01 UTC https://news.ycombinator.com/item?id=43057907 (101 comments)
by allenrb on 3/26/25, 7:17 PM
Right now I’m curious to see how long I can keep up with those using AI for more mundane assistance. So far, so good.
by gatinsama on 3/26/25, 6:14 PM
If you don't understand what's happening, you have no way to know if the system is working as intended. And understanding (and deciding) exactly how the system works is the really hard part for any sufficiently complex project.
by deepfriedchokes on 3/26/25, 11:29 PM
by 0x20cowboy on 3/26/25, 6:07 PM
I only tend to do that when I am tired or annoyed, but when I do it I can feel myself getting dumber. And it’s a weirdly satisfying feeling.
I just need a chair that doubles as a toilet and I’ll be all set.
by _heimdall on 3/26/25, 6:33 PM
Critical thinking is a skill that requires practice to improve at and maintain it. Using LLMs pushes the task that would require critical thinking off to something/someone else. Of course the user will get worse at critical thinking when they try to do it less often.
by jhallenworld on 3/26/25, 7:20 PM
You are tremendously better off getting a bad grade doing your own work than getting a good one using ChatGPT.
by moralestapia on 3/26/25, 5:34 PM
I also believe, however, that humans who are able to reason properly would become much more valuable, because of this same thing.
by DrNosferatu on 3/26/25, 6:23 PM
We have the cognition science to make it happen - or at least learn how to structure it.
by riffic on 3/26/25, 7:05 PM
by ChrisArchitect on 3/26/25, 6:04 PM
Some discussion on the study: https://news.ycombinator.com/item?id=43057907
by AISnakeOil on 3/26/25, 7:07 PM
by _aavaa_ on 3/27/25, 4:56 PM
by derefr on 3/26/25, 7:28 PM
I don't build or rely on pre-prompted agents to automate specific problems or workflows. Rather, I only rely on services like ChatGPT or Claude for their generic reasoning, chat, and "has read the entire web at some point" capabilities.
My use-cases break down into roughly equal thirds:
---
1. As natural-language, iteratively-winnowing-the-search-space versions of search engines.
Often, I want to know something — some information that's definitely somewhere out there on the web. But, from 30+ years of interacting with fulltext search systems, I know that traditional search engines have limitations in the sorts of queries that'll actually do anything. There are a lot of "objective, verifiable, and well-cited knowledge" questions that are just outside of the domain of Google search.
One common example of fulltext-search limitations, is when you know how to describe a thing you're imagining, a thing that may or may not exist — but you don't know the jargon term for it (if there even is one.) No matter how many words you throw at a regular search engine, they won't dredge up discussions about the thing, because discussions about the thing just use the jargon term — they don't usually bother to define it.
To find answers to these sorts of questions, I would have previously ask a human expert — either directly, or through a forum/chatroom/subreddit/Q&A site/etc.
But now, I've got a new and different kind of search engine — a set of pre-trained base models that, all by themselves, perform vaguely as RAGs over all of the world's public-web-accessible information.
Of course, an LLM won't have crystal clarity in its memory — it'll forget exact figures, forget the exact phrasing of quotations, etc. And if there's any way that it can be fooled or misled by some random thing someone made up somewhere on the web once, it will be.
But ChatGPT et al can sure tell me the right jargon term (or entire search query) to turn what was previously, to me, almost deep-web information, into public-web information.
---
2. As a (fuzzy-logic) expert system in many domains, that learned all its implications from the public information available on the web.
One fascinating thing about high-parameter-count pre-trained base models, is that you don't really need to do any prompting, or supply any additional information, to get them to do a vaguely-acceptable job of diagnosis — whether that be diagnosing your early-stage diabetic neuropathy, or that mysterious rattle in your car.
Sure, the LLM will be wrong sometimes. It's just a distillation of what a bunch of conversations and articles spread across the public web have to say about what are or aren't the signs and symptoms of X.
But those are the same articles you'd read. The LLM will almost always outperform you in "doing your own research" (unless you go as far as to read journal papers — I don't know of any LLM base model that's been trained on arXiv yet...). It won't be as good at medicine as a doctor, or as good at automotive repair as an automotive technician, etc. — but it will be better (i.e. more accurate) at those things than an interested amateur who's watched some YouTube videos and read some pop-science articles.
Which means you can just tell LLMs the "weird things you've noticed lately", and get it to hypothesize for you — and, as long as you're good at being observant, the LLM's hypotheses will serve as great lines of investigation. It'll suggest which experts or specialists you should contact, what tests you can perform yourself to do objective differential diagnostics, etc.
(I don't want to under-emphasize the usefulness of this. ChatGPT figured out my house had hidden toxic mold! My allergies are gone now!)
---
3. As a translator.
Large-parameter-count LLM base models are actually really, really good at translation. To the point that I'm not sure why Google Translate et al haven't been updated to be powered by them. (Google Translate was the origin of the Transformer architecture, yet it seems to have been left in the dust since then by the translation performance of generic LLMs.)
And by "translation", I do literally mean "translating entire documents from one spoken/written human language to another." (My partner, who is a fluently-bilingual writer of both English + [Traditional] Chinese, has been using Claude to translate English instructions / documents into Chinese for her [mostly monolingual Chinese] mother to better understand them; and to translate any free-form responses her mother is required to give, back into English. She used to do these tasks herself "by hand" — systems like Google Translate would provide results that were worse-than-useless. But my partner can verify that, at least for this language pair, modern LLMs are excellent translators, writing basically what she would write herself.)
But I also mean:
• The thing Apple markets as part of Apple Intelligence — translation between writing styles (a.k.a. "stylistic editing.") You don't actually need a LoRA / fine-tune to do this; large-parameter-count models already inherently know how to do it.
• Translating between programming languages. "Rewrite-it-in-Rust" is trivial now. (That's what https://www.darpa.mil/research/programs/translating-all-c-to... is about — trying to build up an agentive framework that relies on both the LLM's translation capabilities, and the Rust compiler's typing errors on declaration change, to brute-force iterate across entire codebases, RiiRing one module at a time, and then recursing to its dependents to rewrite them too.)
• Translating between pseudocode, and/or a rigorous description of code, and actual code. I run a data analytics company; I know far more about the intricacies of ANSI SQL than any man ought to. But even I never manage to remember the pile of syntax features that glom together to form a "loose index scan" query. (WITH RECURSIVE, UNION ALL, separate aliases for the tables used in the base vs inductive cases, and one of those aliases referenced in a dependent subquery... but heck if I recall which one.) I have a crystal-clear picture of what I want to do — but I no longer need to look up the exact grammar the SQL standard decided to use yet again, because now I can dump out, in plain language, my (well-formed) mental model of the query — and rely on the LLM to translate that model into ANSI SQL grammar.
by thombles on 3/26/25, 7:19 PM
by RecycledEle on 3/26/25, 10:48 PM
The same is true of managers. I have had managers who yelled at me to do things they did not understand. They rotted on the inside. Other managers learned every trick I brought to the company. They grew.
by sunjester on 3/26/25, 5:58 PM
by rraghur on 3/26/25, 6:48 PM
At the end, I spent probably more time and learnt nothing.. My initial take was that this is the kind of thing I don't care much for so giving it to a llm is OK... However, by the end of it I ended up more frustrated and lost it in the simulation of working things out aa well
by MrMcCall on 3/26/25, 6:43 PM
Comedians' ability diminishes as they take time off.
Ahnold wasn't lounging around all day.
We should understand that fixing crap, unsensible code is not a productive skillset. As Leslie Lamport said the other day, logically developing and coding out proper abstractions is the core skillset, and not one to be delegated to just anything or anyone.
It's ok; the bright side for folks like me is that you're just happily hamstringing yourselves. I've been trying to tell y'all, but I can only show y'all the water, not make you drink.
by AtomBalm on 3/26/25, 8:27 PM
by Deprogrammer9 on 3/26/25, 7:23 PM
by fbn79 on 3/26/25, 6:56 PM