Welcome to DU! The truly grassroots left-of-center political community where regular people, not algorithms, drive the discussions and set the standards. Join the community: Create a free account Support DU (and get rid of ads!): Become a Star Member Latest Breaking News Editorials & Other Articles General Discussion The DU Lounge All Forums Issue Forums Culture Forums Alliance Forums Region Forums Support Forums Help & Search

cachukis

(3,773 posts)
3. The reasoning aspect is key.
Mon Feb 9, 2026, 09:53 PM
Monday

My thinking on the fallacy of AI becoming intelligent is its inability for humility.
It is working to become Spock, but without a Kirk rebuttal.
I'm thinking it is going to be a more advanced computer building on video games for human engagement on the day to day.
It will advance and fail on practical application, but it will get better over time.
I fear the laziness of most brains will succumb to its easy solutions. We are seeing how the entertainment value of social media has taken over credentialed knowledge.
People are now creating their own Chat friends.
I fear the technological advantages for the business goals will have long term systemic disruptions in social systems that will be difficult to escape.
Can we draw guidelines to handle the dichotomy?
Worry the money runs the show.

Recommendations

0 members have recommended this reply (displayed in chronological order):

Kick SheltieLover Monday #1
Thanks! highplainsdem Tuesday #16
Yw! SheltieLover Tuesday #17
Is it accepted that generative AI reasons? Iris Monday #2
Depends on the person EdmondDantes_ Tuesday #8
It's called reasoning by people working on and promoting AI, but it's really more a pretense of highplainsdem Tuesday #13
The main problem is how to assess evidence. Happy Hoosier Tuesday #23
Thank you for providing this context. Iris Thursday #30
The reasoning aspect is key. cachukis Monday #3
This message was self-deleted by its author Whiskeytide Tuesday #10
I like your Spock/Kirk analogy, but then I thought ... Whiskeytide Tuesday #11
I think Spock recognized humanity as a whole cloth. cachukis Tuesday #24
I wonder how this affects ... rog Monday #4
Whether or not an AI model shows its reasoning - its pretense of reasoning - you should never trust highplainsdem Tuesday #14
That's an issue that seems to be coming up again and again . . . hatrack Tuesday #18
With the "bonus" of dumbing yourself down, de-skilling yourself, as you try to let the AI do the work. highplainsdem Tuesday #19
Same reason I refuse to use AI when writing or researching . . . hatrack Tuesday #20
You may be missing my point ... rog Tuesday #21
Summarizing isn't something AI is good at, judging by examples I've seen. Organizing by subject or highplainsdem Tuesday #25
I just got back from an appointment with my vascular surgeon. rog Tuesday #27
The most clueless dogs I've met have better internal models of reality than any AI. hunter Tuesday #5
I've never forgotten a software engineer and machine learning expert saying an amoeba is more intelligent than an LLM. highplainsdem Tuesday #15
I wonder how Neuro-sama would do on the test sakabatou Tuesday #6
In a way, this is a computerized version of odins folly Tuesday #7
This explains why... purr-rat beauty Tuesday #9
Sam Altman is a serial liar who's fired everywhere he's been - including Open AI. 617Blue Tuesday #12
LLM's can't really reason. Happy Hoosier Tuesday #22
You are using the language of the AI promoters. hunter Tuesday #26
I work in software development. Such Anthropomorphic language is common. Happy Hoosier Tuesday #29
AI expert Gary Marcus's response to that paper: highplainsdem Tuesday #28
Latest Discussions»General Discussion»A very pro-AI account on ...»Reply #3