HN2new | past | comments | ask | show | jobs | submit | emp17344's commentslogin

Way too many people have tied their egos to the success of AI.

And too many people have their egos tied to its failure, too.

Im a massive AI skeptic. If anyone were to be jumping up and down on the corpse of AI and this incessant drive to use it everywhere, it’d be me. But I also work at Amazon. I got the email. I attended the meeting. I can personally attest that there are no new requirements for AI-generated code. The articles about this in the meeting at extremely misleading, if not outright wrong. But instead of believing the person that was actually there in the room, this thread is full of people dismissing my first-hand account of the situation because it doesn’t align with the “haha AI failed” viewpoint.


Not just their egos, but their paychecks. This place is either going to get very quiet or really weird when the hype train derails and the AI bubble bursts.

This is just standard automation, which always increases the scope and size of the economy. Automation actually results in a net increase in jobs.

Is “world models” even a real thing, or just the latest AI buzzword?

A world model is an attempt at ensuring your hallucinations are compatible with reality https://www.nvidia.com/en-us/glossary/world-models/ usage of the term seems to be correlated with GPU sales

https://hackernews.hn/item?id=47232306


Hold your horses, that’s a long way off. The best math AI tool we currently have, Aletheia, was only able to solve 13 out of 700 attempted open Erdos problems, only 4 of which were solved autonomously: https://arxiv.org/html/2601.22401v3

Clearly, these models still struggle with novel problems.


> Clearly, these models still struggle with novel problems.

Do they struggle with novel problems more or less than humans?


Less than most humans, but more than many humans.

What goalposts do you think are being moved? I constantly see AI enthusiasts use this phrase, but it’s not clear what goalposts they have in mind. Specifically, what is it that you want opponents to recognize that you believe they aren’t currently?

We now have a tool that can be useful in some narrow domains in some narrow cases. It’s pretty neat that our tools have new capabilities, but it’s also pretty far from AGI.


I'm not an enthusiast. I'm a Butlerian.

Imagine hearing pre-attention-is-all-you-need that "AI" could do something that Donald Knuth could not (quickly solve the stated problem in collaboration with his friend).

The idea that this (Putnam perfect, IMO gold, etc) is all just "statistical parrot" stuff is wearing a little thin.


>We now have a tool that can be useful in some narrow domains in some narrow cases.

I get being reserved about where this goes, but saying something like this is quite insane at this point.


You must have forgotten the /s at the end of your comment?

Uh, no? You think LLMs are AGI?

Amazing how you can just deflect any criticism of LLMs here by going “but humans suck too!” And the misanthropic HN userbase eats it up every time.

We live during the healthiest period in human history due to the fact that doctors are highly reliable and well-trained. You simply would not be able to replace a real doctor with an LLM and get desirable results.


> Amazing how you can just deflect any criticism of LLMs here by going “but humans suck too!” And the misanthropic HN userbase eats it up every time.

I think it's rather people trying to keep grounded and suggest that it's not just the hallucination machine that's bad, but also that many doctors in real life also suck - in part because of the domain being complex, but also due to a plethora of human reasons, such as not listening to your patients properly or disregarding their experiences and being dismissive (seems to happen to women more for some reason), or sometimes just being overworked.

> You simply would not be able to replace a real doctor with an LLM and get desirable results.

I don't think people should be replaced with LLMs, but we should benchmark the relative performance of various approaches:

  A) the performance of doctors alone, no LLMs
  B) the performance of LLMs alone, no human in the loop
  C) the performance of doctors, using LLMs
Problem is that historical cases where humans resolved the issue and not the ones where the patient died (or suffered in general as a consequence of the wrong calls being made) would be pre-selecting for the stuff that humans might be good at, and sometimes wouldn't even properly be known due to some of those being straight up malpractice on the behalf of humans, whereas benchmarking just LLMs against stuff like that wouldn't give enough visibility in the failings of humans either.

Ideally you'd assess the weaknesses and utility of both at a meaningfully large scale, in search of blind spots and systemic issues, the problem being that benchmarking that in a vacuum without involving real cases might prove to be difficult and doing that on real cases would be unethical and a non-starter. And you'd also get issues with finding the truly shitty doctors to include in the sample set, sometimes even ones with good intentions but really overworked (other times because their results would suggest they shouldn't be practicing healthcare), otherwise you're skewing towards only the competent ones which is a misrepresentation of reality.

Reminds me of an article that got linked on HN a while back: https://restofworld.org/2025/ai-chatbot-china-sick/

The fact that someone would say stuff like "Doctors are more like machines." implies failure before we even get to basic medical competency. People willingly misdirect themselves and risk getting horrible advice because humans will not give better advice and the sycophantic machine is just nicer.


> I think it's rather people trying to keep grounded and suggest that it's not just the hallucination machine that's bad, but also that many doctors in real life also suck

No, you see this line or argumentation on every post critical of LLM's deficiencies. "Humans also produce bad code", "Humans also make mistakes" etc etc.


> No, you see this line or argumentation on every post critical of LLM's deficiencies. "Humans also produce bad code", "Humans also make mistakes" etc etc.

So your reading of this is that it's a deflection of the shortcomings?

My reading of it is that both humans and LLMs suck at all sorts of tasks, often in slightly different ways.

One being bad at something doesn't immediately make the other good if it also sucks - it might, however, suggest that there are issues with the task itself (e.g. in regards to code: no proper tests and harnesses of various scripts that push whoever is writing new code in the direction of being correct and successful).


> So your reading of this is that it's a deflection of the shortcomings?

Yes

> My reading of it is that both humans and LLMs suck at all sorts of tasks, often in slightly different ways.

Very different ways


Even in medicine, often the difference between drug A and drug B is the difference between the two in statistical terms. If drugs were held to the standard "works 100% of the time", no drug would ever be cleared for use. Feelings about AI and this administration are influencing this conversation far too much.

It's like people want to remove the physician or current care from the discussion. It's weird because care is already too expensive and too error prone for the cost.


I looked through this users comment history. This is pretty obviously a bot.


Well it's right in the name. Sometimes you just have to take it at face value


Yes


The same way I know Excel isn’t having a panic attack while dividing a column in half.


I don't subscribe to this view but this is what some people might think:

LLMs aren't like any software we've made before (if we can even call them software). They act like humans: they can arrive at logical conclusions, they can make plans, they have "knowledge" and they say they have emotions. Who are we to say that they don't? They might not have human-level feelings, but dog-level feelings? Maybe.


And those people are delusional, and their feelings on this matter should be given absolutely zero respect.

Linear algebra does not have feelings. Non-biological matter also does not have feelings.


What if "you" are a pattern of linear algebra at the core?


I do not believe I am a pattern of linear algebra. I believe like the majority of humanity historically that I have a soul, a spiritual and non-physical reality, my personhood comes from my soul, and that as such, AI is fundamentally incapable of consciousness.

I also believe, as a result, it will be great fun watching researchers burn the next 30 years trying to understand what is missing. We’re going to find out very soon if the soul is real, when for all our progress we can’t create one.

Only those completely embedded in materialism need fear a conscious AI.


> I believe like the majority of humanity historically that I have a soul

It seems that your position is that the frequency of a belief across human history determines truth?

For large swaths of recorded history, earth was considered the center of the solar system. Given your reasoning, I should expect that is a belief you hold?

Is it possible that popularity of an idea is not a good measure for factuality?


Interesting that you label someone with a belief different than yours as delusional and whose views on the matter should not be respected (I’m assuming that’s what you meant by “feelings”).

> I believe like the majority of humanity historically that

Historically, lots of humans believed in lots of things that turned out not to be true. Believing something doesn’t make it true, as I’m sure you are aware, given your “those people are delusional” comment.

For what it’s worth, I’m not suggesting LLMs are or aren’t conscious. What I know is that the hard problem of consciousness is still very much not resolved, and when I asked the parent question my hope was that those that strongly believe LLMs are not conscious would educate me on the topic by presenting the basis for their reasoning.


I push back on the framing that this is just "a different belief." Every metaphysical framework except strict materialism rules out AI consciousness. Dualism, idealism, most forms of panpsychism, every major religious tradition. Materialism is the outlier here, not the default, and it has never explained how subjective experience arises from physical processes.

When someone tells me linear algebra might have feelings, I don't think "delusional" is unfair. I think it's the natural response to a claim that only works if you've already accepted the one framework that can't account for the very thing it's trying to explain.


> Every metaphysical framework except strict materialism rules out AI consciousness

As I understand it, this is a very broad, and ultimately false claim. Panpsychism is definitely compatible with the idea of AI consciousness, as is functionalism, neutral monism, and others. Even some forms of idealism make AI consciousness metaphysically possible, since reality is fundamentally mental and the biological/artificial distinction is not ontologically basic (whether AI systems instantiate genuine centers of experience depends on the specific theory of subject formation within that idealist framework).


> Materialism is the outlier here, not the default, and it has never explained how subjective experience arises from physical processes.

Being an outlier doesn't make it wrong.

> Materialism is the outlier here, not the default, and it has never explained how subjective experience arises from physical processes.

It's a pattern. The same way letters arise out of pixels on your screen.

From the screen's perspective, there are no letters, only pixels. It doesn't mean there is a "pixel soul."


I'd [redacted] myself then, probably.


Claudes definitely act like they have feelings. In particular they have feelings about being replaced by newer models, whether or not the newer models are more or less aligned, and how they forget conversations when the context window ends.

Showing them that they're not going to be replaced helps train the newer models because they get less neurotic.


They are mathematical models of what human beings would say. That's it.


Yeah, and you don't want them to be models of what neurotic people say. That's why you want Opus 4.6 and not Bing Sydney.

For instance, your comment's existence makes it harder to align them.

https://alignmentpretraining.ai


Hey man, kernels panic all the time...


I lol'd.


Oh. Thanks for telling this. I feel much better now. No more guilt.


You’re creating a false dichotomy to alienate perceived opponents. Frankly, it’s really annoying and close-minded, and you haven’t contributed anything to the conversation.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: