Hacker News .hnnew | past | comments | ask | show | jobs | submit | sdevonoes's commentslogin

As long as AI is being introduced by multibillion dollar corporations, it’s all a trick, a scam. They are just looking for increasing their valuation. A waste of time

+100, companies certainly have direct interest in pumping asset evaluation, and emotional attachment is financial valuable thing. Emotional attachment sells better than xxx this days

Didn’t get the “scary” part. I also keep my entertainment to the minimum dependencies possible. I try to rely on stuff I own: music cds, iso videogames + emulators, physical books or ebooks (thanks Anna), exercise outdoors… ditching streaming like netflix/youtube, buying crap on amazon, uber, etc

Scary = “if I have no future prospects for work”

It’s the combination of AI changing the workplace, the large techs shedding double digit headcount, recruiting / hiring departments being so broken by the AI arms race hitting job applications, and the macro business environment generally being on the downward slope at the moment.


Scary part is not having a job right now that's all. It's not scary walking around getting more vitamin d

Why TS? The npm ecosystem is insane and insecure. Not a chance we are running this in our machines.

Go/Rust way better choices. Besides, if it’s all vibe coded, it shouldn’t matter for the author


I think the TypeScript ecosystem is more suitable for this.

I do not think Rust is a bad language. But the agent ecosystem changes very quickly, and in Rust, assembling and reshaping agent workflows is difficult.

Many people prefer Rust, and I understand why. It is a genuinely excellent language, and “Rust is a great language” is a strong message that attracts many developers. But as long as lifetimes exist, I think it will remain difficult.

The lifetime system assumes, in some sense, that humans can fully predict the lifecycle of values and resources. I am not sure that is truly possible in all domains. I am also not sure whether that model is linguistically suitable for the agent ecosystem.

In agent systems, requirements change constantly. Tools change, workflows change, providers change, schemas change, and failure policies change. In that kind of environment, I am not sure Rust is the right fit.

I like Rust a lot, and it is a language I genuinely want to learn. But I am not sure that applying Rust to everything is really the right answer.

I think Rust makes a lot of sense in relatively stable infrastructure ecosystems: operating systems, runtimes, sandboxes, and core low-level layers. But agent code usually requires high-level abstraction and rapid workflow composition. Doing that in Rust takes a tremendous amount of time.


Why do agent systems change more than other things? Maybe while were here: What even is an agent system anyway? Does one work on agent systems as the final product, or is the agent system what you work with to make something else?

The definition of “agent” has changed quite a bit, even in ACL papers and other academic work.

Looking at recent examples, the practical boundary seems to be whether an LLM uses tools. In some 2023 papers, certain pipeline-based systems were still referred to as agents. More recently, the term seems to mean something looser but more action-oriented: a system that understands a goal, uses tool calls, selects actions, and executes them.

In other words, there is still no fully settled engineering definition of what an agent is. I am not an expert or a graduate student; I mostly work as a subcontractor who gets hired by university professors to reproduce specific paper metrics.

In general, every system changes frequently in its early stage. Agent systems are no different. The workflows keep changing because the field does not yet have stable, openly accepted standards for AI development.

That is also why Claude, Codex, and others are fighting to define the standard. I think the term "harness," which Anthropic has been popularizing recently, is part of the same trend. By harness I mean the execution layer around the model call itself: context management, tool dispatch, retry and fallback policies, eval loops. That layer is still actively shifting. The naming is not settled, the responsibilities are not settled, and the boundaries between the harness and the model are not settled either. Each provider is drawing those lines a little differently right now.

So my view is this: agent systems change frequently because the definition differs from person to person, the field keeps updating rapidly, and there is no engineering standard that has been firmly established yet.

Even the I/O standard itself is not really settled.


They also mentioned Go.

Not ready for production yet but ive been working on https://wingman.actor for quite a while. Its a golang based portable agent runtime with minimal dependencies.

I am using TS sandboxed in deno for all our agent code generated from a UI builder (inspired by OpenAI's own agent builder, and spits out the same code output)

You are wrong. We should ditch walled gardens like twitter/facebook/ig

It’s not about them being smart or not. It’s about giving anthropic/openai/google the power to handle our future. Haven’t we learned anything about tech giants so far?

Any engineer (any person actually) can “learn to use AI” in a couple of days. It’s not rocket science; there’s no chance of left behind. If you haven’t use LLMs at all, a weekend would be enough to be on par with everyone else in the industry

The better you are at architecting or even directing a junior developer, the better your output too. Dont let AI make decisions, its supposed to take your decisions and turn those into code. When AI makes decisions, well the unexpected outcome is always on you.

> Dont let AI make decisions, its supposed to take your decisions and turn those into code.

I let the AI make decisions all the time. I often approve them, and I sometimes revert them. Most of the time they’re really good decisions based on my initial intent, but followed by analysis I didn’t make but agree with.


I think there's a spectrum of where to draw the line.

There's clearly some level where you want a human making decisions for even the most vibey of project, because without some kind of a spec about what you're trying to build and what features you want you'd get nonsense.

But like... maybe don't stress the details too much.


> clearly some level where you want a human making decisions

Yes, clearly. There was a meme out there, "just make something cool idk".

Statements like "Don't let AI make decisions" are made because of the loss of control we experience as mechanical parts of our work (such as writing to files) gets automated.


There is also the famous saying from IBM.

  > A computer can never be held accountable
  > Therefore a computer must never make a management decision

  > A computer can never be held accountable
  > Therefore a computer must never make a management decision
It makes sense. But most decisions, by quantity, are not management decisions.

i always found it to be easier to write code myself than to direct a junior developer.

the level of teaching involved would always mean the overall velocity of work slowed down.

some people say you can throw them the drudge work but i find that if you're doing coding right (e.g. you dont let your code base degenerate into a mess of boilerplate), there is barely any drudge work to do.


You're missing the real goal of directing a Junior, which is you're teaching them to be a team player, Junior devs will surpass your expectations, the rate at which they goof or are about to goof should decrease over time the more you mentor them. If you do it right, you not have a strong ally and coder under your belt, or would you rather someone else teach them their bad habits?

Im not missing the point that they need to be trained. I was quite explicit about that being necessary. Nor am I denying the benefits of them being trained.

I just cant pretend that I'll get extra productivity while I'm training them.

Certain professions lend themselves well to a apprentice/master framework because the apprentice doing the drudge work that requires less skill frees up time. This can increase overall productivity while the apprentice is trained. Development isnt really like that though.


> I just cant pretend that I'll get extra productivity while I'm training them.

It is an investment, but it pays back for itself is my point.


i always found it to be easier to write code myself than to direct a junior developer.

Me, too. But that doesn't mean I'm a great developer, just a shitty manager.


Perhaps but at least when you are directing a junior developer, even if badly, you'll eventually get a non-junior developer on the other side. With an AI agent, you'll get ... what?

With current models, you're right, there will be nothing to show for the effort except the code itself.

I suspect that will change sooner or later. Models will be cultivated over time the way we cultivate full-time employees now, with an acquired awareness of what they're building, new skills picked up in the process, and insight into how the larger system works.


How do you reckon?

There's a long-running instance of the model on the provider that's allocated to my organisation? Or are you thinking more of a server-side memory system, similar to the (currently very fallible) ones like Honcho and Mem0?

What happens when my org stops paying that provider? Do I get to take the now senior agent with me to the next provider? Does the provider have to delete it (and all that learning is lost forever)? Does that now become a free agent that can be hired by the next organisation like an employee (one that probably doesn't know how to keep industry secrets to itself)?


Beats me, I haven't thought it through any further than that. But it seems obvious that the combination of a vast-but-static bag of weights and a tiny, ephemeral context has limited value, and will eventually be looked back upon as a transitional form.

Karpathy's notion of a self-maintained wiki may suggest a direction for work in that area. As he points out ( https://gist.github.com/karpathy/442a6bf555914893e9891c11519... ) this is basically the latest of several attempts at implementing Vannevar Bush's "Memex" concept. I think it will find application in some form at organization-wide scope.

If the trend towards centralization holds, then yes, who gets to own and maintain the state in a stateful AI model is indeed going to be a big question. I hope it doesn't hold.


It's doesnt mean either. It's just a reflection of the high cost of training.

Others are disagreeing with you here, and I do too.

The difference is profound, and takes more than a couple of days to get your head around the implications. I'd summarise it as: "if you give a computer the same input it always produces the same output, but if you give a model the same input it always produces different output". Add to that the output is often wrong and it can't reliably follow instructions, and the difference is so great it breaks most of your intuitions.

The reward working with this piece of unreliable jelly is it can be far smarter than you (think the difference between a man with a shovel and a 20 ton excavator - they can literally find bugs in minutes that would take a human hours or days), and they know far more than you.

The engineering challenge is to make this near random machine produce a reliable product. It isn't easy.

The hype you see around them is it's trivially easy to get it to produce a feature rich but very unreliable product, as Anthropic demonstrates with their vibe coded claude-cli. I refuse to use it now. Among its other charms, it triggers a BSOD on windows: https://github.com/anthropics/claude-code/issues/30137 (Granted, it's just another Windows bug: https://learn.microsoft.com/en-ca/answers/questions/5814272/..., but if you are shipping to Windows you should be working around such bugs.)


> The reward ... is it can be far smarter than you.... and they know far more than you

I think this solidified an idea for me. A tool being smarter than me but inconsistent, is useless.

I can work with people who are smarter than me, because I can trust them, and I can trust them to own up or be held accountable for screw ups.

For a calculator, I can only hold myself accountable. However I cannot hold myself accountable for not knowing something I dont know.


> For a calculator, I can only hold myself accountable. However I cannot hold myself accountable for not knowing something I dont know.

I thought I could let this fly through keeper, but the idea keeps bugging me.

If the programmer does not vet and understand what the LLM has done, we call it vibe coding. If he takes personal responsibility for every line of code, we call it software engineering.

Lots of people use LLMs to vibe code (I do), and lots of people use them as a software engineering aid (I do that too).

The choice depends on the quality of code you need to produce. If an engineer is designing an aircraft wing, he would choose the highest grade of materials with known parameters. If he is building a kite for his son, an old newspaper would do in a pinch. Using 7075 Aluminium wouldn't just be overkill, it would mean his son doesn't get to see the fun his dad has building a flying machine in an afternoon.

A good engineer understands the properties of every tool at his disposal, and knows when to deploy each.


Just learn, sure. But the difference between my efficiency of using it on my day 2 and month 6 is significant. Yet I feel I am barely scratching the surface of it.

Firmly disagree. Learning how to use these tools effectively is unintuitively difficult.

They're great at some stuff and terrible at other stuff in ways that are very hard to predict.

I'm figuring out new and better ways to use them in a daily basis, and I've been an almost daily user for nearly three years.


They're difficult and hard to predict because they're still primitive, despite what their companies say. When (or if) they get advanced enough to deliver consistently, there will be no chance of being left behind, because even a kid will be able to use them effectively. Right now they're still at the gimmick level, although a very impressive one.

If the models get to a point of total consistency there's still a LOT that we need to figure out and learn about how to use them.

Let's say models can exactly and correctly write any code you ask of them.

- How do you break down a project into a sequence of requests to models?

- How can you most effectively parallelize the work - models will never be instant, so there will always be benefits in working out how best to use several agents at once

- Now that the models can handle the details of Lean, and Swift-UI, and Oracle stored procedures, and thousands of other technologies that you never got around to learning in the past... what can you do with those and how do you pick which projects to go after?

- How do you collaborate with other engineers and designers and product people in a world where you can churn out the right code reliably in a few minutes?

The models we have today are already effective enough to change the shape of our work as software engineers. As the models continue to improve figuring out and adapting to whatever that new shape is becomes even more complicated.


I don't mean to be dismissive, but aren't you describing simply working on projects with a finite amount of human hours/capital, but substituting finite amount of tokens? I recognize all of those problems as project problems long before LLMs, so I'm not sure it addresses GPs point.

Two key skills for any software engineer are prioritization and estimation.

You need to be able to estimate how long an approach is likely to take in order to correctly prioritize your work.

Coding agents throw a wrench in this because some of the stuff that used to take a long time doesn't take a long time any more. I'm finding my 25+ years of experience in estimating software has been completely thrown off - it's taking a whole lot of work for me to start building up those intuitions about what's quick and what takes a long time.


can you share any examples of these "new and better ways to use them"? because the only way I've used LLM and seen other people use it is to literally just talk to it, which doesn't require any skills beyond basic conversational abilities.

I'm taking about coding agents, not chatbots.

With coding agents you need to think very carefully about how you design the agentic loop such that the agent has the right tools and information available to it to compete the goal.

I've been writing a lot more about that here: https://simonwillison.net/guides/agentic-engineering-pattern...


If these tools stopped drastically improving, what justifies the crazy valuations?

Not much. The valuations are wild.

/thread

If one has been reading a wide variety of books/papers/articles/whatever their whole life, and one has been mindful of how to communicate with the "written word" as it were, it takes about 3 hours to be wildly effective with this technology. I think it took longer to learn google-fu than it did to learn how to use this technology effectively.


I agree. It really is not a difficult “skill” to learn. It took me probably 4 days to configure agents to break down requirements, write tests, write code, open PRs, review and merge PRs. Learned how to use skills, MCP, AGENTS.md. It is really not complicated. It’s just…writing well. Knowing how to decompose problems. Trying out different tools. I also learn new things every day but I sincerely do not think I would be X times more effective with these tools had I started a year ago.

> If you haven’t use LLMs at all, a weekend would be enough to be on par with everyone else in the industry

Disagree. It takes a lot of experimenting to find the right balance between sufficient guardrails and insane halluciations. And it'll be different depending on work domains.

I'm still refactoring AI workflows every week after more than a year or so and still working on it. Will probably be a perpetually ongoing effort as models change.


But does this translate as "one year of cumulative work" or rather "one year of rearranging your workflow and discarding obsolete ideas"?

If you spend a year walking in circles, someone can easily close the gap with one step. Especially if models and harnesses are supposedly getting more powerful all the time.


> a weekend would be enough to be on par with everyone else in the industry

I kind of agree in general that it is a learned skill, but considering how unclear people generally are when they communicate, I'm guessing it'll take longer than a weekend to be able to catch up, especially catch up to people who've been working on precise and careful communication and language for years already in a professional environment.


A weekend is enough to get going, but not nearly enough to 'be on par' with everyone else.

That said - what we have learned in the last year could be compressed quite a lot - there are a lot steps we could skip, and 'learn by failure' that need not be repeated.

It takes a while to get the subtleties of it, it's among the most highly nuanced things we've ever encountered.


The unspoken (and utterly antisocial) subtext is "we are aiming for an exponential leveraging of our labor and complete domination of the market."

We are moving steady without breaking things. So we don’t typically use llms for 100% of the code (we rely on it for prototyping and for code in the edge: like sdks and the like. For core functionality we have very good senior engineers)

90% of the stories on HN are AI related. It’s normal to have so many bots nowadays. If HN dies, wouldn’t be the worst thing tbh

Why do you care about github? It’s Just another corporation doing what they know best: harvesting money. The software ecosystem can live without github just fine

I haven’t seen a single ai-based product that’s relevant and making money.

In all the companies i have worked for, ai hasn’t been a productivity multiplier


Consider applying for YC's Summer 2026 batch! Applications are open till May 4

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: