Hacker News .hnnew | past | comments | ask | show | jobs | submit | eldenring's commentslogin

I'm not sure what you mean by menial coding but all my employers have supported this in the past. This was a variety of companies, big tech, startups, etc. I think its more likely your employer is the outlier.

Interestingly, the person at Microsoft states in a reply that even most of them have to pick this up in their spare time. Judging from other replies, it seems that there are quite some differences in how companies approach this.

What I meant with "menial coding" is those jobs where people have to submit TPS reports on how much hours they spend for each customer. Reading a paper such as this one [1] is typically not directly necessary for being a good frontend developer, but it might stimulate someone to develop into a more fruitful employee in the long run. Managers would have to explain to their customer why time is being spent on that, and that requires some vision and creativity, which is not always a given.

[1] https://arxiv.org/abs/1706.03762


I’ve been scolded for reading books and documentation for the tasks and software I was asked to build (at a startup) during my regular work hours

No company I’ve worked at has ever had dedicated time for reading papers or articles

Maybe I’ve only worked at outliers?


All the companies I've worked at implicitly assume that you're supposed to use your working hours for more than just coding, including learning what you need for the task at hand, although if you're looking at very beginner material that might raise some suspicion.

In the case I mentioned above, the company wanted me to build a search engine before elastic search existed, and before there was full-text search in popular dbs like Postgres or MySQL. The CTO/founder gave me his credit card and told me to buy whatever books I needed. I bought about 5 different relevant books. Work days were about 10-12hrs, they still wanted me to read/research on my own time

In 35 years in the industry, reading and studying during work hours were always supported. Frankly, most places would let us play video games during work hours as long as we met our deadlines.

I've had mandated gaming on Friday after lunch. But this was in the gaming industry so it's "market research"!

We also often played board games. My favourite was playing secret Hitler with my team that one time. That was fun! (I managed to become "untouchable" while also being Hitler. That's a memorable moment!)


Thats my experience as well. Of course not ten paper a day but some learning is always encouraged.

One company had a +1 day. You worked 4 days, had 1 day for learning - everything relevant for the job was fine.


Its not subsidized, in fact, they probably have very healthy margins on Claude Code.


Yeah. If you ignore the negligible fact that some investor may want a return on all that money that is going into capex I am pretty sure you can, Enron style, get to the conclusion that any of those companies have “healthy” margins.


Why do you think that?


DeepSeek had a theoretical profit margin of 545 % [1] with much inferior GPUs at 1/60th the API price.

Anthropic's Opus 4.6 is a bit bigger, but they'd have to be insanely incompetent to not make a profit on inference.

[1] https://github.com/deepseek-ai/open-infra-index/blob/main/20...


American labs trained in a different way than the Chinese labs. They might be making profit on inference but they are burning money otherwise.


> they'd have to be insanely incompetent to not make a profit on inference.

Are you aware of how many years Amazon didn’t turn a profit?

Not agreeing with the tactic - just…are you aware of it?


Amazon was founded in 1994, went public in 1997 and became profitable in 2001. So Anthropic is two years behind with the IPO but who knows, maybe they'll be profitable by 2028? OpenAI is even more behind schedule.


How much loss did they accumulate until 2001? Pretty sure it wasn't the 44 billion OpenAI has. And Amazon didn't have many direct competitors offering the same services.


Did Amazon really not turn a profit, or apply a bunch of tricks to make it appear like they didn't in order to avoid taxes? Given their history, I'd assume the later: https://en.wikipedia.org/wiki/Amazon_tax_avoidance

Anyway, this has nothing to do with whether inference is profitable.


It has everything to do with whether they make profit on paper,

vs give away the farm via free tier accounts, free trials,

and last but not least: subsidized compute to hook entire organizations………


Deepseek lies about costs systematically. This is just another fantasy.


What do you base your accusations on? Is there a specific number from the link above that you claim is a lie?

And how are 7 other providers able to offer DeepSeek API access at roughly the same price as DeepSeek?

https://openrouter.ai/deepseek/deepseek-v3.2


Their price is not a signal of their costs, it is the result of competitive pressure. This shouldn't be so hard to understand. Companies have burned investor money for market share for quite some time in our world.

This is the expected, the normal, why are you so defensive?


> why are you so defensive?

Because you made stuff up, did not show any proof, and ignored my proof to the contrary.

You made the claim:

    > Deepseek lies about costs systematically.
DeepSeek broke down their cost in great detail, yet you simply called it "lies", but did not even mention which specific number of theirs you claim is a lie, so your statement is difficult to falsify. You also ignored my request for clarification.


You’re citing deepseek unaudited numbers. This is not even close to a proof. Unless proven otherwise it is propaganda. Meanwhile we have several industry experts pointing not only towards DeepSeek ridiculous claims of efficiency, but also the lies from other labs.


Again, your claims are impossible to verify or falsify, because they are too unspecific.

> Meanwhile we have several industry experts pointing not only towards DeepSeek ridiculous claims of efficiency, but also the lies from other labs.

What are those "industry experts" saying that is made up and what is their basis for that?

> You’re citing deepseek unaudited numbers.

Which specific number are you claiming to be fake?

I could just guess blindly and find alternative sources for random numbers from DeepSeek's article.

For example, the tokens-per-second efficiency can also be calculated based on the 30k tps from this NVIDIA article: https://developer.nvidia.com/blog/nvidia-blackwell-delivers-...

But looking for other sources is a waste of my time, when you could just be more precise.


Because if you don't then current valuations are a bublle propped inflated by burning a mountain of cash.


That's not how valuations work. A company's valuation is typically based on an NPV (net present value) calculation, which is a power series of its time-discounted future cash flows. Depending on the company's strategy, it's often rational for it to not be profitable for quite a long while, as long as it can give investors the expectation of significant profitability down the line.

Having said that, I do think that there is an investment bubble in AI, but am just arguing that you're not looking at the right signal.


And that's OpenAI's biz model? :)


The issue is that they have already paid off their datacenter 5x over compared to cloud. For offline, batch training, I don't ses how any amount of risk could offset the savings.


It’s no issue and it’s right in the front for their situation, if your business is computer it makes little sense for cloud.

That said from the risk perspective I assume for what their doing in the data center there is low risk if downtime happens.


Model training, perhaps. Web hosting? Not so much.


The ideal shape would be a shaded, flat panel perpendicular to the sun right?


That should be better than a sphere. Though I imagine there could be some fancier 3D geometry designs.

Even for a simple sphere, if we give it different surface roughnesses on the sun-facing side and the "night" side, it can have dramatically different emissivity.


This is a common way of thinking. In practice this type of thing is more like optimizing flop allocation. Surely with an infinite compute and parameter budget you could have a better model with more intensive operations.

Another thing to consider is that transformers are very general computers. You can encode many many more complex architectures in simpler, multi layer transformers.


What you're describing is one of two mechanisms of shedding heat which is convection, heating up the environment. What the long comment above is describing is a _completely_ different mechanism, radiation, which is __more__ efficient in a vacuum. They are different things that you are mixing up.


Solar in space is a very different energy source in terms of required infrastructure. You don't need batteries, the efficiency is much higher, cooling scales with surface area (radiative cooling doesn't work as well through an atmosphere vs. vacuum), no weather/day cycles. Its a very elegant idea if someone can get it working.


Only if you also disregard all the negatives.

The panels suffer radiation damage they don't suffer on Earth. If this is e.g. the same altitude orbits as Starlink, then the satellites they're attached to burn up after around tenth of their ground-rated lifetimes. If they're a little higher, then they're in the Van Allen belts and have a much higher radiation dose. If they're a lot higher, the energy cost to launch is way more.

If you could build any of this on the moon, that would be great; right now, I've heard of no detailed plans to do more with moon rock than use it as aggregate for something else, which means everyone is about as far from making either a PV or compute factory out of moon rock as the residents of North Sentinel Island are.

OK, perhaps that's a little unfair, we do actually know what the moon is made of and they don't, but it's a really big research project just to figure out how to make anything there right now, let alone making a factory that could make them cost-competitive with launching from Earth despite the huge cost of launching from Earth.


> The panels suffer radiation damage they don't suffer on Earth.

I don't think this is true, Starlink satellites have an orbital lifetime of 5-7 years, and GPUs themselves are much more sensitive than solar panels for rad damage. I'd guess the limiting factor is GPU lifetime, so as long as your energy savings outpace the slightly faster gpu depreciation (maybe from 5 -> 3 years) plus cost of launch, it would be economical.

I've said this elsewhere, but based on my envelope math, the cost of launch is the main bottleneck and I think considerably more difficult to solve than any of the other negatives. Even shielding from radiation is a weight issue. Unfortunately all the comments here on HN are focused on the wrong, irrelevant issues like talking about convection in space.


> I don't think this is true, Starlink satellites have an orbital lifetime of 5-7 years,

That's better than I thought, but still means their PV is only lasting order-of 20% of their ground lifespans, so the integrated lifetime energy output per unit mass of PV isn't meaningfully improved by locating them in space, even if they were launched by an efficient electromagnetic system rather than by a rocket.


Yeah the main benefits are:

1. solar is very efficient at generating energy, no moving parts, simple physics etc.

2. in space you don't deal with weather or daylight cycle, you can just point your panels at the sun and generate very stable energy, no batteries required

3. environmental factors are simpler, no earthquakes, security, weather. Main problem here is radiation

In theory its a very elegant way to convert energy to compute.


2 is wrong. At a Lagrange point you can do this. Not in low earth orbit - in LEO sunset is every 60 minutes or so, and you spend the next 60 minutes in darkness.

Satellites are heavily reliant on either batteries or being robust to reboots, because they actually do not get stable power - it's much more dynamic (just more predictable too since no weather).


Interesting.

According to this other source https://www.satellitetoday.com/connectivity/2026/02/02/space...

the filing mentions this

> these satellites would operate between 500 km and 2,000 km altitude and 30 degrees and Sun-Synchronous Orbit inclinations (SSO)


You can line the solar panels and radiators facing away from each other, and the radiators would take up less surface area. I think maybe the tricky part would be the weight of water + pipes to move heat from the compute to the radiators.


Water is not needed to move heat. Heat pipes do it just fine. There's one in your laptop and one in your phone too. It does scale up.


Interesting, That could surely simplify things.


Google tested the radiation tolerance of tpus which include hbm and they performed fine. https://research.google/blog/exploring-a-space-based-scalabl...


That is not a realistic test, as any space engineer could’ve told them. First of all that’s on the very low end for a cosmic ray, an order of magnitude below the average energy. But the average energy doesn’t matter because there is a very wide distribution and the much more intensive cosmic rays do far more damage. It was also not a fully integrated test with a spacecraft, which matters because a high energy cosmic ray, striking other parts of the spacecraft, generates a shower of secondary particles that do most of the damage of a cosmic ray strike.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: