Hacker Newsnew | past | comments | ask | show | jobs | submit | h14h's commentslogin

A lack of tool & die isn't the West "forgetting how to make things" so much as it is the West choosing not to make the things that make things.

To the extent that AI is analogous to automation in manufacturing, and "writing code" to working on an assembly line, it's hard to argue the West is any thing other than a global leader in "software tool & die", so to speak.


Been meaning to check out Railway for a while, but now feeling happy about dragging my feet.

As flashy as their DX seems to be, the fact that a sketchy single VPS node with a server, a SQLite instance, and a LiteStream hookup has a better recovery story really makes me not trust their platform.


What sells me on it is I get to take a spare Gen4 m.2 ssd out of my gaming PC I wasn't fully utilizing instead of paying for 1TB of storage.

Being able to drive the price down by re-using parts I already have is a pretty big selling point, IMO.

Also, I think Apple is benefitting from scale, since they're able to maintain the (usually too high) storage and memory prices they've had for years. At this moment in time, framework have the misfortune of being forced to pass inflated wholesale prices onto the consumer.

Make this comparison one calendar year ago and the F13 Pro could very easily beat the MBP on price spec-for-spec.


This seems huge for subscription customers. Looking at the Artificial Analysis numbers, 5.5 at medium effort yields roughly the intelligence as 5.4 (xhigh) while using less than a fifth the tokens.

As long as tokens count roughly equally towards subscription plan usage between 5.5 & 5.4, you can look at this as effectively a 5x increase in usage limits.


As someone who always leaves intelligence at default, and am ok with existing models, should I be shifting gears more manually as providers sell us newer models? Is medium or lower better than free/cheaper models?

SOTA models on medium are probably still better than free or cheap models, but you should experiment.

For some real data, Artificial Analysis reported that 4.6 (max) and 4.7 (max) used 160M tokens and 100M tokens to complete their benchmark suite, respectively:

https://artificialanalysis.ai/?intelligence-efficiency=intel...

Looking at their cost breakdown, while input cost rose by $800, output cost dropped by $1400. Granted whether output offsets input will be very use-case dependent, and I imagine the delta is a lot closer at lower effort levels.


This is the right way of thinking end-to-end.

Tokenizer changes are one piece to understand for sure, but as you say, you need to evaluate $/task not $/token or #tokens/task alone.


If your use-cases don't benefit from RSC performance characteristics then they probably aren't outright better.

But I do think they're a compelling primitive from a DX standpoint, since they offer more granularity in specifying the server/client boundary. The TanStack Composite/slots API is the real selling point, IMO, and as far as I can tell this API is largely (entirely?) thanks to RSCs.


TanStack uses streams as the basis for loading RSC data, and recommends using a route loader to access them:

https://tanstack.com/start/latest/docs/framework/react/guide...

AFAIK, at least when using TanStack Router, this RSC implementation seems just as capable as the others when it comes to reducing server round trips.


"Critical" even feels strong. The article was essentially a collection of statements others have made about Sam.


Right, but the picture those statements painted collectively was not flattering. And that was certainly intended by the authors. Thus, critical, but not at all "incendiary."

Update: To clarify, my personal stance is that the critical tone was both intended by the authors and, in my opinion, appropriate given how much power Mr. Altman holds. If he has a history of behaving inconsistently, that deserves daylight.


Are you arguing that because the authors knew the pattern they were documenting was unflattering, the piece is somehow compromised? That they clearly had an agenda? That's called reporting. They called a hundred-plus named sources and the picture those sources independently painted was damning. Altman has a history of telling repeated, easily-checked lies, followed by fresh lies when caught in the first ones.

Are you suggesting that they should have "both sides"-ed by reporting company PR and Sam-friendly sources and giving them equal weight? Sometimes the facts point in one direction.


> Are you arguing that because the authors knew the pattern they were documenting was unflattering, the piece is somehow compromised?

Uh, no? Lol, I'm on your side, bud. Put away the pitchfork. I thought it was a really good and fair article. I am not the adversary you're looking for.


> my personal stance is that the critical tone was both intended by the authors

You may think we are on the same side. You don't understand what side I'm on. "Lol".

Your "personal stance" is that you can get inside the heads of the reporters? Obviously not. So you're going by the idea that an article that leads to critical conclusions is inherently slanted. This is an insidious and damaging idea. It has led to the belief by journalists and editors that they need to twist themselves into pretzels to present "both sides", which is easily exploited by people of bad faith to launder outright lies. There's a direct line between this and authoritarianism. I'm quite serious about this. The fact that you agree with the authors in this case is completely orthogonal.

Jay Rosen has written a lot about this, well worth reading: https://pressthink.org/2010/11/the-view-from-nowhere-questio...


Every article is inherently biased due to the fact that there are inclusions and omissions. This is just a fact.

You're injecting your own personal view into GP's statement by adding a lot of weight into the distinction between the words "critical" and "incendiary" and "neutral", when GP made a very neutral and not as charged statement.


Look if you're looking for a fight just visit a local martial arts gym.


Bud. Put the keyboard down and relax. I have no idea what you're talking about. You've extrapolated all this just from what I wrote?


> You've extrapolated all this just from what I wrote?

says the guy who said "certainly intended by the authors" based on... what they wrote?

On top of that "Put the keyboard down and relax" from the guy who keeps replying?

<chef's kiss>

> I have no idea what you're talking about.

The one point I'll concede!


I love reading stuff like “Critical, slanted, and compromised mean the same thing. They are interchangeable words.”

Given that, it looks like your position on davesque’s posts is slanted. Your take is critical of those posts, which means your assessment is compromised, and as such should not be taken as valid.


And I love seeing sentiments attributed to me, in quotes even, that I didn't state or imply, and certainly don't believe. "Critical" by itself is not a synonym for "slanted". However the post I was commenting on was:

> Right, but the picture those statements painted collectively was not flattering. And that was certainly intended by the authors. Thus, critical, but not at all "incendiary."

The key there is "certainly intended by the authors". The full sentiment here IS equivalent to "slanted".


It is clearly your intent to be critical of davesque’s posts. QED your analysis is compromised


"Despite unprecedented capital investment in our R&D, our core product isn't getting meaningfully better so now we're building an app."

Doesn't really strike me as the kind of statement that comes out of a company that can sustain a ~$1T market cap...


My (potentially naive) take is that open models will save us. The biggest markets for LLMs (e.g. coding) are narrow-enough to be served well by smaller models with proper RL. Cursor's Composer 2 (created from a Kimi K2.5 base) is a great example, and I expect it to be the first of many.

The wealth of great open models provide an excellent base for fine-tuning, distillation, and RL. I see a lot of untapped potential in the field of bespoke, purpose-built models that can be served far more cheaply than the frontier competition. I would not be surprised if we see frontier-adjacent experiences running comfortably on a Mac Mini by year end.

With frontier models seemingly hitting diminishing returns in quality, I struggle to see a world in which gigantic, expensive, general-purpose models don't become increasingly niche.


Consider applying for YC's Summer 2026 batch! Applications are open till May 4

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: