Hacker Newsnew | past | comments | ask | show | jobs | submit | ACCount37's commentslogin

Modern phones pack a good bit of compute, and can run things like VLAs decently well.

Of course, that would require today's phones to age out of "being used as a phone" bracket, and robotics VLAs to become actually useful. But things like the Comma AI autopilot hardware use slightly obsolete smartphone chips internally - so it's not like it's impossible to run a useful AI on this kind of HW.


Apple only uses Qualcomm chips as modems. Almost everyone else uses Qualcomm chips as main SoCs.

Now, could hardware vendors tell Qualcomm to go pound sand and run their own support for old SoCs? Yes they could. Do they want to? Hell no, supporting old devices doesn't make any money.


One of the very few genuinely bad takes Linus had.

Bootloader unlocking should be a basic consumer right, and if Linux went GPLv3, it would be closer to reality.


I think expecting software licenses to enforce your rights outside of the realm of software is a pretty bad take. I think Linus's take is quite solid: "I give you my source code, you give me your changes back, and we're even". There are a lot of us who don't think that FOSS should be weaponized as a poison pill to enact the authors worldview on topics outside of the realm of software alone.

If it should be a consumer right, why limit it only to devices certain types of software? Why not consumer protection law that applies to all devices? I think software licenses are the wrong tool for this problem.

There's a lot of crazy crayon licenses out there that try to fix the whole world by tacking on a whole lot of restrictions to their software licenses, prohibiting use for a long list of reasons... to me it sounds like a bunch of newspeak, as if "more restrictions = more freedom"


Is being able to replace software on the devices I own not in "the realm of software" somehow?

"Sure, you can have the sources, you just can't use them on your own devices because the vendor that shipped it has decided to bar you from doing that with a 2048-bit RSA key" just feels like GPL was upheld in letter, but not in spirit.


Yeah, it is -- you're asking for restrictions on pieces of hardware unrelated to the original software other than the fact that someone decided to install it on there.

How would you feel if a piece of hardware came with a license prohibiting software developers from using encryption to secure their systems?

The root of the issue here is that phone hardware landscape is effectively a duopoly. It is an antitrust issue. Trying to use software licenses to do this 1) won't be effective because the duopoly will never use them, and 2) is like going around your ass to get to your elbow. Even if it did work it wouldn't get to the root of the issue. The law needs to fix the fact that almost all phones on the planet are controlled either directly or indirectly by two companies.


It’s not Linus Torvalds’ duty to make bootloader unlocking a reality.

It isn't. But he could have contributed massively to it, and hasn't, and I can and will hold it against him.

There will always be more ways for companies to extract value without contributing. Linus would have to continuously upgrade licenses from GPLV2 to V3 to Affero and so on. It is not really practical.

What Linus has contributed is already huge. We can't put all the burden of making the world right on him.


"Good enough" open weights models were "almost there" since 2022.

I distrust the notion. The bar of "good enough" seems to be bolted to "like today's frontier models", and frontier model performance only ever goes up.


The generation of frontier models from H1 2025 is the good enough benchmark.

Flash forward one year and it'll be H1 2026.

I don’t see why. Today frontier models are already 2 generations ahead of good enough. For many users they did not offer substantial improvement, sometimes things got even worse. What is going to happen within 1 year that will make users desire something beyond already working solution? LLMs are reaching maturity faster than smartphones, which now are good enough to stay on the same model for at least 5-6 years.

Any considerable bump in model capability craters my willingness to tolerate the ineptitude of less capable models. And I'm far from being alone in this.

Ever wondered why those stupid "they secretly nerfed the model!" myths persist? Why users report that "model got dumber", even if benchmarks stay consistent, even if you're on the inference side yourself and know with certainty that they are actually being served the same inference over the same exact weights on the same hardware quantized the same way?

Because user demands rise over time, always.

Users get a new flashy model, and it impresses them. It can do things the old model couldn't. Then they push it, and learn its limitations and quirks as they use it. And then it feels like it "got dumber" - because they got more aggressive about using it, got better at spotting all the ways it was always dumb in.

It's a treadmill, and you pretty much have to keep improving the models just to stay ahead of user expectations.


> users report that "model got dumber"

I have seen this with ChatGPT progression from 4o to 5.2 applied to the newest model. Old prompts stop working reliably, different hallucination modes etc.


Why go for long synthetic chains?

Methane has good energy density, doesn't demand cryogenics or diffuse through steel, burns very cleanly, and can be used in modified gasoline ICEs - without even sacrificing the gasoline fuel capability.


Isn't the point that it is as simple and convenient as normal gasoline and also that you can use your gasoline car? If you are using gases it is a hassle for everyone and you need a new car or a full retrofit. At some point we have to ask ourself why we would even do that. Is it really worth it compared to just using a battery?

Without cryogenics, methane has such low energy density that a low-pressure fuel tank would still have to be as big as a bus for your compact methane-powered vehicle to go as far as you could on a few gallons of gasoline.

Why?

CNG ICE vehicles exist, especially in parts of the world that have cheap natural gas and expensive gasoline - often as dual fuel retrofits.

They have to deal with high pressure tanks, but compared to the woes of hydrogen storage, that's downright benign.


Good question.

It's just the physical properties of methane.

That's why they use high-pressure tanks because with low-pressure gas storage, the tank needs to be bigger than the car.

Energy density of methane is still lower than any other hydrocarbons.


Again: current day CNG ICEs don't actually have "fuel tanks bigger than the car".

That's because they use high-pressure tanks.

Actually higher working presure than regular welders' oxygen tanks.

Too bad the higher the pressure, the heavier the tank, especially steel.

I had a pretty good job offer from a company that was going to start building CNG tanks out of carbon fiber.

This was a few years before the failure of the Titan submersible.

CNG has made some progress since then, here's the kind of thing there is now:

https://steelheadcomposites.com/sites/default/files/2024-01/...


COPVs are not exactly a new tech. They're just taking a while to proliferate to those lower end applications.

Not like it's a necessary thing. CNG conversions were already viable even with 00s steel tanks.


Agreed on "green steel".

In general, "green hydrogen" makes the most sense if used as a chemical feedstock that replace natural gas in industrial processes - not to replace fossil fuels or be burned for heat.

On paper, hydrogen has good energy density, but taking advantage of that in truth is notoriously hard. And for things that demand energy dense fuels, there are many less finicky alternatives.


You do need to research "obvious" things every once in a while. They have this annoying tendency of being proven wrong occasionally.

What you're saying is "different LLMs recommend different things".

Replace "LLMs" with "random schmucks online" and what changes exactly?


No one is arguing to replace everything with random schmucks.

Why would one argue for the status quo?

Modern ML is old school mad science.

The lifeblood of the field is proof-of-concept pre-prints built on top of other proof-of-concept pre-prints.


Sounds like you agree this “evidence” lacks any semblance of scientific rigor?

(Not GP) There was a well recognized reproducibility problem in the ML field before LLM-mania, and that's considering published papers with proper peer-reviews. The current state of afairs in some ways is even less rigourous than that, and then some people in the field feel free to overextend their conclusions into other fields like neurosciences.

Frankly, I don't see a reason to give a shit.

We're in the "mad science" regime because the current speed of progress means adding rigor would sacrifice velocity. Preprints are the lifeblood of the field because preprints can be put out there earlier and start contributing earlier.

Anthropic, much as you hate them, has some of the best mechanistic interpretability researchers and AI wranglers across the entire industry. When they find things, they find things. Your "not scientifically rigorous" is just a flimsy excuse to dismiss the findings that make you deeply uncomfortable.


Strange that they raised money at all with an idea like this.

It's a bad idea that can't work well. Not while the field is advancing the way it is.

Manufacturing silicon is a long pipeline - and in the world of AI, one year of capability gap isn't something you can afford. You build a SOTA model into your chips, and by the time you get those chips, it's outperformed at its tasks by open weights models half their size.

Now, if AI advances somehow ground to a screeching halt, with model upgrades coming out every 4 years, not every 4 months? Maybe it'll be viable. As is, it's a waste of silicon.


Poverty of imagination here, plenty uses of this and its a prototype at this stage.

What uses, exactly?

The prototype is: silicon with a Llama 3.1 8B etched into it. Today's 4B models already outperform it.

Token rate in five digits is a major technical flex, but, does anyone really need to run a very dumb model at this speed?

The only things that come to mind that could reap a benefit are: asymmetric exotics like VLA action policies and voice stages for V2V models. Both of which are "small fast low latency model backed by a large smart model", and both depend on model to model comms, which this doesn't demonstrate.

In a way, it's an I/O accelerator rather than an inference engine. At best.


With LLMs this fast, you could imagine using them as any old function in programs.

You could always have. Assuming you have an API or a local model.

Which was always the killer assumption, and this changes little.


Even if this first generation is not useful, the learning and architecture decisions in this generation will be. You really can't think of any value to having a chip which can run LLMs at high speed and locally for 1/10 of the energy budget and (presumably) significantly lower cost than a GPU?

If you look at any development in computing, ASICs are the next step. It seems almost inevitable. Yes, it will always trail behind state of the art. But value will come quickly in a few generations.


maybe they're betting on improvement in models to plateau, and that having a fairly stablized capable model that is orders of magnitude faster than running on GPU's can be valuable in the future?

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: