Hacker Newsnew | past | comments | ask | show | jobs | submit | smlacy's commentslogin

Is it me or is that aluminum already developing some stress cracking?

Post actual results, make a blog post. Don't just say "this sucks" without tangible evidence.

Otherwise you're doomed to "sample size of one" level of relevance.


I have the opposite experience: random HN/Reddit comments saying “this sucks” or “whoa this is a huge improvement” are the only benchmark that means anything. Standard benchmarks are all gamed and don’t capture the complexity of the real world.

Then your internal benchmarks will be in the post-training set and you’ll have to make new ones.

I may already have but I'm pseudonymous on this website.

The synth engine is an implementation of https://soniccharge.com/microtonic Not sure what "accurate" means in your context though?

How closely it replicates microtonic's, would be my assumption

Yes, and with virtually zero context, which makes an enormous difference for TTFT on the MoE models.


Total gimmick. I guess we're "making progress", but this is will never lead to any useful application other than "Yes, you're absulotely right" bots. What's needed for real applications is 10000× the input token context and 10× the output token speed, so we're off by a factor of ... 100,000×?


Correct, also with the context growing, the conversations cannot continue at the initial speed either. Gimmick or not, this is very sci-fi compared to 10-20 years ago.


This should be the top comment


And with only like a dozen tokens of context. What happens when this thing gets the ~100k tokens of context needed to actually make it useful?


I like the analogy but which 2 is AI coding?

Fast & Cheap (but not Good?) - I wouldn't really say that AI coding is "cheap"

Cheap & Good (but not Fast) - Again, not really "cheap"

Fast & Good (but not Cheap) - This seems like maybe where we're at? Is this a bad place?


The proper idiom is "You can only pick two". It doesn't say that everything is two of them, or even one.


It's hitting all three, right _now_.

Eventually, it will be just Fast and Good. It won't be cheap, as companies start moving towards profitability.

Remember when Uber was super cheap? I do. They're fast and good though.


It's not cheap or good, it's just fast.


It's fast. It's cheap compared to employees. It's really the latter that people are upset about.

As for good. Well, how much software is really good? A lot of it is sewn together APIs and electron-like runtimes and 5,000 dependencies someone else wrote. Not exactly hand-crafted and artisanal.

I'm sure everyone here's projects are the exception, but engineering is always about meeting the design requirements. Either it does or it doesn't.


What's your concern?


Have you ever programmed with AI? It needs a lot of hand holding for even simple things sometimes. Forgets basic input, does all kinds of brain dead stuff it should know not to do.

>"good catch - thanks for pointing that out"


Can you clarify how, at all, that’s relevant to the article?


Both the curl and the SQLite project have been overburdened by AI bug reports. Unless the Google engineers take great care to review each potential bug for validity the same fate might apply here. There have been a lot of news regarding open source projects being stuffed to the brim with low effort and high cost merge requests or issues. You just don't see all the work that is caused unless you have to deal with the fallout...


This project has nothing to do with bug reports... it's an opt-in tool for reviewing proposed changes that kernel developers can decide to use (if they find it useful).


Well, if it doesn't find anything it's just a waste of time at best.


Prevention paradox.


i think it's a skill.


That background looks like AI for sure though?


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: