Hacker Newsnew | past | comments | ask | show | jobs | submit | svantana's commentslogin

> 80mph to cars that push past 600mph

I have yet to see evidence that this is really the case. Already 15 years ago, people were creating impressive software over the course of a hackday, by glueing open source repos together in a high level language. Now that process has been sped up even more, but does it matter that much if the prototype takes 4 or 24 hours to make? The real value is in well-thought-out, highly polished apps, and AFAICT those still take person-years to complete.


The REAL speed up comes to efforts that are already well-designed, but require lots of human busy work. I've personally seen multi-day human efforts reduced to a 15-minute session with an LLM. In a way, LLMs are reducing implementation costs to the kolmolgorov complexity -- you can get what you prompt for, but you have to remember to prompt for everything you want to get -- which comes easiest if you already took time to consider the design.

Now. For example in Algoriddim DJay

> At some level, the simplest thing to do is to give up and crash if things are no longer sane.

The problem with this attitude (that many of my co-workers espouse) is that it can have serious consequences for both the user and your business.

- The user may have unsaved data - Your software may gain a reputation of being crash-prone

If a valid alternative is to halt normal operations and present an alert box to the user saying "internal error 573 occurred. please restart the app", then that is much preferred IMO.


> If a valid alternative is to halt normal operations and present an alert box to the user saying "internal error 573 occurred. please restart the app", then that is much preferred IMO.

You can do this in your panic or terminate handler. It's functionally the same error handling strategy, just with a different veneer painted over the top.


Crashing is bad, but silently continuing in a corrupt state is much worse. Better to lose the last few hours of the user's work than corrupt their save permanently, for example.

> Your software may gain a reputation of being crash-prone

Hopefully crashing on unexpected state rather than silently running on invalid state leads to more bugs being found and fixed during development and testing and less crash-prone software.


So you don't get a crash log? No, thanks.

- The user may have unsaved data

That should not need to be a consideration. Crashing should restore the state from just before the crash. This isn't the '90s, users shouldn't have to press "save" constantly to avoid losing data.


That's probably because "Gemini 3.5 Pro" doesn't exist


The silly verbiage can be excused but not the graphs with completely unlabeled data points, IMO.


Yep that's what I mean - looks like AI slop to me.


That's mentioned in the article, but is the lock-in really that big? In some cases, it's as easy as changing the backend of your high-level ML library.


That is like how every ORM promises you can just swap out the storage layer.

In practice it doesnt quite work out that way.


That's what it is on paper. But in practice you trade one set of hardware idiosyncrasies for another and unless you have the right people to deal with that, it's a hassle.


On top, when you get locked into Google Cloud, you’re effectively at the mercy of their engineers to optimize and troubleshoot. Do you think Google will help their potential competitors before they help themselves? Highly unlikely considering their actions in the past decade plus.


Given my Fitbit's inability to play nice with my pixel phone, I have zero faith in Google engineers.

What else would one expect when their core value is hiring generalists over specialists* and their lousy retention record?

*Pay no attention to the specialists they acquihire and pay top dollar... And even they don't stick around.


I thin k you can only run on google cloud not aws bare metal azure etc


According to that site, there were more tech layoffs in 2022 than in 2024 or 2025. Doesn't that speak against the "AI is taking tech jobs" hypothesis?


Massive, embarrassingly shortsighted overhiring in 2020 and 2021 seems like the more likely culprit.


I agree, I think AI taking jobs is all smoke and mirrors by companies trying to gas up their stock prices


Doesn't work in this case because the 'talk' (github PR comments) is also computer generated. But in person (i.e. at work) it's a good strategy


SWEBench-Verified is probably benchmaxxed at this stage. Claude isn't even the top performer, that honor goes to Doubao [1].

Also, the confidence interval for a such a small dataset is about 3 percent points, so these differences could just be up to chance.

[1] https://www.swebench.com/


claude 4.5 gets 82% on their own highly customized scaffolding. (parallel compute with a scoring function). That beats Doubao


Grok got to hold the top spot of LMArena-text for all of ~24 hours, good for them [1]. With stylecontrol enabled, that is. Without stylecontrol, gemini held the fort.

[1] https://lmarena.ai/leaderboard/text


Is it just me or is that link broken because of the cloudflare outage?

Edit: nvm it looks to be up for me again


Grok is heavily censored though


Is it censored... or just biased towards edge-lord MechaHitler nonsense whenever Musk feels like tinkering with the system prompt?


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: