Hacker Newsnew | past | comments | ask | show | jobs | submit | kumarm's commentslogin

Seems true and really wish the project included some sample PDF output.

My Text to Speech app uses bounding box to display what text in PDF is being read and would not work well PDF's from this project.


OP here, I added a sample PDF output in the project assets and put screenshots in the ReadMe. The text is selectable after rehydration. would this work with your app?


Amazing. Worked really well. Thank you.


Wait! what? This is incredible. Amazing work.


[even in programming it's inconclusive as to how much better/worse it makes programmers.]

Try building something new in claude code (or codex etc) using a programming language you have not used before. Your opinion might change drastically.

Current AI tools may not beat the best programmer, they definitely improves average programmer efficiency.


> Try building something new in claude code (or codex etc) using a programming language you have not used before. Your opinion might change drastically.

Try changing something old in claude code (or codex etc) using a programming language you have used before. Your opinion might change drastically.


I have! Claude is great at taking a large open source project and adding some idiosyncratic feature I need.


this is literally how i maintain the code at my current position, if i didn't have copilot+ i would be cooked


Same here. Coworker left. I now maintain a bunch of Go code and I had never written any before.

I use copilot in agent mode.


...what were you doing before?

That's bread and butter development work.


didnt have to maintain any code before thankfully


I did just that and I ended up horribly regretting it. The project had to be coded in Rust, which I kind of understand but never worked with. Drunk on AI hype, I gave it step by step tasks and watched it produce the code. The first warning sign was that the code never compiled at the first attempt, but I ignored this, being mesmerized by the magic of the experience. Long story short, it gave me quick initial results despite my language handicap. But the project quickly turned into an overly complex, hard to navigate, brittle mess. I ended up reading the Rust in Action book and spending two weeks cleaning and simplifying the code. I had to learn how to configure the entire tool chain, understand various cargo deps and the ecosystem, setup ci/cd from scratch, .... There is no way around that.

It was Claude Code Opus 4.1 instead of Codex but IMO the differences are negligible.


AI can be quite impressive if the conditions are right for it. But it still fails at so many common things for me that I'm not sure if it's actually saving me time overall.

I just tried earlier today to get Copilot to make a simple refactor across ~30-40 files. Essentially changing one constructor parameter in all derived classes from a common base class and adding an import statement. In the end it managed ~80% of the job, but only after messing it up entirely first (waiting a few minutes), then asking again after 5 minutes of waiting if it really should do the thing and then missing a bunch of classes and randomly removing about 5 parenthesis from the files it edited.

Just one anecdote, but my experiences so far have been that the results vary dramatically and that AI is mostly useless in many of the situations I've tried to use it.


This is exactly my experience. We wanted to modernize a java codebase by removing java JNDI global variables. This is a simple though tedious task. And we tried Claude Code and Gemini. Both of these results were hilarious.


LLMs are awful at tedious tasks. Usually because it involves massive context.

You will have much more success if you can compartmentalize and use new LLM instances as often as possible.


One thing I like for this type of refactoring scenario is asking it to write a codemod (which you can of course do yourself but there's a learning curve). Faster result that takes advantage of a deterministic tool.


Yeah I've used it for personal projects and it's 50/50 for me.

Some of the stuff generated I can't believe is actually good to work with long term, and I wonder about the economics of it. It's fun to get something vaguely workable quickly though.

Things like deepwiki are useful too for open source work.

For me though the core problem I have with AI programming tools is that they're targeting a problem that doesn't really exist outside of startups, not writing enough code, instead of the real part of inefficiency in any reasonably sized org, coordination problems.

Of course if you tried to solve coordination problems, then it would probably be a lot harder to sell to management because we'd have to do some collective introspection as to where they come from.


>Of course if you tried to solve coordination problems, then it would probably be a lot harder to sell to management because we'd have to do some collective introspection as to where they come in.

Sad but true. Better to sell to management and tagline it as "you don't need a whole team anymore.", or going so far as "you can do this all by yourself now!".

Sadly managers usually have more money to spend than the workers too, so it's more profitable.


> For me though the core problem I have with AI programming tools is that they're targeting a problem that doesn't really exist outside of startups

If you work in science it's great to have s.th. that spits out mediocre code for your experiments.


> Try building something new in claude code (or codex etc) using a programming language you have not used before. Your opinion might change drastically.

So it looks best when the user isn't qualified to judge the quality of the results?


> using a programming language you have not used before

haven't we established that if you are layman in an area AI can seem magical. Try doing something in your established area and you might get frustrated. It will give you the right answer with caveats - code which is too verbose, performance intensive or sometimes ignoring best security practices.


> they definitely improves average programmer efficiency

Do we really need more efficient average programmers? Are we in a shortage of average software?


Average programmers do not produce average software; the former implements code, the latter is the full picture and is more about what to build, not how to build it. You don't get a better "what to build" by having above-average developers.

Anyway we don't need more efficient average programmers, time-to-market is rarely down to coding speed / efficiency and more down to "what to build". I don't think AI will make "average" software development work faster or better, case in point being decades of improvements in languages, frameworks and tools that all intend to speed up this process.


> Are we in a shortage of average software?

Yes. The "true" average software quality is far, far lower than the average person perceives it to be. ChatGPT and other LLM tools have contributed massively to lowering average software quality.


I don’t understand how your three sentences mesh with each other. In any case, making the development of average software more efficient doesn’t by itself change anything about its quality. You just get more of it faster. I do agree that average software quality isn’t great, though I wouldn’t attribute it to LLMs (yet).


> using a programming language you have not used before

But why would I do that? Either I'm learning a new language in which case I want to be as hands-on as possible and the goal is to learn, not to produce. Or I want to produce something new in which case, obviously, I'd use a toolset I'm experienced in.


There are plenty of scenarios where you want to work with a new language but you don't want to have to dedicate months/years of your life to becoming expert in it because you are only going to use it for a one-time project.

For example, perhaps I want to use a particular library which is only available in language X. Or maybe I'm writing an add-on for a piece of software that I use frequently. I don't necessarily want to become an expert in Elisp just to make a few tweaks to my Emacs setup, or in Javascript etc. to write a Firefox add-on. Or maybe I need to put up a quick website as a one-off but I know nothing about web technologies.

In none of these cases can I "use a toolset I'm experienced in" because that isn't available as an option, nor is it a worthwhile investment of time to become an expert in the toolset if I can avoid that.


How can I possibly assess the results in a programming language I haven’t used before? That’s almost the same as vibe coding.


The same way you assess results in a programming language you have used before. In a more complicated project that might mean test suites. For a simple project (e.g. a Bash script) you might just run it and see if it does what you expect.


The way I assess results in a familiar programming language is by reviewing and reasoning through the code. Testing is necessary, but not sufficient by any means.


Out of curiosity, how do you assess software that you didn't write and just use, and that is closed source? Don't you just... use it? And see if it works?

Why this is inherently different?


You are correct that this is indeed a mostly unsolved problem. In chapter 15 of "The Mythical Man-Month", Fred Brooks called for all program documentation to not only for how to use a program, but also for how to modify a program [1] and, relevant to this discussion, for how to believe a program. This was before automated tests and CI/CD were a thing, so he advocated for shipping testcases with the program that the user could review and execute at any time. It's now 50 years later, and this is one of the many lessons in that book that we've collectively not picked up on enough.

[1] Side-note: This was written at a time when selling software as a standalone product was not really a thing, so everything was open-source and the "how to modify" part was more about how to read and understand the code, e.g. architecture diagrams.


As you said, this is in a very different context. He was building an OS, which was sold to highly technical users running their own programs on it.

I'm talking about "shrinkwrap" software like Word or something. There's nothing even close to testing for that this is not just "system testing" it.


The question is: is that value worth the US$400bi per year of investment sucking out all the money from other ventures?

It's a damn good tool, I use it, I've learned the pitfalls, it has value but the inflation of potential value is, by definition, a bubble...


It is not worth it and it is not even that impressive considering the cost.

If you told me that you would spend half a trillion and the best minds on reading the whole internet, then with some statistical innovation try to guess the probable output of an input. The way it works now seems about right, probably a bit disappointing even.

I would also say, it seems cool and you could do that, but why would you? At least when the training is done it is cheap to use right? No!? What the actual fuck!


The UI presentation of Schema is really nice. Is there a javascript based UI for web that is used?


It is same on PlayStore and AppStore.

You would be surprised to know Apple started this in AppStore before Google on PlayStore. I assume it is because Google wanted to be safe from Antitrust lawsuits (Follow Apple rather than going there first).


Since API currently is not working (seems rate limits not set for Image Generation yet) I tried on fal.

Definitely inferior to results I see on AI Studio and image generation time is 6s on AI Studio vs 30 seconds on Fal.AI


> Definitely inferior to results

Quality or latency?


Latency. Started using API. Gets response directly on Gemini API in under 6 seconds. Fal.ai takes about 30 seconds.


Seems to be failing at API Calls right now with "You exceeded your current quota, please check your plan and billing details. For more information on this error,"

Hope they get API issues resolved soon.


I am in Bay Area. I donated my cars to Kars4Kids before and even recommended to others since their process to donate is simple.


I'm sorry to hear that.


Most likely. But that's exactly what someone who hasn't experienced enough cycles in industry would come up with :).


Exactly this. My startup count is zero, but even I’m well enough informed to know that casting such a wide net is diverting attention away into too many tangents, too quickly, and will drain your balance sheet dry before you can capitalize on any singular success to build a moat or secure another funding round.

They were banking on AI coding being better than it was, and the snowball effect happening faster than it ever could. And now, they’re toast.


He is second only to Elon in this case (SolarCity, X/XAI).


That's why they so despise each-other; they're the same



No - Elon is much more successful


All my Veo 3 videos has sound missing. No idea why. Seems like a common problem.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: