Hacker Newsnew | past | comments | ask | show | jobs | submit | realityfactchex's commentslogin

Isn't the easy way just a development VM? As in:

Install your OS of choice in a virtual machine, e.g. even hosted on your main machine.

Install the AI coding tool in the virtual machine.

Set up a shared folder between host+guest OS.

Only let the VM access files that are "safe" for it to access. Its own repo, in its own folder.

If you want to give the AI tool and VM internet access and tool access, just limit what it can reach to things it is allowed to go haywire on. All the internet and all OS tools are ok. But don't let this AI do "real things" on "real platforms" -- limit the scope of what it "works on" to development assets.

When deploying to staging or prod, copy/sync files out of the shared folder that the AI develops on, and run them. But check them first for subterfuge.

So, don't give the AI access to "prod" configs/files/services/secrets, or general personal/work data, etc. Manage those in other "folders" entirely, not accessible by the development VM at all.

Is that close?


Here are my open-source (MIT) solutions for Mac development:

SandVault [0]: Run AI agents isolated in a sandboxed macOS user account

ClodPod [1]: Run AI agents isolated inside an OSX virtual machine

0: https://github.com/webcoyote/sandvault

1: https://github.com/webcoyote/clodpod


Thanks for sharing. Which one do you use for what?

Did somewhat exactly that for apple container based sandbox - Coderunner[1]. You can use it to safely execute ai generated code via an MCP at http://coderunner.local:8222

A fun fact about apple containers[2], it's more isolated than docker containers as in it doesn't share the VM across all containers.

1. https://github.com/instavm/coderunner

2. https://github.com/apple/container


Automating this setup is also somewhat easily possible with, e. G., Lima[0] or HashiCorp vagrant[1].

[0]: https://lima-vm.io/

[1]: https://developer.hashicorp.com/vagrant


I'd just do it over a Docker mount (or equivalent) to keep it a bit more lightweight. Can keep the LLM running local; and teach it how to test/debug via instruction files.

Whether they do some kind of reasoning or not, they have all the biases included that come from their training/programming: what material was included, how that material is handled, etc.

AFAIAA, there's certainly not even a single frontier model, trained on "the internet", that is able to process information factually and in an unbiased manner.

So, they're not really reasoning impartially, as a computer "should" be wont to do. They're regurgitating biases. In a word: parroting.

Can you coax a model into seeming fair, via context? Sure. But, the baseline would be need to be based on reasoning ab initio, to qualify as reasoning. Otherwise, they are, again, parroting.

It's important to not mis-state what is or is not "emergent reasoning", or else people will think we have something that we don't, because some expert said so.

Disagree? Do you think that there is at least one accessible frontier model trained on the internet that is not parroting the baises of its creators and users, and performs its own "emergent reasoning" (instead of just doing something that mimics doing that)? Then please link to it.


It is no fun to have old iCloud photos deleted unexpectedly. Apple has provided plenty of footguns, even if they really are user errors. For examples: (1) during device restores and (2) premium subscription management fumbles.

Product idea: Apple should offer a paid service to restore the "old backups" of photos that are no longer accessible via iCloud UI/API, which were soft-removed for missing the subscription quota or whatever, if Apple happens to have that data tucked away in cold storage somewhere.

Case in point, I had some c. 2016 era photos in iMessages that I thought I handled right to not lose from iCloud, but they are apparently nowhere to be found in iCloud API based on recent checks. More than mildly irritating.

I should have used an iCloud photos backup tool like this much sooner.

Print what you want to keep onto archival paper with archival dyes. Everything else will atrophy.


That's cool and useful.

IMO, the best alternative is Chatterbox-TTS-Server [0] (slower, but quite high quality).

[0] https://github.com/devnen/Chatterbox-TTS-Server


Chatterbox-TTS has a MUCH MUCH better output quality though, the quality of the output from Sopro TTS (based on the video embedded on GitHub) is absolutely terrible and completely unusable for any serious application, while Chatterbox has incredible outputs.

I have an RTX5090, so not exactly what most consumers will have but still accessible, and it's also very fast, around 2 seconds of audio per 1 second of generation.

Here's an example I just generated (first try, 22 seconds runtime, 14 seconds of generation): https://jumpshare.com/s/Vl92l7Rm0IhiIk0jGors

Here's another one, 20 seconds of generation, 30 seconds of runtime, which clones a voice from a Youtuber (I don't use it for nefarious reasons, it's just for the demo): https://jumpshare.com/s/Y61duHpqvkmNfKr4hGFs with the original source for the voice: https://www.youtube.com/@ArbitorIan


You should try it! I wouldn’t say it’s the best, far from that. But also wouldn’t say it’s terrible. If you have a 5090, then yes, you can run much more powerful models in real time. Chatterbox is a great model though

> But also wouldn’t say it’s terrible.

But you included 3 samples on your GitHub video and they all sound extremely robotic and have very bad artifacts?


I've been using Higgs-Audio for a while now as the primary TTS system. How would you say does Chatterbox compare to it if you have experience?

I haven't used it. I compared it with T5Gemma TTS that came out recently and Chatterbox is much better in all aspects, but especially in voice cloning where T5Gemma basically did not work.

I quite like IndexTTS2 personally, it does voice cloning and also lets you modulate emotion manually through emotion vectors which I've found quite a powerful tool. It's not necessarily something everyone needs, but it's really cool technology in my opinion.

It's been particularly useful for a model orchestration project I've been working on. I have an external emotion classification model driving both the LLM's persona and the TTS output so it stays relatively consistent. The affect system also influences which memories are retrieved; it's more likely to retrieve 'memories' created in the current affect state. IndexTTS2 was pretty much the only TTS that gives the level of control I felt was necessary.


Wow, the IndexTTS2 demo is very good. Definitely going to check that out. Thanks.

[0] https://indextts2.org


Strong disagree

What's achieved is impressive as heck, but it ain't AGI.

It's easy (and good marketing) to convince people that this is AGI. And, for sure, it's fairly easy to become convinced that this is AGI.

But AGI is far, far more magical than this. (Not to discount what this is, which is amazing, too!)


You probably built enough to start finding out from others. Try to find one person who wants it (or why they don't). Then another, and another.

Some ideas, distilled from [0]:

  1. Find subreddits who would find it helpful to solve their real problems
  2. Build karma there for a week or more, be helpful without advertising
  3. Only if the sub allows self-promo, offer it for free for first 500 people (or however many)
  4. Collect their feedback, build on wahat they demand, always be shipping and listenign (have easy way to receive feedback
  5. Take the top requests on the public issue-lists/ideaboards for your compeitors; nail those
  6. On socials (YouTube, TikToik, post in 2 or 3 consistent formats, don't just copy stuff)
  7. When you find places with people you suspect would be receptive (on social media), suggest "Hi, I made this ____" in a helpful way, don't be overly self promotional, but be modest and share what you have to help
So,

  a) Keep it free long enough for users to know what product is and you to learn what they need
  b) Constantly be listening and improving.
  c) Always be launching
[0] https://www.reddit.com/r/MobileAppDevelopers/comments/1pzchh...

The parallel between microwave and AI skepticism is even broader: Plenty of people today "know" that "you shouldn't use a microwave" *at all*, either.


Overall, HN is good.

Nowhere else really can you get that concentrated & large a place to weigh in on something and get (fairly) intelligent, (fairly) rapid, (fairly) reasonable, (fairly) fast feedback -- even if you don't agree with all of it.

Or, if you're not weighing in and getting responses, just to read the temperature on any discussion (of interest) and see reasonable views raised and discussed.

PG always said something like, if you want to see how silly some people are just read the bottom-ranked comments on HN.

Well maybe HN doesn't upvote the best. But hype gonna hype. It's the best we have AFAIK.


PG also always said huuuummmm


Ok, here are some starting links and summaries:

In 2012, South Korean judges found the following to be unconstitutional [0], e.g. based on the person who complained about not being allowed to comment anonymously on YouTube and other sites (since Korean YouTube and other sites needed them to identify their real identity first):

1. Act to Promote Use of Communications Network and to Protect Information (as amended by Act No. 9119 of 13 June 2008)

2. Article 29 and Article 30, Paragraph 1 of the Enforcement Decree of the said Act (as amended by the Presidential Decree No 21278 of 28 January 2009)

Also, the Korean "Real Name" requirement was "rolled back", as reported at [1], which describes that "Article 44-5 of the Act on Promotion of Information and Communications Network Utilization and Data Protection, etc. (the 'ICN Act') was enacted in 2007... It required large-scale portal sites with more than 100,000 visitors on average a day to record the real name identities of visitors posting comments, usually via the poster's resident registration number (RRN).".

  [0] "Constitutional Court of Korea, 2010 Heon Ma 47, 252 (consolidated), Re: Confirmation of unconstitutionality", https://www.opennetkorea.org/wp-content/uploads/2023/05/Korean-real-name-law-decision-english.pdf
  [1] "Korea Rolls Back ‘Real Name’ and ID Number Surveillance" (2012), https://papers.ssrn.com/sol3/papers.cfm?abstract_id=2187232


Advent of Computing:

  https://adventofcomputing.com/

  https://podcasts.apple.com/us/podcast/advent-of-computing/id1459202600

  https://adventofcomputing.libsyn.com/rss

  https://www.youtube.com/@adventofcomputing4504/videos


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: