Nowhere near it. There's parts I don't like but it's not like Homesteading, slavery, Chinese exclusion, redlining, Japanese internment, the klan, and Jim Crow were great.
This is American behavior: crude, cruel, hostile, arrogant, and proudly ignorant.
Richard Hofstadter wrote about Americans acting this way in the 1960s.
Look at the Johnson-Reed Act of 1924, stood for decades. It's not like those sentiments went away...
And there's no "good states" either - the California Constitution in 1879 set up a racial apartheid system against Chinese people. Even had a second called "The Chinese".
Oregon was admitted to the Union explicitly with a "whites only" clause.
The Declaration of Independence even has wild conspiracy theories about "merciless Indian savages"
No amount of empirical evidence will make Americans realize this because it gives them a frowny face.
I see people treating LLMs like programming languages and trying to give very precise and detailed instructions. Essentially pseudo-coding or writing english instead of C++. I find that being vague and iterating is more powerful. If you want to give a detailed spec that fully describes the program then you might as well write that program?
Basically treat the LLM as a human. Not as a computer. Like a junior developer or an intern (for the most part).
That said you need to know what to ask for and how to drive the LLM in the correct direction. If you don't know anything you're likely not going to get there.
The other simple method is to only accept certain system prompts
I've been meaning to do some dumb little proxy system where all your i/o can pass through any specified system such as a web page, harness, whatever...
Essentially a local model toolcalls to an "Oracle" which is just something like a wrapper around Claude code or anything you've figured out how to scrape and then you talk to the small model that mostly uses the Oracle and.... There you go.
There's certainly i/o shuffling and latency but given model speeds and throughput it'll be relatively very small
Now people probably care
Doesn't mean I know how to market it, I'll certainly fail at that, but at least I can build it
It needs to support tool calling and many of the quantized ggufs don't so you have to check.
I've got a workaround for that called petsitter where it sits as a proxy between the harness and inference engine and emulates additional capabilities through clever prompt engineering and various algorithms.
They're abstractly called "tricks" and you can stack them as you please.
I've got an interesting hack brewing for extremely hassle free tool orchestration - basically think along the lines of .bash_profile level simplicity... Maybe I'll get that out tomorrow
I've gone through about 500M tokens on this model already. They've got some free inferencing options (such as on openrouter) ... $0 is hard to beat and it's creating not-crap.
reply