Sounds like you're betting that the performance users experience today will be the same as the performance they'll expect tomorrow. I wouldn't take that bet.
You mean that if you were Anthropic, you'd build the data centers on every continent? Can you explain your reasoning?
We're talking about billions of dollars of extra capex if you take the "let's build them everywhere" side of the bet instead of "let's build them in the cheapest possible place" side. It seems to me that you'd have to be really sure that you need the data center to be somewhere uneconomical. I think if you did build them in the cheap place, it's a safe bet that you'll always have at least enough latency-insensitive workloads to fill it up. I doubt that we would transition entirely to latency-sensitive workloads in the future, and that's what would have to happen for my side of the bet to go wrong. The other side goes wrong if we don't see a dramatic uptick in latency-sensitive inference workloads. As another comment pointed out, voice agents are the one genuinely latency-sensitive cloud inference workload we have right now; they do need low latency for it. Such workloads exist, but it's a slim percentage so far.
I believe I'm taking the safe bet that lets Anthropic make hay while the sun shines without risking a major misstep. Nothing stops them from using their own data centers for cheap slow "base load" while still using cloud partners for less common specialized needs. I just can't see why they would build the international data centers to reduce cloud partner costs on latency-sensitive workloads before those workloads actually show up in significant numbers.
I suspect that the touch bar served its likely real purpose: to ship an ARM CPU with a secure enclave in the machines so that we could have Touch ID without needing to wait for Apple Silicon. Everything other than that was gravy, an interesting experiment.
Yeah, congress forces the military to contract out to companies in enough congressional districts to secure passage of the legislation. We basically force these companies into byzantine and inefficient supply chains because we treat it all as a jobs program.
Maybe in the future but with the current models I found the constantly accessible memories to be an impediment. I don't want models to record and repeat mistakes or suboptimal strategies.
Gemini (just in the browser) has been really bad about conflating a bunch of similar projects. It remembers "oh, you have a home server that does XYZ", so my new home server that's doing ZYX instead must be the same system.
Is it lifting him up? It's certainly irrelevant, is my point. My assumption then is that it's because it's supposed to be surprising. 'Hobbyist spends 20y on their hobby' isn't that surprising, even if the hobby is interesting; instead of letting the story stand on that interest, they're attempting to add 'shock and awe'.
reply