AI Datacenters are not how all or probably even most HNers get paid...
> Most data centers colo multiple types of compute, not just those dedicated to inference or model training. Additonally, strangling the economics of the infrastructure layer makes entire ecosystems move abroad.
Sure but we are talking about whether the enormous investment into AI infrastructure is prudent or not. Also I reckon most people on here made a living just fine before everything moved to remote data centers, and many if not most HNers workloads could run on individual machines... But that's another conversation.
I think language grammars are the an interesting way to define a ruleset too. Forget REST API's or MCP Servers for a second... Define a domain specific language, and let the language model generate a valid instruction within the confines of that grammar.
Than pass the program, your server or application can parse the instructions and work from the generated AST to do all sorts of interesting things, within the confines of your language features.
It's verifiable, since you are providing within the defined grammar, and with the parser provided.
It is implicitly sandboxed by the powers you give (or rather exclude) to your runtime via an interpreter/compiler
I've tried this before for a grammar I defined for searching documents, and found it to be quite good at creating valid often complex search instructions.
Previously I made a chrome extension that removes them from web... But I haven't updated it in a while. Basically just inspects the HTML/CSS patterns of the shorts components and removes them from the page. You could probably code/vibe code a similar extension in 10m.
Definitely could be, but in the time I spent talking to the 4-bit models in comparison to the 16-bit original it seemed surprisingly capable still. I do recommend benchmarking quantized models at the specific tasks you care about.
yes, but the difference between one model and one 4x larger is usually a lot more than that.
It is not a question of do a run Qwen 8b at bf16 or a quantized version. It more of a question of do I run Qwen 8b at full precision or do I run a quantized version of Qwen 27b.
You will find that you are usually better off with the larger model.
I'm no expert but to me, what's particularly silly about "breaking encryption" is it does nothing to prevent using user agents from employing their own encryption layers over other messaging system like gpg/pgp or others. So this does nothing to stop someone who is intent on hiding illegal content and it decreases security and privacy for the average user.
It should, but as far as I know you have zero guarantees about that. I just hope there's privacy organizations and / or hackers that continuously verify these claims. Of course, Amazon can push an update at any time to change this, at which point it'll be too late to think "hmm, 1984 warned us about this".
reply