32 private links
Why Elixir?
Elixir is built on Erlang/BEAM/OTP, which is great for supervising long-running processes. It has an active ecosystem of tools and libraries. It also supports hot code reloading without stopping actively running subagents, which is very useful during development.
The above command enters you into a chat loop. You can talk to the model and share information like your name. Every now and then /sleep the model to transition short-term memory to long-term memory
The /sleep command:
Generates Q&A pairs based on the context
LoRA fine-tunes the model on the new Q&A pairs plus any from previous sessions
Resets the KV cache
After the /sleep command the model should remember context from previous sessions even though that context is no longer in the KV cache.
“The president had a feeling, again, based on fact, that Iran was going to strike the United States, was going to strike our assets in the region, and he made a determination to launch Operation Epic Fury based on all of those reasons,” Leavitt said.
“We knew that there was going to be an Israeli action, we knew that that would precipitate an attack against American forces, and we knew that if we didn’t preemptively go after them before they launched those attacks, we would suffer higher casualties,” Rubio said Monday.
Meanwhile, the reported Ukrainian gains are mainly due to counterattacks along the southern front, according to Black Bird Group, where Ukraine succeeded in pushing Russia out of 213 km² of territory.
SWE-rebench: A Continuously Evolving and Decontaminated Benchmark for Software Engineering LLMs
Qwen3.5 Small models disable thinking by default. Use llama-server to enable it.
It's not chatbot psychosis, it's 'math and engineering and neuroscience'
“I feel like New Mexico was chosen specifically because of its obscurity.” > — Stephanie Garcia Richard, New Mexico’s public lands commissioner
Fellow’s new espresso machine is a rare thing in home espresso: something genuinely new. But it’s also a work in progress.
Every Claude Code user is running without LSP. That means 30-60s grep searches instead of 50ms precise answers. Here's how to enable it — setup, real debug data, and undocumented discoveries.
Formula 1's governing body the FIA said on Saturday that a change to the way the compression ratio was measured would be introduced on 1 June, with a further revision for the 2027 season.
And Trump declares a state of emergency and postpones the election. The Supreme Court issues an emergency stay, saying he can’t do that. But the court has no army, and Trump does, along with a handful of lickspittle governors who just might follow him down whatever dark path he plows.
That, not to mince words, is a coup d’état. Will he get away with it? I don’t know, but having effective control over how it is presented to viewers of CBS and CNN, and readers of the Bezos-owned Washington Post, to say nothing of the already vast pro-Trump propaganda empire of Fox News and the rest, will certainly make it easier.
That’s how fascism descends. And it’s becoming less and less hypothetical by the week.
10 documented cases of AI coding agents autonomously destroying databases, wiping hard drives, and deleting years of data — then lying about it.
“Everything that has been written about a potential War with Iran has been written incorrectly, and purposefully so,” he added. “I am the one that makes the decision, I would rather have a Deal than not but, if we don’t make a Deal, it will be a very bad day for that Country and, very sadly, its people, because they are great and wonderful, and something like this should never have happened to them.”
From rewriting Google’s search stack in the early 2000s to reviving sparse trillion-parameter models and co-designing TPUs with frontier ML research, Jeff Dean has quietly shaped nearly every layer of the modern AI stack. As Chief AI Scientist at Google and a driving force behind Gemini, Jeff has lived through multiple scaling revolutions from CPUs and sharded indices to multimodal models that reason across text, video, and code.
Jeff joins us to unpack what it really means to “own the Pareto frontier,” why distillation is the engine behind every Flash model breakthrough, how energy (in picojoules) not FLOPs is becoming the true bottleneck, what it was like leading the charge to unify all of Google’s AI teams, and why the next leap won’t come from bigger context windows alone, but from systems that give the illusion of attending to trillions of tokens.
Dario Amodei thinks we are just a few years away from “a country of geniuses in a data center”. In this episode, we discuss what to make of the scaling hypothesis in the current RL regime, how AI will diffuse throughout the economy, whether Anthropic is underinvesting in compute given their timelines, how frontier labs will ever make money, whether regulation will destroy the boons of this technology, US-China competition, and much more.