AI stream

AI Posts

A readable stream of AI posts. Open one post to focus on the original content.

This week
@emollick
@emollick Mar 29, 2026 Research paper

Good news; The same research team in a more recent RCT found that AIs prompted to act as a tutor improved learning outcomes! https://papers.ssrn.com/sol3/papers.cfm?abstract_id=6423358

Likes: 85 Reposts: 1 Views: 2,666 Images: 1
Score 6
@elonmusk
@elonmusk Mar 29, 2026 Tool announcement

Grok automatically translating and recommending 𝕏 posts from other languages is starting to work

Likes: 130,535 Reposts: 7,885 Views: 30,233,584
Score 3
@RoundtableSpace
@RoundtableSpace Mar 29, 2026 Tool announcement

Anthropic just dropped the Claude Certified Architect exam. 13 free courses, 60 questions, 2 hours. Build, orchestrate, and master multi-agent AI systems.

Likes: 792 Reposts: 59 Views: 201,158 Videos: 1
Score 5
@zenorocha
@zenorocha Mar 29, 2026 Tool announcement

Agents were getting confused by our pricing. Turns out JS-based sliders are not the best way for them to parse things. So today I’m shipping https://resend.com/pricing.md Plus, content negotiation (Accept: text/markdown). Pricing is now machine-readable for AI agents and LLMs.

Likes: 403 Reposts: 10 Views: 30,039 Images: 1
Score 6
@fchollet
@fchollet Mar 29, 2026 Opinion editorial

Let me explain what I mean using your chess analogy... Imagine a world where chess doesn't exist. In this world, humanity encounters an alien species, and they say "let's play a game of Glurg, it's our traditional pastime. Here are the rules, see you tomorrow" -- and it's the rules of chess. My claim is that following this interaction, a working group of the world's best minds, leveraging current externalized cognitive infrastructure (computers, the internet, etc.) would be able to analyze the rules and develop a working 3000 Elo chess engine within 24 hours, in time for the match. Give them an extra 3 weeks and they'd have a 3500 Elo engine that's 10x more compute efficient. So human intelligence is already at a level where we can go from "here are the rules" to "I can play at 3000 Elo" immediately. Not optimal yet, but not too far off.

Likes: 368 Reposts: 20 Views: 76,321
Score 4
@GoogleCloudTech
@GoogleCloudTech Mar 29, 2026 Tutorial

Vertex AI helps improve model performance with minimal infrastructure overhead. Check out our new codelab to learn how to fine-tune Gemini 2.5 Flash and walk through the complete SFT workflow using the Vertex AI SDK for Python → https://cloud.google.com/blog/topics/developers-practitioners/mastering-model-adaptation-a-guide-to-fine-tuning-on-google-cloud?utm_source=twitter&utm_medium=unpaidsoc&utm_campaign=fy26q1-googlecloud-blog-ai-in_feed-no-brand-global&utm_content=-&utm_term=-&linkId=59008331

Likes: 91 Reposts: 6 Views: 6,276 Images: 1
Score 4
@RoundtableSpace
@RoundtableSpace Mar 29, 2026 Code sample

SOMEONE PASTED GOOGLE'S TURBOQUANT PAPER INTO CLAUDE & BUILT A TRADING BOT IN MINUTES THE BOT MADE 3,317 PREDICTIONS AND TURNED $1,500 INTO $83,115 ON POLYMARKET IN 72 HOURS THE PAPER WAS FREE. CLAUDE COSTS $20 A MONTH

Likes: 990 Reposts: 92 Views: 203,991 Videos: 1
Score 5
@negligible_cap
@negligible_cap Mar 29, 2026 Performance

Yesterday I gave Claude Cowork everything it needed to do my taxes and sent it loose on TurboTax Claude absolutely flew through the process. Incredibly confident, very few follow up questions It’s so over for accountants

Likes: 11,019 Reposts: 166 Views: 1,606,759 Images: 1
Score 3
@thetripathi58
@thetripathi58 Mar 29, 2026 Research paper

🚨 Cambridge researchers just tested what happens when you overload an AI's memory with irrelevant data. They found a complete collapse of modern RAG systems. Not a minor hallucination. A total failure of the exact retrieval architecture that every enterprise AI relies on to access private data. The models simply drowned in the noise. The researchers tested standard Retrieval-Augmented Generation (RAG) and filtering models like Self-RAG. They fed them information but slowly increased the ratio of distracting, low-quality documents. Here is what they found. Current read-time filtering failed completely. When the ratio of distractors hit 8:1, the accuracy of standard RAG systems plummeted to 0%. The AI lost the ability to find the truth. It exposed a massive architectural flaw. We currently store every single document an AI reads, regardless of quality, and force the model to sort through the garbage at query time. It is highly inefficient and fundamentally broken. The biological fix. The researchers built a new system called "Write-Time Gating" modeled after the human hippocampus. Instead of saving everything, it evaluates novelty, reliability, and source reputation before the data is even stored. And then there is the finding that changes how we build AI: hierarchical archiving. When beliefs update, the system does not delete the old data. It deprioritizes it, maintaining a version history just like the human brain. The result? The write-gated system maintained 100% accuracy even at massive distractor scales, all while costing one-ninth the compute of current systems. The researchers made it clear. When you dump raw, unfiltered data into a database and expect the LLM to figure it out later, you are building a system designed to fail at scale. No reliable retrieval. No cost control. No accuracy guarantees. Nothing. Right now, companies are building massive vector databases, throwing every piece of corporate documentation into them, and assuming the AI will magically find the signal in the noise. Stop treating AI memory like a hard drive. Start treating it like a biological filter. Build the gate at the entrance, not the exit.

Likes: 90 Reposts: 47 Views: 19,261 Images: 1
Score 5
@midudev
@midudev Mar 29, 2026 Tutorial

Curso de Microsoft para crear Agentes de IA. En Español, desde cero y son 12 lecciones. MCP, RAG, multiagentes y más ↓

Likes: 936 Reposts: 155 Views: 29,491 Videos: 1
Score 3
@ihtesham2005
@ihtesham2005 Mar 29, 2026 Tool announcement

🚨BREAKING: Someone compiled every free AI agents resource from Microsoft, Google, OpenAI, Anthropic, and Hugging Face in one place. You can learn: - LangChain, LangGraph, CrewAI, AutoGen, OpenAI Swarm - RAG agents, multi-agent systems, task management, conversational AI - No-code agents with n8n, Vapi, and low-code workflows - 100+ resources with full courses, notebooks, and video lectures 100% Opensource.

Likes: 128 Reposts: 31 Views: 12,553 Images: 1
Score 5
@marcusyul
@marcusyul Mar 29, 2026 Tutorial

🚨ÚLTIMA HORA: CLAUDE YA PUEDE CREAR APPS COMPLETAS DESDE CERO Sin equipo. Sin código. Sin dinero. Solo necesitas saber como usarlo. Aquí tienes 8 prompts para empezar a probarlo des de ya: 🔖 Guárdalos, lo agradecerás.

Likes: 2,216 Reposts: 332 Views: 137,766 Images: 2
Score 4
@svpino
@svpino Mar 29, 2026 Opinion editorial

I still remember when people thought "prompt engineering" was going to become a real career.

Likes: 8,603 Reposts: 316 Views: 354,334
Score 3
@cgtwts
@cgtwts Mar 29, 2026 Ai tools

Anthropic CEO: “ I have engineers within anthropic who don’t write any code, they just let Claude write the code and they edit it and look it over” “At anthropic writing code means designing the next version of Claude it self, so we essentially have Claude designing the next version of Claude itself, not completely but most of it”. In the last 52 days, the Claude team dropped 50+ major feature launches. This is literally INSANE.

Likes: 7,316 Reposts: 616 Views: 1,747,124 Videos: 1
Score 3
@HuggingPapers
@HuggingPapers Mar 29, 2026 Research paper

MSA breaks the 100M token barrier Memory Sparse Attention achieves unprecedented 100M token context lengths with near-linear complexity. The architecture maintains 94% accuracy at 1M tokens while outperforming RAG systems and frontier models, using end-to-end sparse attention with document-wise RoPE.

Likes: 248 Reposts: 35 Views: 13,060 Images: 1
Score 4
@gregisenberg
@gregisenberg Mar 29, 2026 Tip trick

How to 10x your Claude with 4 .md files

Likes: 2,766 Reposts: 230 Views: 183,659 Videos: 1
Score 4
@marmaduke091
@marmaduke091 Mar 29, 2026 Model release

🚨 DeepSeek just got a HUGE upgrade It's been upgraded on the web app, and the outputs are much better. Check it out! They could be fine-tuning the outputs with their new generation (V4) model, or this could even be our first look at the V4 series of models, not confirmed yet.

Likes: 422 Reposts: 16 Views: 64,843 Images: 3
Score 3
@mokyusub
@mokyusub Mar 29, 2026 Fine tuning

今日も1日、お疲れ様でした(`・ω・´)ゞ 色々調整しながら、最カワを目指して\( *'ω'*)/ pony系からillustrious系に移行する為に、頑張ってみる! 自作LoRAで、だいぶもなかちゃんの顔になってると思う🥴✨️

Likes: 515 Reposts: 20 Views: 8,950 Images: 3
Score 4
@_vmlops
@_vmlops Mar 29, 2026 Ai tools

someone just dockerized an entire AI coding workstation and it's kind of insane one docker compose up and you get: → Claude Code with a browser UI → Gemini, Codex, Cursor, TaskMaster CLIs → Playwright + Chromium, pre-configured → 50+ dev tools (pandas, ffmpeg, prisma, gh...) no config. no debugging "why won't chromium run in docker" uses your existing Claude Max/Pro subscription

Likes: 152 Reposts: 24 Views: 10,484 Images: 1
Score 5
@AiCreatorS1881
@AiCreatorS1881 Mar 29, 2026 Fine tuning

ZITのLoRA、LTX-2.3用LoRAを作成しました。 Ruriさんです。 久々のモデル公開です。モデル制作者であることを思い出しました(笑) https://huggingface.co/Kotajiro/ZIT_ruri_LoRA https://huggingface.co/Kotajiro/LTX23-ruri_LoRA

Likes: 100 Reposts: 9 Views: 6,489 Images: 1
Score 5
@ujjwalscript
@ujjwalscript Mar 29, 2026 Opinion editorial

Prompt Engineering is a SCAM. Please take it off your resume. The biggest lie on Tech Twitter right now is that you need to be an "AI Whisperer" to build software in 2026. Here is the reality check: If you need a 600-word prompt with 14 bullet points just to generate a stable React component... the AI isn't the problem. Your architecture is garbage. We spent the last few years teaching people to type "Act as a senior 10x developer and..." Modern models are now smart enough to ignore the fluff. They don't need magic words. They need Constraints. What actually separates a Senior Engineer from a "Prompt Bro" today: 1System Boundaries: Knowing exactly where your Next.js frontend stops and your backend microservice begins. 2Data Contracts: Defining strict schemas and types before you let the AI write a single loop. 3State Management: The one thing autonomous agents still hallucinate on a daily basis. Stop trying to trick the machine with psychological hacks. Start feeding it clean, modular system architecture. If your only technical moat is "writing really good prompts," someone who actually understands database indexing is going to take your job by Q3. Good engineering fixes bad prompting. Good prompting cannot fix bad engineering.

Likes: 145 Reposts: 7 Views: 15,731
Score 5
@emollick
@emollick Mar 29, 2026 Research paper

The research team (including @hamsabastani who is on X) found that letting students just use AI resulted in them using it to accidentally shortcut learning But both that study and a separate RCT found that AIs prompted to act as a tutor improved learning https://papers.ssrn.com/sol3/papers.cfm?abstract_id=6423358

Likes: 382 Reposts: 62 Views: 61,792 Images: 1
Score 4
@emollick
@emollick Mar 29, 2026 Ai research

I wish the corpus from the era was big enough that we could do counterfactual history. I wonder if you can generate enough synthetic data to get it to work. If I could, I'd love to assign a Victorian agentic scientist to discover the luminous aether. Very @nealstephenson

Likes: 142 Reposts: 3 Views: 21,428
Score 5
@emollick
@emollick Mar 29, 2026 Ai research

Very cool work, I wonder what other eras have a large enough corpus for training? https://huggingface.co/spaces/tventurella/mr_chatterbox

Likes: 145 Reposts: 3 Views: 19,971
Score 5
@emollick
@emollick Mar 29, 2026 Ai research

Want to talk to the past? Here is an LLM "trained entirely from scratch on a corpus of over 28,000 Victorian-era British texts published between 1837 and 1899, drawn from a dataset made available by the British Library." Quite different from an LLM roleplaying a Victorian.

Likes: 2,518 Reposts: 216 Views: 215,937 Images: 2
Score 3
@ClaudeCodeLog
@ClaudeCodeLog Mar 29, 2026 Release announcement

Claude Code 2.1.87 is now available. 1 CLI change Highlights: • Cowork Dispatch messages deliver reliably, ensuring dispatched communications reach recipients Full details are in thread ↓

Likes: 926 Reposts: 49 Views: 129,927 Videos: 1
Score 5
@bcherny
@bcherny Mar 29, 2026 Tool announcement

Working on improving this. A bunch of efficiency wins incoming.

Likes: 708 Reposts: 5 Views: 40,832
Score 3
@fchollet
@fchollet Mar 29, 2026 Opinion editorial

I do believe that a large collective of the smartest humans, aided by external tools, sits very close to the optimality bound -- i.e. humans should be able to solve any solvable problem (where the required information is available) if they pay enough attention to it

Likes: 345 Reposts: 15 Views: 35,542
Score 4
@fchollet
@fchollet Mar 29, 2026 Opinion editorial

One of the biggest misconceptions people have about intelligence is seeing it as some kind of unbounded scalar stat, like height. "Future AI will have 10,000 IQ", that sort of thing. Intelligence is a conversion ratio, with an optimality bound. Increasing intelligence is not so much like "making the tower taller", it's more like "making the ball rounder". At some point it's already pretty damn spherical and any improvement is marginal. Now of course smart humans aren't quite at the optimal bound yet on an individual level, and machines will have many advantages besides intelligence -- mostly the removal of biological bottlenecks: greater processing speed, unlimited working memory, unlimited memory with perfect recall... but these are mostly things humans can also access through externalized cognitive tools.

Likes: 1,804 Reposts: 156 Views: 263,933
Score 3
@bcherny
@bcherny Mar 29, 2026 Tool announcement

A weird part of working at Anthropic: getting a few of these each day

Likes: 7,214 Reposts: 110 Views: 906,636 Images: 1
Score 3