More from @godofprompt
Nov 5
Google Search is so dead â ď¸
Iâve been using Perplexity AI for 6 months it now handles every research brief, competitor scan, and content outline for me.
Hereâs how I replaced Google (and half my workflow) with a single AI tool:
1. Deep Research Mode
Prompt:
âYouâre my research assistant. Find the latest studies, reports, and articles on [topic]. Summarize each source with: Title | Date | Key Finding | Source link.â
â Returns citations + structured summaries faster than any Google search.
2. âExplain Like Iâm Smartâ Mode
Prompt:
âExplain [complex concept] as if I have deep domain knowledge but limited time. Include: key principles, debates, and real-world applications.â
â Replaces 10 tabs of random articles with one expert summarâŚ
More from @godofprompt
Nov 5
Google Search is so dead â ď¸
Iâve been using Perplexity AI for 6 months it now handles every research brief, competitor scan, and content outline for me.
Hereâs how I replaced Google (and half my workflow) with a single AI tool:
1. Deep Research Mode
Prompt:
âYouâre my research assistant. Find the latest studies, reports, and articles on [topic]. Summarize each source with: Title | Date | Key Finding | Source link.â
â Returns citations + structured summaries faster than any Google search.
2. âExplain Like Iâm Smartâ Mode
Prompt:
âExplain [complex concept] as if I have deep domain knowledge but limited time. Include: key principles, debates, and real-world applications.â
â Replaces 10 tabs of random articles with one expert summary.
Read 12 tweets
Nov 4
đ¨ China just built Wikipediaâs replacement and it exposes the fatal flaw in how we store ALL human knowledge.
Most scientific knowledge compresses reasoning into conclusions. You get the âwhatâ but not the âwhy.â This radical compression creates what researchers call the âdark matterâ of knowledge the invisible derivational chains connecting every scientific concept.
Their solution is insane: a Socrates AI agent that generates 3 million first-principles questions across 200 courses. Each question gets solved by MULTIPLE independent LLMs, then cross-validated for correctness.
The result? A verified Long Chain-of-Thought knowledge base where every concept traces back to fundamental principles.
But hereâs where it gets wild... they built the Brainstorm Search Engine that does âinverse knowledge search.â Instead of asking âwhat is an Instanton,â you retrieve ALL the reasoning chains that derive it: from quantum tunneling in double-well potentials to QCD vacuum structure to gravitational Hawking radiation to breakthroughs in 4D manifolds.
They call this the âdark matterâ of knowledge finally made visible.
SciencePedia now contains 200,000 entries spanning math, physics, chemistry, biology, and engineering. Articles synthesized from these LCoT chains have 50% FEWER hallucinations and significantly higher knowledge density than GPT-4 baseline.
The kicker? Every connection is verifiable. Every reasoning chain is checked. No more trusting Wikipediaâs citations you see the actual derivation from first principles.
This isnât just better search. Itâs externalizing the invisible network of reasoning that underpins all science.
The âdark matterâ of human knowledge just became visible.
The pipeline is genius.
A Planner generates problem thumbnails. A Generator expands them into specific questions with verifiable answers. Then multiple independent Solver agents (different LLMs) attack the same problem.
Only answers with consensus survive. Hallucinations get filtered automatically.
This is the architecture that changes everything.
User query â Keywords extraction â LCoT Knowledge Base retrieval â Ranking by cross-disciplinary relevance â LLM Synthesizer weaves verified chains into coherent articles.
âInverse knowledge searchâ discovers HOW concepts connect, not just WHAT they are.
Read 7 tweets
Oct 30
Holy shit... Alibaba just dropped a 30B parameter AI agent that beats GPT-4o and DeepSeek-V3 at deep research using only 3.3B active parameters.
Itâs called Tongyi DeepResearch and itâs completely open-source.
While everyoneâs scaling to 600B+ parameters, Alibaba proved you can build SOTA reasoning agents by being smarter about training, not bigger.
Hereâs what makes this insane:
The breakthrough isnât size itâs the training paradigm.
Most AI labs do standard post-training (SFT + RL).
Alibaba added âagentic mid-trainingâ a bridge phase that teaches the model how to think like an agent before it even learns specific tasks.
Think of it like this:
Pre-training = learning language Agentic mid-training = learning how agents behave Post-training = mastering specific agent tasks
This solves the alignment conflict where models try to learn agentic capabilities and user preferences simultaneously.
The data engine is fully synthetic.
Zero human annotation. Everything from PhD-level research questions to multi-hop reasoning chains is generated by AI.
They built a knowledge graph system that samples entities, injects uncertainty, and scales difficulty automatically.
20% of training samples exceed 32K tokens with 10+ tool invocations. Thatâs superhuman complexity.
The results speak for themselves:
32.9% on Humanityâs Last Exam (vs 26.6% OpenAI DeepResearch) 43.4% on BrowseComp (vs 30.0% DeepSeek-V3.1) 75.0% on xbench-DeepSearch (vs 70.0% GLM-4.5) 90.6% on FRAMES (highest score)
With Heavy Mode (parallel agents + synthesis), it hits 38.3% on HLE and 58.3% on BrowseComp.
Whatâs wild: They trained this on 2 H100s for 2 days at <$500 cost for specific tasks.
Most AI companies burn millions scaling to 600B+ parameters.
Alibaba proved parameter efficiency + smart training >>> brute force scale.
The bigger story?
Agentic models are the future. Models that autonomously search, reason, code, and synthesize information across 128K context windows.
Tongyi DeepResearch just showed the entire industry theyâre overcomplicating it.
Full paper: arxiv. org/abs/2510.24701 GitHub: github. com/Alibaba-NLP/DeepResearch
The architecture is beautifully simple.
Itâs vanilla ReAct (reasoning + acting) with context management to prevent memory overflow.
No complex multi-agent orchestration. No rigid prompt engineering.
Just pure scalable computation exactly what âThe Bitter Lessonâ predicted would win.
Hereâs how they synthesized massive agent behavior data without humans:
1. Question Synthesis - multi-hop reasoning problems 2. Planning Action - problem decomposition 3. Reasoning Action - logical chains across noisy data 4. Decision-Making Action - explicit choice modeling
All generated from an entity-anchored knowledge graph.
Read 12 tweets
Oct 29
deepmind just published something wild đ¤Ż
they built an AI that discovers its own reinforcement learning algorithms.
not hyperparameter tuning.
not tweaking existing methods.
discovering ENTIRELY NEW learning rules from scratch.
and the algorithms it found were better than what humans designed.
hereâs what they did:
⢠created a meta-learning system that searches the space of possible RL algorithms ⢠let it explore millions of algorithmic variants automatically ⢠tested each on diverse tasks and environments ⢠kept the ones that worked, evolved them further ⢠discovered novel algorithms that outperform state-of-the-art human designs like DQN and PPO
the system found learning rules humans never thought of. update mechanisms with weird combinations of terms that shouldnât work but do.
credit assignment strategies that violate conventional RL wisdom but perform better empirically.
the discovered algorithms generalize across different tasks. theyâre not overfit to one benchmark.
they work like principled learning rules should, and theyâre interpretable enough to understand WHY they work.
we are discovering the fundamental math of how agents should learn.
led by david silver (alphago, alphazero creator). published in nature. fully reproducible.
the meta breakthrough: we now have AI systems that can improve the way AI systems learn.
the thing everyone theorized about? itâs here.
why this breaks everything:
RL progress has been bottlenecked by human intuition.
researchers have insights, try variations, publish.
it takes years to go from Q-learning to DQN to PPO.
now you just let the machine search directly.
millions of variants in weeks instead of decades of human research.
but hereâs the compounding part: each better learning algorithm can be used to discover even better ones.
you get recursive improvement in the narrow domain of how AI learns.
humans took 30+ years to get from basic Q-learning to modern deep RL.
an automated system can explore that space and find non-obvious improvements humans would never stumble on.
this is how you get to superhuman algorithm design.
not by making humans smarter, but by removing humans from the discovery loop entirely.
when david silverâs lab publishes in nature about âmachines discovering learning algorithms for themselves,â you pay attention. this is the bootstrap beginning.
paper: nature.com/articles/s4158âŚ
TL;DR for normal people:
imagine youâre teaching a robot to learn. humans spent decades figuring out the âbest waysâ to teach machines (called learning algorithms).
deepmind built an AI that invents its own teaching methods. and they work better than ours.
why it matters: â we donât wait for human breakthroughs anymore â AI searches millions of strategies weâd never think of â each better algorithm helps discover even better ones (compounding) â weâre automating the process of making AI smarter
itâs like having a student who figures out better ways to study, then uses those better methods to figure out even better ones, recursively.
the âAI improving AIâ loop is here. published. working.
the next generation of breakthroughs in how machines learn might be designed entirely by machines.
Read 4 tweets
Oct 21
đ¨ Academia just got an upgrade.
A new paper called Paper2Web might have just killed the static PDF forever.
It turns research papers into interactive websites complete with animations, videos, and embedded code using an AI agent called PWAgent.
Hereâs why itâs a big deal:
⢠10,700 papers analyzed to build the first dataset + benchmark for academic webpages. ⢠Evaluates sites on connectivity, completeness, and interactivity (even runs a âPaperQuizâ to test knowledge retention). ⢠Outperforms arXiv HTML and alphaXiv by 28%+ in structure and usability.
Essentially, it lets you publish living papers where readers can explore, interact, and even quiz themselves.
The PDF era is ending.
Your next research paper might talk back.
github. com/YuhangChen1/Paper2All
Today, most âHTML paperâ attempts fail because they just convert text not meaning.
Paper2Web fixes that.
It built the first dataset of 10,700 paperâwebsite pairs across top AI conferences to actually learn what makes research websites effective.
Itâs not just tech itâs an entire academic web design benchmark.
Every paper in the dataset was labeled as static, multimedia, or interactive.
The findings are wild:
Only 9.8% of academic websites are interactive. Over 42% are still just static text dumps.
Meaning: the research web is still trapped in 2005. Paper2Web is the first system to quantify why and fix it.
Read 7 tweets
Oct 20
đ¨ DeepSeek just did something wild.
They built an OCR system that compresses long text into vision tokens literally turning paragraphs into pixels.
Their model, DeepSeek-OCR, achieves 97% decoding precision at 10Ă compression and still manages 60% accuracy even at 20Ă. That means one image can represent entire documents using a fraction of the tokens an LLM would need.
Even crazier? It beats GOT-OCR2.0 and MinerU2.0 while using up to 60Ă fewer tokens and can process 200K+ pages/day on a single A100.
This could solve one of AIâs biggest problems: long-context inefficiency. Instead of paying more for longer sequences, models might soon see text instead of reading it.
The future of context compression might not be textual at all. It might be optical đď¸
github. com/deepseek-ai/DeepSeek-OCR
1. Vision-Text Compression: The Core Idea
LLMs struggle with long documents because token usage scales quadratically with length.
DeepSeek-OCR flips that: instead of reading text, it encodes full documents as vision tokens each token representing a compressed piece of visual information.
Result: You can fit 10 pages worth of text into the same token budget it takes to process 1 page in GPT-4.
2. DeepEncoder - The Optical Compressor
Meet the star: DeepEncoder.
It uses two backbones SAM (for perception) and CLIP (for global vision) bridged by a 16Ă convolutional compressor.
This allows it to maintain high-res understanding without exploding activation memory.
The encoder converts thousands of image patches â a few hundred compact vision tokens.
Read 8 tweets