π§βπ« ChatGPT launches study mode to encourage academic use
ChatGPT debuts study mode, Google beefs up NotebookLM video tools, Tencent unveils Hunyuan3D, Alibaba ships Wan2.2, GPT-5 rumor, Claude prompt leak.
Read time: 10 min
π Browse past editions here.
( I publish this newletter daily. Noise-free, actionable, applied-AI developments only).
β‘In todayβs Edition (29-July-2025):
π§βπ« ChatGPT launches study mode to encourage academic use
π Google Enhances NotebookLM with AI Video Summaries and Studio Upgrades
π‘ Tencent Hunyuan just released Hunyuan3D World Model 1.0!
π΅ Chinaβs open-source AI engine is still running hot, Alibabaβs Wan2.2 video model launched.
ποΈ Byte-Size Briefs:
As per widely circulating rumor, OpenAI is dropping GPT-5 next week! with 1M token input window, 100k output tokens.
Somebody shared the full Claude code system prompt on GitHub.
π§βπ OPINION: Six Strange Aliases, One Giant Model: The GPTβ5 Scavenger Hunt
π§βπ« ChatGPT launches study mode to encourage academic use
OpenAI announced Study Mode for ChatGPT, a new feature that fundamentally changes how students interact with artificial intelligence by withholding direct answers in favor of Socratic questioning and step-by-step guidance.
Available to logged-in Free, Plus, Pro, Team users, with availability in ChatGPT Edu coming in the coming weeks.
It will guide students with questions instead of answers, aiming for deeper learning. OpenAI hopes the tutor style flow captures a slice of the $80.5B edβtech market. The system even resists studentsβ attempts to obtain quick answers. When prompted with βjust give me the answer,β Study Mode responds that βthe point of this is to learn, not just to give you the answer.β
Instead of spitting out a solution, the bot fires Socratic prompts that probe what the learner already knows, offers hints, then asks for the next step. Quick toggling lets users shift between normal chat and this coaching mode, so it fits homework, exam prep, or new topics.
Behind the curtain, OpenAI injects custom system instructions written with teachers across 40 institutions. The rules control cognitive load, scaffold concepts, check retention with mini quizzes, and adjust depth from novice to advanced. OpenAI will later fuse this behavior into future models after real world feedback.
Early testers report big confidence boosts and finally cracking tough ideas like positional encodings. Rivals Anthropic and Google chase similar aims, showing that guided questioning is the next battleground as 1 in 3 U.S. college students lean on chatbots.
π Google Enhances NotebookLM with AI Video Summaries and Studio Upgrades
πΉ Video Overviews allow users to turn dense multimedia, such as raw notes, PDFs, and images, into digestible visual presentations.
The video overviews feature enables NotebookLM to produce narrated video summaries, complete with visuals and voiceovers, drawing directly from user-uploaded sources like PDFs, web pages, or notes. This isnβt just a simple text-to-video conversion; itβs an AI-orchestrated presentation that highlights key insights, timelines, and connections, making it ideal for quick briefings or educational content.
It also lets each notebook hold multiple custom outputs. NotebookLM already offered Audio Overviews, tiny podcasts that read course packs or legal briefs so users could learn handsβfree. Handy in traffic, but charts, formulas, and process diagrams still begged for visuals.
Video Overviews plugs that hole. An AI host builds short narrated slides, pulling images, diagrams, quotes, and raw numbers straight from the upload, then plays the set like a brisk explainer. Prompts let users set topics, goals, and audience depth.
The updated Studio panel adds 4 tiles and stores many outputs of each type, making multilingual audio, roleβspecific video, or chapterβbased mind maps possible in one notebook. It even runs an Audio Overview while you explore other views, trimming toolβswitching friction.
π‘ Tencent Hunyuan just released Hunyuan3D World Model 1.0!
Tencent just released Hunyuan3D World Model 1.0, which theyβre calling the first open-source AI system that can build full 3D worlds from just a text prompt or an image. Basically, itβs like a 3D version of Soraβbut one you can walk around in, edit, and even drop into a game engine.
What used to take weeks to build can now be generated in just a few minutes. Itβs up on GitHub and Hugging Face, and thereβs a live demo you can play with right now. But thereβs a caveat: the license isn't fully open-source. Itβs more of a βsource-availableβ setup, with some heavy restrictions, especially around where and how it can be used.
This is set to transform game development, VR, digital content creation and so on. Get started nowππ»
360Β° immersive experiences via panoramic world proxies;
mesh export capabilities for seamless compatibility with existing computer graphics pipelines;
disentangled object representations for augmented interactivity.
The core of our framework is a semantically layered 3D mesh representation that leverages panoramic images as 360Β° world proxies for semantic-aware world decomposition and reconstruction, enabling the generation of diverse 3D worlds.
Achieves SOTA performance in generating coherent, explorable, and interactive 3D worlds while enabling versatile applications in virtual reality, physical simulation, game development, and interactive content creation.
Tencent HunyuanWorld-1.0's generation architecture integrates panoramic proxy generation, semantic layering, and hierarchical 3D reconstruction to achieve high-quality scene-scale 360Β° 3D world generation, supporting both text and image inputs.
π΅ Chinaβs open-source AI engine is still running hot, Alibabaβs Wan2.2 video model launched.
π΅ Chinaβs open-source AI engine is still running hot.
Alibabaβs Tongyi Lab just dropped Wan2.2, an open-source video model built for sharp motion and film-like quality in both text-to-video and image-to-video. Licensed under the Apache 2.0
You get user-controlled lighting in a single 4090-sized GPU. Visually equal result to Seedance, Kling, Hailuo, and Sora on looks and motion.
Wan2.2 spreads its load across two expert networks: one sketches each frame, the other polishes fine detail.
This mixture-of-experts keeps compute steady yet lifts capacity, so scenes stay coherent even through rapid camera swings.
The lab fed the model 66% more images and 83% more videos than Wan2.1. That extra variety helps it read text inside frames, follow tricky shots, and keep characters moving naturally.
A new Wan2.2-VAE squeezes footage by 16x16x4, so a 5B parameter system streams 720p 24fps on a single RTX4090. Users can nudge exposure, color, and shot length without retraining, giving near-director-level control.
Chinaβs open models now span speech, language, and video. Wan2.2 shows how more data plus smart routing can close the gap with closed labs while staying fully shareable.
ποΈ Byte-Size Briefs
As per widely circulating rumor, OpenAI is dropping GPT-5 next week!
β1M token input window, 100k output tokens
βMCP support, parallel tool calls
βDynamic short + long reasoning
βUses Code Interpreter and other tools
Codenames are:
o3-alpha > nectarine (GPT-5) > lobster (mini) > starfish (nano)
Somebody shared the full Claude code system prompt on GitHub. its super long at 144K characters / 16.9K+ words. (note, this is Claude Code, not the Claude webui prompt which is open-sourced anyway)
π§βπ OPINION: Six Strange Aliases, One Giant Model: The GPTβ5 Scavenger Hunt
π Snapshot
OpenAI has been spotted fieldβtesting 6 anonymous models on the coding site LMβ―Arena. Community sleuths link those codenames, Zenith, Summit, Lobster, Starfish, Nectarine, and o3βalpha, to different slices of the forthcoming GPTβ5. Each slice shows a standout skill, from giant code dumps to slick SVG animation.
The pattern fits a Mixtureβ―ofβ―Experts blueprint, where specialized βexpertsβ are tuned in public, then stitched together for the final release. The Verge and Windows Central both caught Microsoft wiring Copilot for GPTβ5βs arrival in Augustβ―2025, and leaked Copilot code already references a βSmartβ mode that flips between quick and deep reasoning.
π¨ The mystery lineβup
Right now LMβ―Arenaβs leaderboard suddenly features Zenith, Summit, Lobster, Starfish, Nectarine, and o3βalpha, and they sit above nearly every public model on tough reasoning and coding tasks. A Reddit thread catalogued the names the moment they appeared, noting how Zenith consistently tops the chart while o3βalpha wins coding headβtoβhead challenges.
Somebody pushed Summit with a playful prompt and got a fully interactive starship control panel, more than 2,300β―lines of JavaScript, that ran without tweaks on the first try. He posted the raw code and a live demo, which rocketed through social feeds as proof of a new coding ceiling.
π§© Why the pieces look like MoE specialists
Its a simple theory: each Arena model is an βexpertβ trained for one domain, such as longβform writing (Zenith) or dense code refactoring (o3βalpha). That lines up with mixtureβofβexperts research, where a router sends each token to the few specialists that matter, slashing compute while boosting capacity. Huggingβ―Faceβs explainer and IBMβs primer both show how MoE layers keep only a couple of experts hot per request, letting a giant network act small at inference time. OpenAI can therefore expose individual experts, harvest realβworld feedback, then glue them together for the full GPTβ5 launch.
π©βπ» Coding power is the headline feature
Many report say GPTβ5 is built βto crush coding tasks,β from refactoring monolithic codebases to automating browser work. Public Arena runs back that claim: Summit writes productionβgrade React when asked, Zenith drafts clean TypeScript scaffolds, and o3βalpha debugs obscure stack traces without breaking a sweat. The Information piece adds that GPTβ5 lets callers set how long the model βthinksβ before answering, much like a dimmer control for reasoning depth.
On Microsoftβs side, insiders found Copilotβs UI now sporting a hidden βSmartβ toggle that explicitly calls GPTβ5 to βthink quickly or deeply,β showing Redmond is prepping its flagship assistant for the handβoff.
π Why this public test matters
Running the experts in the open hands OpenAI a mountain of messy, realβworld edge cases it cannot simulate inβhouse. Every strange prompt, every unexpected coding bug, and every refusal fallout lands as fresh gradient data for the final checkpoints. For users, the experiment hints at a new release style: instead of a surprise monolith drop, we may see modules pop up, prove themselves, and then fuse into one generalist megaβmodel. If GPTβ5 emerges with a modular backbone, future upgrades could arrive expertβbyβexpert, cutting retraining bills and letting OpenAI bolt on new skills on the fly.
π Takeaway
The leaderboard stunt shows OpenAI taking a page from game studios that ship public betas: let the crowd break each part early, fix the cracks, then ship the polished whole. For builders, the message is clear.
Expect bigger context windows, richer code support, and a model that can throttle thinking depth per call. For competitors, the clock is ticking, the GPTβ5 assembly line is already humming.
Thatβs a wrap for today, see you all tomorrow.