๐ DeepSeek-R1 fails every safety test thrown at it and failed to block any harmful request
DeepSeek-R1 fails every safety test, ByteDance's OmniHuman-1 stuns with lifelike videos, DeepMind proves RL beats fine-tuning, OpenAI eyes smartphone replacement, and AI outsmarts PhDs.
The above podcast on this post today was generated with Googleโs Illuminate.
Read time: 7 min 27 seconds
๐ Browse past editions here.
( I write daily for my 112K+ AI-pro audience, with 4.5M+ weekly views. Noise-free, actionable, applied-AI developments only).
โกIn todayโs Edition (4-Feb-2025):
๐ DeepSeek-R1 fails every safety test thrown at it and failed to block any harmful request
๐จ ByteDance released OmniHuman-1, itโs insanely good model for creating lifelike human videos based on a single human image
๐ฅ Google DeepMindโs new paper explains why reinforcement learning outperforms supervised fine-tuning for model generalization.
๐๏ธ Byte-Size Briefs:
OpenAI is reportedly developing an AI device to replace smartphones, according to CEO Sam Altman.
xAIโs Grok Android app enters beta testing in Australia, Canada, India, the Philippines, and Saudi Arabia. Users can pre-register on Google Play.
AI now surpasses PhD experts in their own fields, according to viral posts citing data from epoch.ai.
Developers deploy malware-based tarpits like Nepenthes and Iocaine to trap AI web crawlers, waste resources, and mislead models, increasing AI firmsโ data collection costs.
๐งโ๐ Top Github Repo: Oumi:build state-of-the-art foundation models, end-to-end
๐ Educational Resource: A Little Bit of Reinforcement Learning from Human Feedback
๐ DeepSeek-R1 is failing every safety test thrown at it, 100% attack success rate
๐ฏ The Brief
DeepSeek R1, has been found extremely vulnerable to security attacks. A study by Ciscoโs Robust Intelligence and the University of Pennsylvania tested the model with 50 harmful prompts from the HarmBench dataset, covering areas like cybercrime and misinformation. Result: 100% attack success rateโit failed to block a single harmful request.
โ๏ธ The Details
โ Researchers from Cisco applied algorithmic jailbreaking using an automated attack methodology. The 100% attack success rate means DeepSeek R1 failed to block any of the 50 adversarial prompts.
โ Security firm Adversa AI confirmed the results, finding that DeepSeek R1 is susceptible to both basic linguistic tricks and advanced AI-generated exploits.
๐จ ByteDance released OmniHuman-1, itโs insanely good model for creating lifelike human videos based on a single human image
๐ฏ The Brief
ByteDance's OmniHuman-1 is a Diffusion Transformer-based AI model that generates highly realistic human videos from a single image and audio/video input. Unlike previous models limited to facial or static full-body animation, OmniHuman-1 handles diverse aspect ratios, complex human-object interactions, and pose-driven motion. It outperforms existing methods in realism, achieving state-of-the-art lip-sync accuracy, gesture precision, and expressiveness.
โ๏ธ The Details
โ OmniHuman-1 extends traditional human animation by using audio, pose, and reference images in a unified training framework, improving realism across portraits, half-body, and full-body animations.
โ The model is built on DiT (Diffusion Transformer) architecture, leveraging multimodal motion conditioning to enhance training efficiency and video quality.
โ Supports video-driven animation, allowing motion replication from existing videos, or a combination of audio and video to control body parts independently.
โ Outperforms competitors (e.g., Loopy, CyberHost, DiffTED) in lip-sync accuracy (5.255 vs. 4.814), FVD (15.906 vs. 16.134), and gesture expressiveness, while supporting varied body proportions in a single model.
โ Handles diverse styles, including cartoons, stylized characters, and anthropomorphic objects, preserving unique motion characteristics.
๐ฅ Google DeepMindโs new paper explains why reinforcement learning outperforms supervised fine-tuning for model generalization
๐ฏ The Brief
Google DeepMindโs latest study demonstrates that reinforcement learning (RL) significantly outperforms supervised fine-tuning (SFT) for model generalization in both textual and visual domains. Using two benchmark tasks, GeneralPoints (arithmetic reasoning) and V-IRL (real-world navigation), the researchers show that SFT tends to memorize training data, while RL enables models to adapt to novel, unseen scenarios. The findings reinforce RLโs effectiveness in learning generalizable knowledge for complex multimodal tasks.
โ๏ธ The Details
โ The study compares SFT and RL in foundation model post-training and their ability to generalize beyond training data. RL is shown to be significantly better at rule-based and visual generalization, whereas SFT exhibits strong memorization tendencies.
โ In GeneralPoints (text-based arithmetic reasoning), RL enables models to compute unseen rule variations, whereas SFT-trained models struggle with novel rule applications.
โ In V-IRL (real-world navigation), RL improves out-of-distribution generalization by +33.8% (44.0% โ 77.8%), demonstrating better adaptation to new navigation conditions.
โ RL enhances visual recognition capabilities, a crucial factor in vision-language models (VLMs), while SFT degrades visual recognition performance as it overfits to reasoning tokens.
โ Despite RLโs advantages, SFT remains useful for stabilizing the modelโs output format, which helps RL achieve better learning efficiency.
โ The study suggests that scaling up inference-time computation (increasing verification steps) further boosts RL generalization.
โ Without SFT pre-training, RL training fails due to poor instruction-following ability, emphasizing the necessity of a structured initialization before RL fine-tuning.
๐๏ธ Byte-Size Briefs
OpenAI is reportedly developing a new AI device intended to replace smartphones, as stated by CEO Sam Altman.
The Grok Android app, developed by xAI, has been released for beta testing in Australia, Canada, India, the Philippines, and Saudi Arabia.
A viral post in Reddit and Twitter shows AIโs Exponential progress, which now surpasses human PhD experts in their own field. The data for this comes from epoch.ai
Developers are deploying malware-based tarpits like Nepenthes and Iocaine to trap and poison AI web crawlers that ignore robots.txt. These tools waste resources, mislead AI models with gibberish, and inflate AI companies' data collection costs. While AI firms like OpenAI have countermeasures, the rise of AI poisoning tactics signals growing resistance against unchecked AI scraping.
๐งโ๐ Top Github Repo
Oumi:build state-of-the-art foundation models, end-to-end
Oumi is a fully open-source platform designed to train, evaluate, and deploy foundation models end-to-end. It supports models from 10M to 405B parameters, enabling fine-tuning using LoRA, QLoRA, DPO, and other techniques. It integrates with popular inference engines (vLLM, SGLang) and works across laptops, clusters, and cloud platforms (AWS, Azure, GCP, Lambda, etc.). Supports multimodal models like Llama, DeepSeek, and Phi.
โ๏ธ Key Benefits
โ Oumi simplifies model training with a unified API, allowing seamless model fine-tuning, data synthesis, and evaluation. Supports both open-source and commercial APIs like OpenAI, Anthropic, and Vertex AI, making it highly flexible.
โ Enables fast inference with optimized engines such as vLLM and SGLang, ensuring efficient deployment. Installation is straightforward with pip install oumi
, supporting both CPU and GPU setups.
โ Supports cloud-based training with direct job execution on AWS, Azure, GCP, and Lambda. Includes prebuilt ready-to-use training recipes for LLM fine-tuning, distillation, evaluation, and inference.
โ 100% open-source under Apache 2.0 license, with an active community on Discord and GitHub.
๐ Educational Resource: A Little Bit of Reinforcement Learning from Human Feedback
This work systematically breaks down RLHFโs fundamental algorithms and optimizations for training more stable and efficient LLMs.
A concise overview of Reinforcement Learning from Human Feedback (RLHF) with a focus on policy gradient algorithms like PPO, REINFORCE, and GRPO. These algorithms optimize LLMs by directly updating policies based on rewards instead of storing them in replay buffers. The work explores new RLHF variants such as REINFORCE Leave One Out (RLOO) and Group Relative Policy Optimization (GRPO), which improve stability and efficiency for LLMs.
โ๏ธ Key Learning Takeaways
โ Policy gradient methods are core to RLHF, optimizing models by computing gradients based on expected returns from current policies. These include REINFORCE, PPO, and GRPO, which control policy updates differently.
โ Vanilla Policy Gradient methods suffer from high variance, mitigated using advantage functions that normalize rewards. REINFORCE is a Monte Carlo-based gradient estimator, operating without a value function.
โ REINFORCE Leave One Out (RLOO) improves stability by using batch-level reward baselines instead of full averages. This works well in domains where multiple completions per prompt are used.
โ Proximal Policy Optimization (PPO), a widely used RL method, controls updates via clipping to prevent excessive policy shifts. PPO requires a value network to estimate advantages.
โ Group Relative Policy Optimization (GRPO), introduced in DeepSeekMath, simplifies PPO by removing the need for a value function. This reduces memory consumption and avoids challenges in learning value functions for LLMs.
โ Implementation details cover practical aspects of log-probability ratios, KL penalties, and clipping constraints, which refine policy optimization.
โ Double regularization in RLHF ensures stable learning by balancing updates with KL divergence constraints, preventing drastic policy shifts.
Thatโs a wrap for today, see you all tomorrow.