GPT-5 Excels in Medical Reasoning; Perplexity AI Bids for Chrome to Challenge Google
Key Takeaways
- Perplexity AI offered Google $34.5B for Chrome to integrate AI; Google declined.
- GPT-5 outperforms experts in multimodal medical reasoning, achieving state-of-the-art accuracy.
- AI companion apps projected to generate $120M in 2025 with 220M downloads.
- GitHub Copilot Pro supports GPT-5 in VS Code, offering a 30-day free trial.
- Context engineering reduces AI errors by 80% and saves 75%-99% across industries.
Top Stories
Perplexity AI bids $34.5B for Google Chrome acquisition.
On August 12, 2025, Perplexity AI made an unsolicited $34.5 billion cash offer to Google to acquire Chrome, aiming to challenge Google's search dominance by integrating AI into the browser. Google has no plans to sell Chrome.
GPT-5 surpasses experts in multimodal medical reasoning, study shows.
A study submitted to arXiv on August 11, 2025, evaluates GPT-5's performance in multimodal medical reasoning, indicating that GPT-5 surpasses all baselines and pre-licensed human experts, achieving state-of-the-art accuracy.
AI companion apps revenue projected at $120M in 2025.
AI companion apps are projected to generate $120 million in revenue in 2025, with 220 million downloads globally and $82 million in revenue generated during the first half of the year.
GitHub Copilot Pro supports GPT-5 in VS Code with free trial.
GitHub Copilot Pro now supports GPT-5 in VS Code, offering users a 30-day free trial.
Context engineering transforms AI, reduces errors 80%, saves 75%-99%.
Context engineering is transforming AI applications across industries, with Five Sigma Insurance seeing an 80% reduction in claim processing errors and Microsoft seeing a 26% increase in completed software tasks.
AI Breakthroughs
MCPToolBench++: Large-scale AI Agent tool use benchmark introduced.
Researchers proposed MCPToolBench++, a large-scale AI Agent tool use benchmark built upon a marketplace of over 4k MCP servers from more than 40 categories, consisting of single-step and multi-step tool calls.
Mem4D: Novel framework for dynamic scene reconstruction introduced.
Submitted on August 11, 2025, Mem4D is a novel framework for dynamic scene reconstruction that decouples static and dynamic memory using a dual-memory architecture.
DMPO: Data selection aligns LLMs with human values, improves 10%.
Researchers propose Direct Multi-Preference Optimization (DMPO), a data selection principle for aligning Large Language Models (LLMs) with diverse human values, achieving over 10% relative improvement.
Griffon v2: Unified model excels in multimodal perception, object referring.
Researchers introduce Griffon v2, a unified high-resolution generalist model for multimodal perception, enabling flexible object referring with visual and textual prompts, achieving state-of-the-art performance on REC and phrase grounding.
X-evolve: LLM-powered method evolves solution spaces efficiently.
Researchers introduce X-evolve, a method that evolves solution spaces powered by large language models, generating tunable programs and efficiently exploring the space with a score-based search algorithm.
PBD5K: New task, benchmark for power battery detection via X-rays.
Researchers propose a new task called Power Battery Detection (PBD) and release PBD5K, a large-scale benchmark with 5,000 X-ray images for quality inspection.
Grove MoE: Novel MoE architecture expands capacity, manages overhead.
Researchers introduce Grove MoE, a novel MoE architecture with adjugate experts, enabling model capacity expansion while maintaining manageable computational overhead, achieving performance comparable to SOTA open-source models.
InterChart: Benchmark evaluates vision-language models' reasoning across charts.
Researchers introduce InterChart, a diagnostic benchmark evaluating vision-language models' reasoning across multiple related charts, revealing accuracy declines as chart complexity increases.
Deepfake detection papers achieve top performance in ACM challenge.
On August 10-11, 2025, researchers submitted papers addressing audio-visual deepfake detection, with one achieving top performance in temporal localization and a top-four ranking in classification for the ACM 1M Deepfakes Detection Challenge.
Method identifies bias in GPT-2 using topological data analysis.
Researchers propose a method using topological data analysis to identify which parts of GPT-2 contribute to bias towards specific groups, finding biases concentrated in certain attention heads.
SE-LLM enhances LLM interpretability for time series analysis.
Researchers propose a novel Semantic-Enhanced LLM (SE-LLM) that enhances token embedding by exploring inherent periodicity and anomalous characteristics of time series, improving interpretability for LLMs in temporal sequence analysis.
Speculative decoding for Llama models achieves 4ms latency on H100s.
Researchers submit a paper detailing techniques for efficient speculative decoding for Llama models at production scale, achieving a new state-of-the-art inference latency of about 4 ms per token on 8 NVIDIA H100 GPUs.
Fairness enforcement techniques for generative AI introduced.
On August 11, 2025, arXiv announced the v5 revision of a paper introducing characterization and enforcement techniques to address fairness concerns in generative AI (GenAI).
REX-RAG enhances LLM reasoning via reinforcement learning, improves 5.1%.
Wentao Jiang and colleagues introduced REX-RAG, a framework enhancing LLM reasoning via Reinforcement Learning with Retrieval-Augmented Generation (RAG), improving performance by 5.1% on Qwen2.5-3B and 3.6% on Qwen2.5-7B over baselines.
Transformers become symmetry-aware for automated planning via contrastive learning.
Researchers propose a novel contrastive learning objective to make transformers symmetry-aware for automated planning, addressing the limitations of PlanGPT.
In-context info impacts LLM reliability; misleading context induces errors.
Researchers investigate how in-context information influences model behavior and whether LLMs can identify their unreliable responses, finding that misleading context often induces confidently incorrect responses.
MLLMs learn user preferences for personalized image generation.
Researchers propose a new approach for learning user preferences in image generation models using Multimodal Large Language Models, outperforming others in preference prediction accuracy.
OpT-DeUS for LLMs improves performance and training efficiency.
Researchers propose Optimal Transport Depth Up-scaling (OpT-DeUS) for Large Language Models (LLMs), achieving better performance and training efficiency than existing methods.
Study explores user privacy perceptions of LLM's RAG-based memory.
A paper submitted on August 11, 2025, explores user privacy perceptions towards LLM's Retrieval Augmented Generation (RAG)-based memory, revealing diverse mental models and significant concerns regarding privacy, control, and accuracy.
Dataset simulates proactive robots inferring human needs from conversations.
Published on August 11, 2025, a paper introduces a simulated dataset designed to support research on proactive robots that infer human needs from natural language conversations within workplace environments.
Efficient posterior sampling with Annealed Langevin Monte Carlo.
On August 11, 2025, a paper was submitted addressing posterior sampling in score-based generative models, introducing a method to sample from a distribution close to both the noised prior posterior and the true posterior.
CSVG solves zero-shot 3D visual grounding as constraint satisfaction.
On August 11, 2025, arXiv announced the revised version of a paper introducing Constraint Satisfaction Visual Grounding (CSVG), a zero-shot method that reformulates 3D visual grounding (3DVG) as a Constraint Satisfaction Problem (CSP).
Industry Watch
Tahoe Therapeutics raises $30M for AI models of living cells.
Tahoe Therapeutics, a Palo Alto biotech startup, raised $30 million in funding led by Amplify Partners to develop AI models of living cells for cancer research, bringing total funding to $42 million and a valuation of $120 million.
Chinese tech companies actively recruiting AI experts amid rising demand.
As reported by Digitimes on August 12, 2025, Chinese tech companies are actively recruiting AI experts, mirroring a broader trend of young people learning AI skills to stay competitive in the job market.
Study: 75% comfortable with AI agents, 30% want AI managers.
A Workday study from August 12, 2025, found that 75% of employees are comfortable working with AI agents, but only 30% want AI as their manager, with 82% of organizations expanding AI agent use.
AI agents offer better ROI than RAG for businesses, expert argues.
Alon Goren, writing for Forbes Technology Council, argues that AI agents, which possess autonomy and can make independent decisions, are a better option for achieving transformative enterprise value than Retrieval Augmented Generation (RAG).
Real-World AI
FEAT: AI system automates cause-of-death analysis with LLM.
Researchers introduced FEAT, a multi-agent AI framework for automated cause-of-death analysis, which uses a domain-adapted large language model to standardize death investigations and outperformed state-of-the-art AI systems.
Improved plant segmentation model deployed in BASF's phenotyping pipeline.
Researchers developed an improved segmentation model for automated multi-species and damage plant semantic segmentation in herbicide trials, which significantly improved species identification and damage classification and is now deployed in BASF's phenotyping pipeline.
AIS-LLM: Integrates time-series AIS data with LLM for maritime analysis.
AIS-LLM, a novel framework, integrates time-series AIS data with a large language model (LLM) for maritime traffic analysis, enabling simultaneous execution of trajectory prediction, anomaly detection, and collision risk assessment.
Vision, LLM navigation system achieves 96% accuracy indoors.
Researchers from Cornell University present an approach integrating vision-based localization with large language model (LLM)-based navigation for indoor environments, achieving 96% accuracy across waypoints in an office corridor.
NeuroDx-LM: Model for EEG-based neurological disorder detection.
A new paper, NeuroDx-LM, proposes a large-scale model for EEG-based neurological disorder detection, showing state-of-the-art performance in seizure and schizophrenia detection.
CURec enhances recommender systems using LLM fine-tuning.
On August 11, 2025, Yunze Luo et al. submitted a paper proposing CURec, a framework for enhancing recommender systems using large language model (LLM) fine-tuning by generating collaborative-aligned content features.