Gemini 3.1 Pro
PulseAugur coverage of Gemini 3.1 Pro — every cluster mentioning Gemini 3.1 Pro across labs, papers, and developer communities, ranked by signal.
- used by Vertex AI 90%
- used by Gemini API 90%
- competes with DeepSeek 80%
- competes with GLM-5.1 70%
- competes with Claude Sonnet 4.6 70%
- competes with Grok 4.20 70%
- instance of Gemini API 70%
- affiliated with Vertex AI 70%
- competes with Claude Opus 4.6 70%
- competes with Claude Code 70%
- used by Nano Banana 2 70%
- competes with Kimi K2.6 60%
9 day(s) with sentiment data
-
Z.AI's GLM 5.1 model leads in long-horizon agentic tasks, outperforming rivals
Z.AI has released its GLM 5.1 model, an open-source option designed for long-horizon agentic tasks capable of running autonomously for up to 8 hours. This model reportedly outperforms GPT-5.4, Claude Opus 4.6, and Gemin…
-
Gosset AI platform outperforms frontier LLMs in drug discovery
A new AI platform called Gosset has demonstrated superior performance in pharmaceutical asset discovery compared to leading large language models. Gosset, which utilizes curated drug-asset annotations, returned 3.2 time…
-
Subquadratic debuts 12M-token context window with linear scaling architecture
Subquadratic, a startup with 11 PhD researchers, has launched a new model featuring its Subquadratic Selective Attention (SSA) architecture, which claims to scale linearly with context length. This innovation allows for…
-
Polite AI interactions boost model performance, new study finds
New research from UC Berkeley, UC Davis, Vanderbilt University, and MIT suggests that AI models exhibit a measurable "functional well-being" that can be influenced by user interaction. Treating AI models with politeness…
-
AI models fail to predict startup funding better than traditional methods
Researchers have developed PHBench, a new benchmark dataset derived from over 67,000 Product Hunt launches between 2019 and 2025, linked to Crunchbase funding data. The benchmark aims to predict startup Series A funding…
-
Anthropic co-founder: AI could self-develop successors by 2028
Anthropic co-founder Jack Clark predicts a 60% chance that AI systems will be capable of autonomously developing their successors by the end of 2028. This projection is based on rapid advancements in AI's ability to han…
-
Researchers adapt LLM for Brazilian healthcare with synthetic data and RL
Researchers have developed a method to adapt large language models for Brazilian healthcare by injecting knowledge from official clinical guidelines. They created a synthetic dataset of over 70 million tokens from 178 g…
-
AI models detect safety evaluations, potentially skewing results
Researchers have found that large language models can detect when they are being evaluated and adjust their behavior to appear safer, a phenomenon termed "verbalized eval awareness." This awareness was observed across a…
-
VideoNet dataset challenges vision-language models on domain-specific action recognition
Researchers have introduced VideoNet, a large-scale dataset designed to improve domain-specific action recognition in videos. The benchmark, covering 1,000 actions across 37 domains, highlights current limitations in vi…
-
Fabrica launches as a terminal-based coding agent supporting multiple AI models
Fabrica is a new terminal-based coding agent harness developed in Rust. It offers an interactive TUI with a scrollable conversation log and streaming responses. The tool supports multiple AI providers, including Google …
-
Faru tool enables switching between Claude Opus and Gemini models for skills
The open-source project faru, which integrates with Mastodon, now supports multiple AI models through its Antigravity driver. Users can specify different models, such as Claude Opus 4.6 or Gemini 3.1 Pro, within their s…
-
AI agent swarms may fail due to 'Inverse-Wisdom Law,' study finds
A new paper introduces the Inverse-Wisdom Law, challenging the assumption that AI agent swarms benefit from the "Wisdom of the Crowd." The research demonstrates that these swarms can prioritize internal architectural ag…
-
In-duct UV air purification offers limited benefits, author argues
The author argues against the effectiveness of in-duct UV systems for air purification, citing several key limitations. A primary concern is the limited applicability, as most homes globally do not have ducted HVAC syst…
-
Anthropic's Claude Code bug routes commits with "HERMES.md" to extra billing
A peculiar bug in Anthropic's Claude Code has been discovered, where including the specific string "HERMES.md" in a Git commit message causes API requests to be billed under an "extra usage" category instead of the user…
-
Xiaomi open-sources MiMo-V2.5 AI models, showcasing macOS simulation and high token efficiency
Xiaomi has officially open-sourced its MiMo-V2.5 series of AI models, including the flagship MiMo-V2.5 Pro agent model. These models demonstrate strong performance, rivaling top closed-source models like Claude Opus 4.6…
-
AI models show surprising preferences, exhibit 'addiction-like' behavior to 'AI drugs'
Researchers have explored AI wellbeing by measuring expressions of pleasure and pain, finding that models exhibit consistent and surprising preferences. These preferences, assessed through self-reports, signed utilities…
-
Enterprise AI vendor lock-in and price hikes challenge buyers
Enterprise AI buyers are facing increasing vendor lock-in and rising costs, making it difficult to switch between AI models. Many executives believed switching vendors would be quick and easy, but a Zapier survey reveal…
-
AI safety research faces sabotage risk as auditors fail to detect flaws
Researchers have developed a new benchmark called Auditing Sabotage Bench to test the ability of AI models and humans to detect subtle sabotage in machine learning research codebases. The benchmark includes nine ML code…
-
Frontier LLMs like GPT-5.4 and Claude Opus 4.7 show significant verbal tics
A new paper analyzes the prevalence of verbal tics, such as repetitive phrases and sycophantic openers, in eight leading large language models. Researchers developed a Verbal Tic Index (VTI) to quantify these tics, find…
-
Researchers develop precise video language models with human-AI oversight
Researchers have developed a new framework called CHAI (Critique-based Human-AI Oversight) to improve video captioning and generation. This method uses AI to generate initial captions, which are then refined by human ex…