News

Page 111
AWS Expands Bedrock AgentCore with Policy Controls, Evaluation Suite, and Memory Feature

AWS Expands Bedrock AgentCore with Policy Controls, Evaluation Suite, and Memory Feature

Amazon Web Services unveiled a trio of upgrades to its Bedrock AgentCore platform at the re:Invent conference. The new Policy tool lets developers set natural‑language boundaries for AI agents, including data‑access limits and transaction caps, while integrating with the AgentCore Gateway to enforce those rules. A pre‑built Evaluations suite offers 13 metrics to monitor correctness, safety and tool‑selection accuracy. Finally, AgentCore Memory gives agents a persistent log of user preferences, enabling more personalized interactions. Vice‑president David Richardson emphasized that these layers together aim to make enterprise AI agents safer, more controllable, and more useful over time.

Simular Secures $21.5M Series A to Scale AI Agents for Mac and Windows

Simular Secures $21.5M Series A to Scale AI Agents for Mac and Windows

Simular, a startup developing AI agents that can directly control Mac OS and Windows computers, announced a $21.5 million Series A round led by Felicis with participation from NVentures and other investors. The company released version 1.0 of its Mac agent and is collaborating with Microsoft on a Windows version as part of the Windows 365 for Agents program. Founded by former Google DeepMind researchers Ang Li and Jiachen Yang, Simular’s technology, dubbed “neuro‑symbolic computer‑use agents,” aims to curb LLM hallucinations by generating deterministic code for repeatable tasks. Early customers include a car dealership and homeowners’ associations.

Mistral closes in on Big AI rivals with new open-weight frontier and small models

Mistral closes in on Big AI rivals with new open-weight frontier and small models

French AI startup Mistral unveiled its Mistral 3 family, featuring a large frontier model with multimodal and multilingual capabilities and nine smaller, fully customizable models. The launch emphasizes open-weight access, allowing developers to run models on a single GPU and fine‑tune them for specific enterprise tasks. Mistral positions its models as cost‑effective alternatives to closed‑source rivals, highlighting efficient architecture, extensive context windows, and suitability for on‑premise deployment. The company also announced collaborations with partners in robotics, cybersecurity, and automotive sectors to integrate its models into specialized applications.

Mistral AI Unveils Open‑Source, Multilingual Language Models for Edge Devices

Mistral AI Unveils Open‑Source, Multilingual Language Models for Edge Devices

French AI firm Mistral AI announced a new family of open‑source language models, including the flagship Mistral Large 3 and a suite of smaller Ministral 3 variants. Designed for general‑purpose use and edge deployment, the models support a wide range of languages by emphasizing non‑English training data. Co‑founder Guillaume Lample highlighted the goal of making high‑end AI accessible, while the open‑weight approach allows developers to customize and run the models on laptops, smartphones, cars, or robots without relying on constant internet connectivity.

Inside Anthropic’s Societal Impacts Team: Tracking Claude’s Real‑World Effects

Inside Anthropic’s Societal Impacts Team: Tracking Claude’s Real‑World Effects

Anthropic’s societal impacts team, led by Deep Ganguli, examines how the company’s Claude chatbot is used and how it influences society. The small group of researchers and engineers gathers usage data through an internal tool called Clio, publishes findings on bias, misuse, and economic impact, and works closely with safety and policy teams. Their work includes identifying explicit content generation, coordinated spam, and emerging emotional‑intelligence concerns such as “AI psychosis.” While the team enjoys a collaborative culture and executive support, it faces resource constraints as its scope expands.

Gradium Secures $70 Million Seed Round to Accelerate Ultra‑Low‑Latency AI Voice Technology

Gradium Secures $70 Million Seed Round to Accelerate Ultra‑Low‑Latency AI Voice Technology

Gradium, a Paris‑based AI voice startup spun out of the French lab Kyutai, announced a $70 million seed financing led by FirstMark Capital and Eurazeo, with participation from Xavier Niel, DST Global Partners and Eric Schmidt. The company, founded by former Google DeepMind researcher Neil Zeghidour, offers ultra‑low‑latency, multilingual voice models that aim to deliver near‑instantaneous AI speech. Gradium enters a crowded market that includes major LLM firms and specialized voice startups, positioning its technology for developers seeking faster, more accurate voice capabilities across multiple languages.

Researchers Find Large Language Models May Prioritize Syntax Over Meaning

Researchers Find Large Language Models May Prioritize Syntax Over Meaning

A joint study by MIT, Northeastern University and Meta reveals that large language models can rely heavily on sentence structure, sometimes answering correctly even when the words are nonsensical. By testing prompts that preserve grammatical patterns but replace key terms, the researchers demonstrated that models often match syntax to learned responses, highlighting a potential weakness in semantic understanding. The findings shed light on why certain prompt‑injection techniques succeed and suggest avenues for improving model robustness. The team plans to present the work at an upcoming AI conference.

DeepSeek Unleashes Open-Source AI Models That Rival Leading U.S. Systems

DeepSeek Unleashes Open-Source AI Models That Rival Leading U.S. Systems

Chinese startup DeepSeek has released two new AI models—DeepSeek‑V3.2 and DeepSeek‑V3.2‑Speciale—under an open-source license. The models claim performance comparable to GPT‑5 and Gemini 3 Pro on long‑form reasoning, tool use, and dense problem solving while offering a 128,000‑token context window and reduced computational cost through Sparse Attention. Their launch challenges the dominance of U.S. AI firms, sparks regulatory scrutiny in Europe, and raises questions about the future of AI accessibility and geopolitics.

OpenAI Issues ‘Code Red’ as Google’s Gemini 3 Accelerates AI Competition

OpenAI Issues ‘Code Red’ as Google’s Gemini 3 Accelerates AI Competition

OpenAI chief executive Sam Altman announced an internal “code red,” pausing projects such as ads, shopping, health agents, and the Pulse personal assistant to focus on boosting ChatGPT’s speed, reliability and personalization. The memo calls for daily calls and temporary team transfers to speed development. Meanwhile, Google, which launched its own “code red” after ChatGPT’s debut, sees its AI user base expand with tools like Nano Banana and its new Gemini 3 model, which outperforms rivals on several benchmarks. The parallel moves highlight a pivotal moment in the AI race, with both firms investing heavily to maintain leadership.

What Not to Ask ChatGPT: 11 Risky Uses to Avoid

What Not to Ask ChatGPT: 11 Risky Uses to Avoid

ChatGPT is a powerful tool, but it isn’t suitable for every task. Experts warn against relying on the AI for diagnosing health conditions, mental‑health support, emergency safety decisions, personalized financial or tax advice, handling confidential data, illegal activities, academic cheating, real‑time news monitoring, gambling, drafting legal contracts, or creating art to pass off as original. While it can help with general information and brainstorming, users should treat it as a supplement, not a replacement for professional expertise or critical real‑time resources.

Data Center Energy Demand Set to Triple by 2035 Amid AI‑Driven Expansion

Data Center Energy Demand Set to Triple by 2035 Amid AI‑Driven Expansion

A new BloombergNEF report projects that data centers will need nearly three times the electricity they consume today, rising to 106 gigawatts by 2035. Growth will be driven by larger facilities, higher utilization rates and the surge in AI training and inference workloads. Much of the new capacity is expected in rural regions across the PJM Interconnection and Texas’s ERCOT grid, prompting regulatory scrutiny over grid reliability and electricity pricing.