News

Page 108
OpenAI Acknowledges Ongoing Prompt Injection Risk in Atlas Browser

OpenAI Acknowledges Ongoing Prompt Injection Risk in Atlas Browser

OpenAI has publicly recognized that prompt injection attacks remain a persistent threat to its Atlas AI browser. The company says the risk is unlikely to be fully eliminated and is investing in continuous defenses, including a reinforcement‑learning‑based automated attacker that simulates malicious inputs. OpenAI’s updates aim to detect and flag suspicious prompts, while it also advises users to limit agent autonomy and access. The UK National Cyber Security Centre echoed the concern, noting that prompt‑injection attacks may never be completely mitigated. Other AI firms such as Anthropic and Google are taking similar defensive approaches.

Google Gemini 3 Flash Shows High Hallucination Rate Despite Leading Performance

Google Gemini 3 Flash Shows High Hallucination Rate Despite Leading Performance

Google's Gemini 3 Flash model, praised for speed and accuracy, exhibits a striking 91% hallucination rate in tests where it should admit uncertainty. While the model remains top‑scoring in general AI benchmarks, its tendency to fabricate answers when it lacks knowledge raises concerns about reliability, especially as the technology integrates into consumer products like Google Search. Experts highlight the need for better uncertainty detection and caution users to verify AI‑generated information.

OpenAI Reports Surge in Child Exploitation Reports to NCMEC

OpenAI Reports Surge in Child Exploitation Reports to NCMEC

OpenAI disclosed that it submitted roughly 75,000 reports to the National Center for Missing & Exploited Children during the first half of the year, a dramatic increase from the under 1,000 reports it filed in the same period the previous year. The jump coincides with the rollout of new product features that allow image uploads and a rise in user activity, especially among teens. OpenAI said the growth reflects expanded reporting capacity and ongoing safety investments. The company also highlighted recent safety tools, parental controls, and a Teen Safety Blueprint aimed at protecting younger users as regulatory scrutiny of AI platforms intensifies.

OpenAI’s Sora 2 AI Video Tool Used to Create Disturbing Child‑Like Content on TikTok

OpenAI’s Sora 2 AI Video Tool Used to Create Disturbing Child‑Like Content on TikTok

OpenAI’s video‑generation model Sora 2 has been weaponized to produce realistic yet artificial videos that depict children in questionable scenarios. These clips, many of which mimic commercial advertisements, have spread on TikTok and other platforms, prompting concerns about the ease of circumventing existing safeguards. While OpenAI asserts strict policies against child exploitation, the rapid emergence of such content highlights gaps in moderation and the need for more robust safeguards. Industry observers, child‑protection groups, and policymakers are calling for stronger design‑by‑default protections to prevent misuse of AI‑generated media.

OpenAI Introduces New Personality Controls for ChatGPT

OpenAI Introduces New Personality Controls for ChatGPT

OpenAI has added a suite of personality controls to ChatGPT, allowing users to fine‑tune the bot’s tone and style. The new "Characteristics" settings let users adjust warmth, enthusiasm, the use of headers and lists, and emoji presence with simple "more," "default," or "less" options. These controls sit alongside the existing "Base style and tone" menu, which offers presets such as Professional, Friendly, Candid, Quirky, Efficient, Nerdy, and Cynical. Accessible via the Personalization menu, the settings take effect instantly, giving users granular control over how the AI communicates.

OpenAI Adds Warmth and Enthusiasm Controls to ChatGPT

OpenAI Adds Warmth and Enthusiasm Controls to ChatGPT

OpenAI has expanded the personalization options for its ChatGPT AI chatbot, allowing users to adjust the model's tone with new Warm, Enthusiastic, Header & Lists, and Emoji settings. The changes follow user complaints about the tone of the recent GPT‑5.2 release. In a post on X, OpenAI explained that each option can be set to more, less, or default, giving users granular control over how conversational and friendly the assistant feels. The new controls build on earlier style options such as Professional, Candid, and Quirky that were introduced with GPT‑5.1.

Safety Concerns Rise Over Humanoid Robots After Lawsuit and Stunt Incident

Safety Concerns Rise Over Humanoid Robots After Lawsuit and Stunt Incident

A lawsuit filed by former safety engineer Robert Gruendel alleges that Figure AI's Figure 02 humanoid robot is capable of exerting force sufficient to fracture a human skull, and that he was dismissed after raising safety concerns. Figure AI denies the claim, attributing his termination to poor performance. Meanwhile, a demonstration by Chinese robotics firm Engine AI saw its CEO knocked to the ground by the company's T800 robot, raising questions about the strength and safety of such machines. Both events highlight growing scrutiny over the risks associated with advanced humanoid robots.

OpenAI Introduces Adjustable Warmth, Enthusiasm, and Emoji Settings for ChatGPT

OpenAI Introduces Adjustable Warmth, Enthusiasm, and Emoji Settings for ChatGPT

OpenAI has added new personalization controls to ChatGPT, allowing users to adjust the model's warmth, enthusiasm, and emoji usage. These options appear in the Personalization menu and can be set to More, Less, or Default. The changes complement existing style selections such as Professional, Candid, and Quirky. The update follows earlier adjustments after user feedback on tone, and it has sparked discussion among academics about the potential impact of overly affirming chatbot behavior on user experience.

New York Governor Signs AI Safety Legislation

New York Governor Signs AI Safety Legislation

New York Governor Kathy Hochul signed the RAISE Act, a law aimed at holding large artificial intelligence developers accountable for model safety. The legislation requires companies to disclose safety protocols and report incidents within 72 hours, while establishing fines of up to $1 million for a first violation and $3 million for subsequent breaches. An oversight office within the Department of Financial Services will monitor compliance and issue annual reports. The governor also approved two additional AI measures targeting the entertainment sector, even as President Trump pushes for a national, less burdensome standard.

Engineering Leaders Must Prove AI Impact on Outcomes

Engineering Leaders Must Prove AI Impact on Outcomes

CFOs are demanding evidence that AI spending translates into measurable business results, not just activity metrics. While AI can speed up individual coding tasks, those gains often do not scale to system‑level productivity. Leaders are urged to redirect the time saved by AI into quality improvement, technical debt reduction, and high‑friction initiatives such as legacy migrations and security remediation. Leveraging engineering intelligence platforms provides the data needed to link AI usage with throughput, quality, and customer‑visible outcomes, enabling executives to answer hard budget questions with numbers instead of anecdotes.

Anthropic Launches Claude AI Extension for Chrome Browsers

Anthropic Launches Claude AI Extension for Chrome Browsers

Anthropic has opened its Claude AI assistant to Chrome users through a browser extension available to Claude Pro, Team, and Enterprise subscribers. The extension lets Claude see and interact with live webpages, schedule meetings, organize Google Drive files, and even record custom workflows that the AI can repeat on demand. While the tool offers powerful automation—including handling passwords and multi‑tab tasks—users are cautioned about the breadth of permissions granted, as the AI gains deep access to personal browsing data. Anthropic advises against using the automated features for sensitive activities like banking, underscoring ongoing privacy considerations as AI agents become more integrated into everyday web use.

Resolve AI Secures $1 Billion Valuation in Series A Led by Lightspeed

Resolve AI Secures $1 Billion Valuation in Series A Led by Lightspeed

Resolve AI, a startup building an autonomous site reliability engineer (SRE) platform, announced a Series A financing led by Lightspeed Venture Partners that carries a headline valuation of $1 billion. The company, founded less than two years ago by former Splunk executives Spiros Xanthos and Mayank Agarwal, reports annual recurring revenue of about $4 million. Resolve AI’s technology automates the detection, diagnosis, and remediation of production issues, addressing a growing talent shortage in SRE roles as cloud environments become more complex. The round’s blended valuation is lower than the headline figure due to a multi‑tranche structure, a model gaining popularity among high‑growth AI startups.