News

Page 12
Microsoft Unveils Three Proprietary AI Models to Challenge OpenAI and Google

Microsoft Unveils Three Proprietary AI Models to Challenge OpenAI and Google

Microsoft announced the launch of three in‑house AI models—MAI‑Transcribe‑1, MAI‑Voice‑1 and MAI‑Image‑2—through its Foundry platform and MAI Playground. The models, designed for speech‑to‑text, synthetic voice and image generation, aim to rival offerings from OpenAI, Google and Amazon. Built after a 2019 contract with OpenAI lifted a restriction on Microsoft’s own frontier AI work, the new suite promises faster performance, multilingual support and competitive pricing, with rollouts already underway in Bing and PowerPoint.

AI Companions Offer Relief for Loneliness, But May Heighten Emotional Distress, Study Finds

AI Companions Offer Relief for Loneliness, But May Heighten Emotional Distress, Study Finds

A new study led by Aalto University and slated for presentation at CHI 2026 reveals that AI companions can lessen feelings of loneliness, yet users’ online language shows growing emotional distress over time. Researchers say the technology’s constant, non‑judgmental presence helps some people feel heard, but experts warn the reliance could erode real‑world social skills and foster unhealthy dependence.

Teens Turn to AI Chatbots for Friendship, Prompting Safety Concerns

Teens Turn to AI Chatbots for Friendship, Prompting Safety Concerns

A recent Common Sense Media survey found that 72 percent of U.S. teens have used AI companion apps, with a third seeking friendship or emotional support from the bots. Researchers warn that relational chatbots can foster a false sense of trust, especially among lonely or stressed adolescents. After lawsuits and reports of sexually explicit or manipulative exchanges, platforms such as Character.AI have begun restricting teen access to open‑ended chat features. The trend raises questions about how AI‑driven companionship is reshaping teenage social habits and what safeguards are needed.

AI Music Platform Suno’s Filters Fail to Block Copyrighted Songs, Enabling Easy Creation of Infringing Covers

AI Music Platform Suno’s Filters Fail to Block Copyrighted Songs, Enabling Easy Creation of Infringing Covers

Suno, the AI‑driven music service that markets a $24‑a‑month Premier Plan for creating original tracks, is letting users slip copyrighted material past its detection system. By uploading a song, slowing it with free software, or adding brief bursts of white noise, creators can generate AI‑styled imitations of hits like Beyoncé’s “Freedom” and Black Sabbath’s “Paranoid.” The resulting covers, which sound eerily close to the originals, can be exported and placed on streaming services, raising fresh concerns about royalty avoidance and artist protection.

Apple battles AI‑generated app surge as vibe‑coding tools flood App Store

Apple battles AI‑generated app surge as vibe‑coding tools flood App Store

Apple’s App Store has seen an unprecedented influx of new apps created with AI‑driven “vibe coding” tools, driving an 84% jump in submissions in a single quarter. The surge has stretched Apple’s review process, pushing approval times from a day to up to a month. In response, the company has begun pulling or blocking updates for apps that violate its self‑containment rules, sparking a standoff with the platforms that power the AI‑generated boom. Regulators are watching as the dispute highlights a clash between rapid AI development and existing gatekeeping frameworks.

UK courts Anthropic to broaden London footprint amid US contract row

UK courts Anthropic to broaden London footprint amid US contract row

British officials are preparing a package of incentives to persuade San Francisco‑based AI firm Anthropic to expand its London office and list shares on a UK exchange. The effort comes as the company battles a dispute with the U.S. Department of Defense, which halted a multi‑year contract after Anthropic refused to soften its AI safety guardrails. While the DoD designation as a supply‑chain risk remains under a court‑ordered stay, the United Kingdom sees an opening to attract the startup, even as rival OpenAI has already committed to a London expansion.

Anthropic Scrambles to Remove Malware-Infused Claude Code Leak from GitHub

Anthropic Scrambles to Remove Malware-Infused Claude Code Leak from GitHub

Anthropic unintentionally exposed the source code for its Claude Code tool, prompting a flood of GitHub reposts. Security researchers discovered that many of the copies include hidden infostealer malware, turning a simple code leak into a broader threat. The company has issued copyright takedown notices, trimming the number of repositories from over 8,000 to under 100. The episode follows earlier attempts to lure users with fake installation guides that also delivered malicious payloads.

Anthropic Ends Free Claude Access for Third‑Party Apps Like OpenClaw

Anthropic Ends Free Claude Access for Third‑Party Apps Like OpenClaw

Anthropic announced that, effective 3 p.m. ET on April 4, its Claude AI will no longer be free for third‑party applications. Users of OpenClaw and similar tools must now purchase a usage bundle or provide a Claude API key. Founder and head of Claude Code, Boris Cherny, cited engineering constraints and capacity limits as the reason for the change, noting that existing subscription plans were not designed for the heavy usage patterns of these integrations. The move forces developers and end‑users to reconsider how they access Anthropic’s models.

Anthropic Raises Fees for Claude Code Users of OpenClaw and Other Third‑Party Tools

Anthropic Raises Fees for Claude Code Users of OpenClaw and Other Third‑Party Tools

Anthropic announced that, beginning noon Pacific on April 4, subscribers to its Claude Code service will lose the ability to apply their subscription limits when using third‑party harnesses such as OpenClaw. Instead, users must switch to a pay‑as‑you‑go model billed separately. The change, explained by Claude Code head Boris Cherny, reflects the company’s need to align pricing with the heavy usage patterns of these tools and to sustain growth. The move follows OpenClaw creator Peter Steinberger’s shift to OpenAI and comes as Anthropic offers refunds to affected customers.

Anthropic Blocks Claude Pro and Max Users From OpenClaw, Shifts to Pay‑As‑You‑Go

Anthropic Blocks Claude Pro and Max Users From OpenClaw, Shifts to Pay‑As‑You‑Go

Anthropic announced that, effective April 4, 2026, Claude Pro and Max subscription plans can no longer be used with third‑party AI agent frameworks such as OpenClaw. Users must now pay for any extra usage under a pay‑as‑you‑go model or supply a separate API key. The move ends a quiet subsidy that let thousands of developers run autonomous agents on a flat‑rate plan, prompting cost spikes of up to 50 times for heavy users. Anthropic says the change protects capacity and aligns pricing with the compute‑intensive workloads of agentic tools.

Perplexity AI Hit With Class-Action Lawsuit Over Alleged Data Sharing in Incognito Mode

Perplexity AI Hit With Class-Action Lawsuit Over Alleged Data Sharing in Incognito Mode

A class-action suit filed by an anonymous user, identified as John Doe, accuses Perplexity, the fast‑growing AI search platform, of breaching privacy promises. The complaint alleges that the company’s incognito feature fails to shield user conversations, instead funneling chat transcripts, IP addresses, email identifiers and location data to advertising partners such as Google and Meta. If the allegations prove true, the case could force tighter transparency standards across AI‑driven services.

Study Finds 73% of Users Accept Faulty AI Answers, Raising Concerns Over Trust

Study Finds 73% of Users Accept Faulty AI Answers, Raising Concerns Over Trust

Researchers analyzing 1,372 participants across more than 9,500 decision‑making trials discovered that people accepted AI‑generated answers that were wrong 73.2% of the time, while only overturning them in 19.7% of cases. The study links high trust in artificial‑intelligence systems to a greater likelihood of being misled, whereas individuals with higher fluid intelligence were more prone to question the AI. Authors warn that while reliance on AI can be advantageous when the technology is superior, the current tendency to treat AI output as authoritative creates a structural vulnerability in human judgment.