News

Gen AI Live

A lot happens in Gen AI. Gen AI Live is the definitive resource for executives who want only the signal. Just curated, thoughtful, high impact Gen AI news.
Thank you! Your submission has been received!
Oops! Something went wrong while submitting the form.
Models
March 18, 2026

Meta-backed Manus has launched a desktop application

Manus “My Computer” brings AI directly to your desktop, letting it access files, run commands, and automate tasks locally, turning your computer into a powerful, always-on AI assistant.
Expand

Manus “My Computer” is a desktop AI capability that moves AI from the cloud to your local machine, allowing it to directly access files, run terminal commands, and control applications. It can automate repetitive tasks like organizing files, renaming documents, or even building full applications using local tools.

By leveraging your computer’s resources, including GPUs, it unlocks faster processing and continuous background execution.

You can also trigger tasks remotely while your system handles the work. This creates a seamless bridge between cloud intelligence and local computing, turning your personal computer into an active AI workspace.

#
Meta
Models
March 18, 2026

Google's latest investment in open source security for the AI era

Google is using AI to strengthen open-source security by automatically finding vulnerabilities, predicting threats, and helping developers fix issues faster, improving overall software safety across the ecosystem.
Expand

Google is advancing open-source security using AI systems that can detect vulnerabilities, predict potential exploits, and assist in fixing issues before they are widely used in attacks. Tools like its AI agent “Big Sleep” have already identified critical flaws in widely used open-source projects and even helped prevent real-world exploitation.

This approach shifts security from reactive to proactive, allowing developers to address risks earlier in the lifecycle.

It also helps security teams scale their impact by automating repetitive tasks, improving protection across widely used software without increasing manual effort.

#
Google
Models
March 18, 2026

Introducing GPT‑5.4 mini and nano

OpenAI introduced GPT-5.4 Mini and Nano as fast, low-cost models designed for high-volume workloads, offering strong reasoning, coding, and multimodal performance while improving efficiency for real-time applications.
Expand

OpenAI launched GPT-5.4 Mini and Nano as compact, high-efficiency versions of its flagship GPT-5.4 model, built for speed, scalability, and lower cost. These models are designed for high-volume workloads such as real-time applications, coding tasks, and enterprise automation.

Despite their smaller size, they deliver strong reasoning, coding, and multimodal capabilities, with Mini achieving performance close to the full model in many cases.

The focus is on reducing latency and compute costs while maintaining quality, making them suitable for developers and businesses that need reliable AI at scale without the expense of larger models.

#
OpenAI
Models
March 17, 2026

NVIDIA Rubin Platform

NVIDIA’s Vera Rubin platform is a full-stack AI infrastructure designed for large-scale reasoning and agentic workloads, delivering faster inference, lower cost per token, and improved efficiency for enterprise AI systems.
Expand

NVIDIA’s Vera Rubin platform is a next-generation AI infrastructure built to support large-scale reasoning and agentic AI workloads. It combines GPUs, CPUs, networking, and data processing into a unified system, enabling faster training and more efficient inference.

The platform reduces bottlenecks in memory and communication, delivering higher performance with lower cost per token compared to previous architectures. Designed for AI factories and enterprise deployments, Rubin allows organizations to run complex, long-context workflows at scale.

This shift reflects NVIDIA’s move from standalone chips to fully integrated AI systems optimized for real-world, production-level AI applications.

#
Nvidia
Models
March 17, 2026

NVIDIA announces NemoClaw for the OpenClaw community

NVIDIA introduced NemoClaw, an open-source platform for building secure enterprise AI agents, enabling businesses to deploy autonomous systems with strong privacy, control, and scalable automation capabilities.
Expand

NVIDIA announced NemoClaw, an open-source AI agent platform designed for enterprise use, focusing on secure, scalable deployment of autonomous systems. It enables businesses to build and run AI agents that can automate workflows, execute tasks, and integrate with tools while maintaining strong privacy and data control.

The platform is hardware-agnostic and integrates with NVIDIA’s broader AI stack, including NeMo and Nemotron models, making it flexible for different environments.

By combining open-source access with enterprise-grade security and automation capabilities, NemoClaw reflects NVIDIA’s push toward agentic AI as the next stage of enterprise software and workflow transformation.

#
Nvidia
Models
March 16, 2026

New Microsoft AI Agents to help modernize enterprises

Microsoft introduced new AI agents to help enterprises modernize operations. These agents automate complex workflows, manage tasks across apps, and integrate organizational data to improve productivity and streamline business processes.
Expand

Microsoft has introduced a new generation of AI agents designed to help enterprises modernize their operations and workflows. Built within Microsoft 365 Copilot, these agents can complete complex, multi-step tasks across applications by accessing business data and tools.

The initiative includes capabilities like Copilot Cowork and an expanded ecosystem that allows organizations to create and deploy custom agents tailored to their workflows. These agents automate routine activities such as document creation, data analysis, scheduling, and collaboration tasks.

By reducing manual work and improving decision support, Microsoft aims to help enterprises accelerate digital transformation and increase employee productivity across modern workplaces.

#
Microsoft
Spotlight
March 13, 2026

Top Gen AI Competency Partners for AWS AI Services in 2026

AWS Generative AI Competency partners help enterprises build and scale AI applications on AWS. Leading companies include GoML, Caylent, Quantiphi, Loka, TCS, and Deloitte, known for deploying AI with services like Bedrock and SageMaker.
Expand

Many enterprises rely on AWS Generative AI Competency partners to implement AI solutions using services such as Amazon Bedrock and Amazon SageMaker. These partners provide expertise in designing, deploying, and scaling AI systems across industries.

Leading companies in this space include GoML, Caylent, Quantiphi, Loka, TCS, and Deloitte. Each brings specialized strengths, such as rapid generative AI pilot deployment, industry specific consulting, and large scale cloud integration.

Organizations choose these partners to accelerate AI adoption, ensure security and compliance, and integrate AI into existing enterprise systems. Selecting the right partner depends on business goals, industry requirements, and the complexity of the AI solutions being deployed.

#
GoML
Ecosystem
March 13, 2026

Twenty years of Amazon S3 and building what’s next

Amazon S3 turned 20 years old as a core AWS storage service. It evolved from simple cloud storage into a large data platform that supports analytics, AI workloads, and applications at global scale.
Expand

Amazon Simple Storage Service, known as Amazon S3, launched in March 2006 to provide scalable cloud storage for developers and businesses. Over two decades, it grew from a basic object storage service into a foundation for modern data infrastructure.

S3 now stores hundreds of trillions of objects and supports workloads such as backups, data lakes, analytics, and AI applications. AWS improved the service with features like stronger consistency, faster storage classes, and tools for managing large datasets.

These upgrades allow developers to build high performance applications directly on S3. The next phase focuses on tighter integration with analytics and AI systems so data stored in S3 can power more advanced cloud workloads.

#
AWS
Spotlight
March 12, 2026

Elixia’s Transportation Management System optimization and 3D cargo modeling for fleet logistics

Elixia partnered with GoML to modernize its Transportation Management System with scalable cloud infrastructure and 3D cargo visualization. The upgrade improved delivery planning, increased efficiency, and helped logistics teams optimize vehicle cargo utilization.
Expand

Elixia worked with GoML to upgrade its Transportation Management System to support large scale logistics operations. The platform manages fleet workflows, telematics data, billing, and compliance for more than 85,000 connected vehicles.

As delivery volumes increased, the system needed stronger infrastructure and better cargo planning tools. GoML introduced a containerized cloud architecture using AWS ECS Fargate to enable automatic scaling and reliable performance.

The team also built an interactive 3D cargo visualization module that shows how goods fit inside vehicles. This helped dispatch teams inspect loads, plan shipments more accurately, and improve delivery efficiency. The upgrade led to better cargo space utilization, faster dispatch validation, and reduced infrastructure management effort.

#
GoML
Industries
Models
March 12, 2026

Introducing Copilot Health

Copilot Health is a Microsoft AI tool that analyzes wearable data, lab results, and medical records to provide personalized health insights, helping people understand their health information before visiting a doctor.
Expand

Copilot Health is a new AI feature from Microsoft that helps people understand their health information by analyzing data from multiple sources. Users can connect wearable devices, medical records, and lab results in one secure space inside the Copilot app.

The system then summarizes the data and provides clear explanations, insights, and suggestions so users can better understand their health status.

It can also help users prepare questions before doctor visits and find healthcare providers based on factors such as specialty, location, and insurance. Microsoft designed the tool with strong privacy controls, and it does not replace doctors but helps users make better use of medical consultations.

#
Healthcare
#
Microsoft
Models
March 12, 2026

How we’re reimagining Maps with Gemini

Google introduced Ask Maps and Immersive Navigation in Google Maps. Ask Maps uses Gemini AI to answer questions about places and routes, while Immersive Navigation shows realistic 3D views with detailed road guidance.
Expand

Google Maps now includes two major AI features called Ask Maps and Immersive Navigation. Ask Maps uses Google’s Gemini AI to let users ask natural language questions such as where to meet friends, find charging spots, or plan road trips.

The system analyzes user preferences, reviews, and location data to provide personalized suggestions and route plans. Immersive Navigation improves the driving experience with realistic 3D visuals of buildings, terrain, lanes, traffic lights, and intersections.

This helps drivers understand complex turns and road layouts more easily. Together, these updates transform Google Maps from a simple navigation tool into an intelligent assistant that helps users explore places and plan journeys more effectively.

#
Google
Models
March 12, 2026

New spaces for AI innovation and discovery

Platform 37 is Google’s renamed London headquarters near King’s Cross. It includes “The AI Exchange,” a public space where people can explore artificial intelligence through exhibitions, events, and educational programs.
Expand

Platform 37 is the new name for Google’s London headquarters located near King’s Cross station. The name references its location and a milestone in artificial intelligence history. As part of this initiative, Google introduced “The AI Exchange,” a public space designed to help people understand and engage with AI technologies.

Visitors can attend exhibitions, workshops, and educational programs that explain how AI works and how it impacts society.

The goal is to encourage open discussion about artificial intelligence, promote learning, and help the public interact with emerging technologies developed by Google and its research teams.

#
Google
Models
March 12, 2026

Introducing Groundsource

Groundsource is a Google Research project that uses the Gemini AI model to convert unstructured news reports into structured datasets. It extracts details about events like floods to improve forecasting and research.
Expand

Groundsource is a research initiative by Google that uses the Gemini AI model to transform news reports into structured data. News articles often contain valuable information about real-world events such as floods, disasters, and infrastructure damage, but this information exists in unstructured text.

Groundsource analyzes millions of news reports and extracts key details like location, timing, and event impact. This process creates large datasets that researchers and machine learning systems can use to study patterns and improve forecasting models.

One key use case is flash flood prediction, where news reports help fill gaps in traditional weather data collected from sensors and monitoring stations.

#
Google
Models
March 12, 2026

Designing AI agents to resist prompt injection

OpenAI outlines strategies for designing AI agents that resist prompt injection attacks. The guide explains risks, safe system design, and layered defenses that help agents follow user intent instead of malicious instructions.
Expand

OpenAI explains how developers can design AI agents that resist prompt injection attacks, a security threat where hidden instructions manipulate an AI to ignore user intent. These attacks often appear in emails, webpages, or documents that the agent processes.

To reduce risk, OpenAI recommends layered defenses such as separating trusted system instructions from external content, restricting tool access, validating inputs, and continuously testing agents through automated red teaming.

The company also uses reinforcement learning based automated attackers to discover new vulnerabilities before they appear in real world deployments. While defenses are improving, prompt injection remains a long term security challenge for AI agents that interact with external data.

#
OpenAI
Models
March 11, 2026

Introducing Nemotron 3 Super

NVIDIA Nemotron 3 Super is a 120B-parameter open model designed for agentic AI systems. It combines Mamba, Transformer, and Mixture-of-Experts architectures to deliver efficient reasoning, long context handling, and high-throughput AI workflows.
Expand

NVIDIA Nemotron 3 Super is an open-weight 120-billion-parameter AI model designed to power large-scale agentic AI systems. It uses a hybrid architecture that combines Mamba sequence modeling, Transformer attention, and Mixture-of-Experts routing to improve reasoning accuracy and computational efficiency.

The model supports a 1-million-token context window, enabling AI agents to process long documents, codebases, and complex workflows. By activating only a subset of experts per token, the system delivers faster inference and higher throughput while controlling compute costs.

Nemotron 3 Super is built for enterprise workloads such as software development automation, cybersecurity analysis, and multi-step task execution by autonomous AI agents.

#
Nvidia
Models
March 11, 2026

Everything is Computer

“Everything is Computer” explains Perplexity’s vision that AI will become the main interface for computing. Instead of manually using apps, people describe goals and AI agents complete tasks across tools and data automatically.
Expand

In “Everything is Computer,” Perplexity presents a new view of computing where artificial intelligence becomes the central interface for getting work done. Instead of opening multiple apps and completing tasks step by step, users simply describe their goal in natural language and an AI system executes the workflow.

The company introduced systems like Perplexity Computer and Personal Computer, which act as digital workers that can research information, manage files, coordinate tools, and complete multi-step tasks automatically.

These systems combine multiple AI models and software tools to handle complex requests, marking a shift from traditional software interfaces to autonomous AI agents that manage digital workflows.

#
Perplexity
Models
March 11, 2026

New ways to learn math and science in ChatGPT

OpenAI introduced new ways to learn math and science in ChatGPT with interactive visual explanations. Users can explore formulas, adjust variables, and see equations and graphs update instantly to understand concepts better.
Expand

OpenAI has introduced new ways to learn math and science in ChatGPT through interactive visual explanations. The feature allows users to explore formulas, change variables, and instantly see how graphs, equations, and diagrams respond in real time.

This approach helps students understand how mathematical and scientific concepts work rather than only reading explanations. The learning modules currently cover more than 70 core topics, including ideas like the Pythagorean theorem, circle equations, and the ideal gas law.

By combining written guidance with visual interaction, ChatGPT supports deeper conceptual understanding and hands-on exploration of complex topics for students, educators, and lifelong learners.

#
OpenAI
Expert Views
March 10, 2026

WebMCP and AI orchestration

WebMCP introduces a structured way for AI agents to interact with websites. Instead of fragile screen scraping, agents call defined actions directly, making web automation more reliable for enterprise AI workflows.
Expand

WebMCP represents a shift in how AI agents interact with the web. Today, most agents rely on screen scraping, DOM parsing, or simulated clicks to complete tasks on websites. These methods often break when page layouts change or dynamic elements behave differently.

WebMCP replaces this fragile approach with a structured interface where websites expose actions that agents can call directly.

With this model, agents no longer guess how to navigate a page. Instead, they access defined tools that trigger workflows such as submitting forms, retrieving data, or completing transactions. This improves reliability and enables AI orchestration systems to run complex enterprise workflows across web platforms more consistently.

#
GoML
Models
March 10, 2026

OpenAI to acquire Promptfoo

OpenAI plans to acquire Promptfoo, an AI security and testing platform. The technology will help enterprises evaluate AI systems, detect risks such as prompt injection, and improve reliability of AI coworkers.
Expand

OpenAI has announced plans to acquire Promptfoo, a platform that helps developers test and secure AI applications. Promptfoo provides tools to evaluate prompts, simulate adversarial attacks, and detect risks such as prompt injection, jailbreak attempts, and data leakage.

The acquisition will integrate Promptfoo’s technology into OpenAI Frontier, a platform designed to build and manage AI coworkers. With automated evaluation and security testing built into development workflows, enterprises will be able to identify risks earlier and maintain oversight of AI systems.

As companies deploy AI agents in real workflows, testing, security, and governance have become critical for reliable and responsible AI deployment.

#
OpenAI
Spotlight
March 9, 2026

AI email response generator boosts efficiency by 62% for Civic

Civic partnered with GoML to build an AI email response generator that helps congressional offices manage large volumes of constituent messages. The system automates classification and response drafting, improving communication efficiency by 62%.
Expand

Civic partnered with GoML to enhance its Revere platform with an AI email response generator designed for congressional offices. These offices receive thousands of constituent emails each week, which previously required manual review, topic tagging, and response writing.

GoML built a system that automatically classifies messages by policy topic and viewpoint, groups similar emails together, and generates draft responses aligned with each congressperson’s communication style.

The system uses large language models and retrieval pipelines that pull policy context from legislative data and news sources. Staff review and approve responses before sending. The solution improved operational efficiency, reduced manual classification work, and enabled faster responses to constituent communication.

#
GoML
Spotlight
March 9, 2026

AI powered coding for kids platform for AngelQ

AngelQ partnered with GoML to build an AI powered coding platform for children. The system converts natural language commands into visual programming blocks and animations, helping kids learn coding through storytelling.
Expand

AngelQ partnered with GoML to develop an AI powered coding platform designed for children aged 5 to 12. Traditional coding tools rely on manual block programming, which often slows experimentation and reduces engagement for young learners.

The new platform allows children to type simple commands such as “make the dog walk,” and the system automatically converts them into visual programming blocks and animated character actions. This helps kids connect their ideas with the logic behind code while creating interactive stories.

The proof of concept also showed strong learning impact, including faster interaction, better understanding of programming logic, and higher engagement through animation driven storytelling.

#
GoML
Ecosystem
March 9, 2026

Amazon announced general availability of Policy in Amazon Bedrock

Amazon announced general availability of Policy in Amazon Bedrock AgentCore, allowing organizations to manage agent tool access and input validation through centralized policies using natural language that converts into Cedar rules.
Expand

Amazon released Policy in Amazon Bedrock AgentCore with general availability, introducing centralized governance for AI agent tool interactions. Security and compliance teams can now define access permissions and input validation rules outside the agent code using natural language instructions.

These rules automatically translate into Cedar, the AWS open source policy language used for fine grained authorization. The feature allows organizations to control how agents interact with tools, services, and external systems while maintaining strong security and compliance standards.

With centralized policy management, companies can scale AI agent deployments more safely while ensuring consistent governance across applications built on Amazon Bedrock.

#
AWS
#
Bedrock
Ecosystem
March 9, 2026

Introducing Amazon connect health

Amazon launched Amazon Connect Health with five AI agents for healthcare tasks such as patient verification, appointment management, documentation, patient insights, and medical coding. The HIPAA eligible system integrates into clinical workflows quickly.
Expand

Amazon introduced Amazon Connect Health, a healthcare focused AI solution that includes five specialized AI agents designed to support clinical operations. These agents handle tasks such as patient verification, appointment management, patient insights, ambient documentation, and medical coding.

The platform is HIPAA eligible and can integrate with existing healthcare systems and workflows within days. By automating administrative and documentation tasks, healthcare teams can reduce operational workload and improve patient care efficiency.

Amazon Connect Health is built to help hospitals and healthcare providers streamline communication, manage patient interactions, and support clinicians with AI powered assistance across daily clinical processes.

#
AWS
Models
March 9, 2026

Powering Frontier Transformation with Copilot and agents

Microsoft announced new Copilot and AI agent capabilities to help organizations become “Frontier Firms.” These tools automate workflows, analyze work data, and enable human-AI collaboration across Microsoft 365 apps.
Expand

Microsoft introduced new Microsoft 365 Copilot and AI agent capabilities designed to help organizations transform into “Frontier Firms,” where humans and AI agents collaborate to improve productivity and business processes. The platform uses an intelligence layer called Work IQ that analyzes emails, files, meetings, and chats to understand a user’s work context and suggest actions.

Copilot can recommend or activate specialized agents that handle tasks such as research, document creation, data analysis, and workflow automation directly within apps like Word, Excel, and Teams.

These capabilities enable multi step workflows, personalized assistance, and faster decision making across organizations adopting AI powered work systems.

#
Microsoft
Models
March 9, 2026

Copilot Cowork

Microsoft introduced Copilot Cowork, a new AI powered way to complete tasks in Microsoft 365. It helps users research, prepare documents, manage emails, and automate workflows across workplace apps.
Expand

Microsoft announced Copilot Cowork, a new approach to getting work done with AI inside Microsoft 365. The system acts like an AI teammate that can research information, prepare meeting briefs, organize files, and manage emails across workplace tools such as Outlook, Word, and Excel.

Copilot Cowork focuses on completing multi step tasks rather than responding to single prompts. It can analyze documents, gather context from different sources, and generate outputs like reports or summaries while users supervise the process.

This model reflects a shift toward agent based productivity tools where AI works alongside employees to automate routine tasks and improve workflow efficiency.

#
Microsoft
Models
March 6, 2026

Introducing GPT‑5.4

OpenAI introduced GPT-5.4, an updated version of its GPT-5 model family designed to improve reasoning, coding, and complex knowledge work. The model is available in ChatGPT and through developer APIs, with variants such as GPT-5.4 Thinking and GPT-5.4 Pro for more demanding tasks.
Expand

GPT-5.4 is OpenAI’s latest model designed to improve reasoning, coding, and large scale knowledge tasks. It introduces stronger multi step reasoning capabilities, allowing the model to solve complex problems in areas such as engineering, research, and data analysis.

A major advancement is its expanded context window, which can process extremely large documents, datasets, or codebases within a single interaction. This helps users analyze long reports, legal files, or technical systems more efficiently.

GPT-5.4 also supports agent driven workflows, enabling AI systems to coordinate tools and complete multi stage tasks. Overall, the model focuses on reliability, deeper reasoning, and practical applications for professional and enterprise environments.

#
OpenAI
Models
March 5, 2026

The five AI value models driving business reinvention

OpenAI outlines five AI value models that help businesses move beyond isolated pilots to full transformation, focusing on workforce empowerment, expert systems, and agent-driven process reengineering for long-term impact.
Expand

OpenAI introduces five AI value models as a structured roadmap for business transformation, moving beyond scattered AI pilots to system-wide impact. These models include workforce empowerment, expert capability, AI-native distribution, systems management, and process reengineering.

Each stage builds on the previous, helping organizations gradually scale from productivity gains to full operational reinvention. The framework highlights that true value comes from sequencing AI adoption correctly, rather than jumping directly to automation.

Companies that follow this approach can improve efficiency, enable smarter decision-making, and eventually redesign how they deliver products and services using AI-driven workflows and autonomous agents.

#
OpenAI
Expert Views
March 5, 2026

OpenAI teams up with AWS on a new Stateful Runtime Environment for smarter enterprise AI

OpenAI and AWS are developing a Stateful Runtime Environment for enterprise AI. It lets AI agents keep memory, context, and workflow state so they can run complex multi step tasks reliably at production scale.
Expand

OpenAI and Amazon Web Services announced a partnership to build a Stateful Runtime Environment designed for enterprise AI agents. The system runs inside Amazon Bedrock and uses OpenAI models while integrating with AWS infrastructure.

Unlike traditional stateless APIs that respond to single prompts, the runtime allows agents to keep memory, track past actions, and maintain context across long workflows. This enables AI systems to operate across tools, data sources, and enterprise applications while preserving identity, permissions, and governance controls.

The goal is to help developers build reliable production systems such as customer support automation, IT operations workflows, and finance processes where agents must manage multi step tasks over time.

#
GoML
Models
March 5, 2026

Things stand with the Department of War

Anthropic outlines its position in a dispute with the US Department of Defense over military use of AI. The company opposes unrestricted use of its models for surveillance and autonomous weapons.
Expand

Anthropic explains its stance in an ongoing conflict with the US Department of Defense about how AI should be used in military systems. The company argues that its models should not be deployed for mass domestic surveillance or fully autonomous weapons without strict safeguards.

US officials pushed for the right to apply AI technology for any lawful military purpose, which Anthropic refused to accept. This disagreement led the Pentagon to label the company a “supply chain risk” and restrict its technology from defense contracts.

Anthropic says the move sets a dangerous precedent and has challenged the decision through legal action.

#
Anthropic
Models
March 5, 2026

Introducing the Adoption news channel

OpenAI launched the Adoption news channel to help business leaders turn AI capabilities into real operational results. It provides practical frameworks, case studies, and insights on scaling AI adoption across organizations.
Expand

OpenAI introduced the Adoption news channel to help organizations move from AI experimentation to real business impact. The channel focuses on practical insights, frameworks, and examples that guide leaders in implementing AI across workflows and operating models.

It targets executives, AI adoption leaders, and transformation teams responsible for scaling AI within enterprises. Topics include identifying where AI creates business value, scaling adoption beyond pilots, redesigning roles and governance, and separating lasting trends from market hype.

The goal is to provide decision making guidance that helps companies turn AI progress into measurable outcomes, stronger workflows, and new business advantages.

#
OpenAI
Models
March 5, 2026

The latest AI news announced in February

Google announced new AI updates in February 2026. Highlights include Gemini 3.1 Pro for stronger reasoning, Nano Banana 2 for faster image generation, and Lyria 3 for AI music creation in the Gemini app.
Expand

Google shared several AI product updates in February 2026 focused on creativity, reasoning, and developer tools. The company introduced Gemini 3.1 Pro, a new model designed to handle complex reasoning tasks and advanced problem solving.

Google also launched Nano Banana 2, a faster image generation model that powers creative workflows across the Gemini app and other Google services. Another highlight is Lyria 3, an advanced music generation model that lets users create short AI generated tracks directly in the Gemini app.

These updates aim to improve productivity, expand creative possibilities, and give developers stronger AI tools to build new applications and experiences.

#
Google
Models
March 4, 2026

Extending single-minus amplitudes to gravitons

The article explains how researchers extended new formulas for single-minus gluon scattering amplitudes to gravitons. The work shows interactions previously assumed impossible can occur under special conditions, advancing theoretical physics research.
Expand

The article discusses research showing how formulas for single-minus scattering amplitudes can be extended from gluons to gravitons. Scattering amplitudes measure the probability that particles interact in a certain way.

For decades, physicists believed tree-level interactions where one gluon has negative helicity and the rest positive had zero amplitude, meaning they should not occur. Researchers found that in a specific kinematic setting called the half-collinear regime, these amplitudes are actually nonzero and can be expressed with a closed-form formula.

The result was discovered with help from GPT-5.2 and verified using known physics constraints. The same mathematical approach can now be applied to gravitons, opening new research directions.

#
OpenAI
Models
March 4, 2026

New tools for understanding AI and learning outcomes

The article explains how AI tools affect student learning outcomes. It highlights both benefits and risks, showing that AI can support learning when used responsibly alongside teachers, guidance, and proper educational design.
Expand

The article explores how artificial intelligence influences student learning outcomes in education. AI tools can provide personalized support, instant feedback, and interactive learning experiences that help students understand concepts more effectively.

They can also assist teachers by automating routine tasks and improving lesson design. However, the article also notes challenges such as overreliance on AI, potential inaccuracies, and concerns about academic integrity.

Effective use requires thoughtful integration into teaching practices, clear guidelines, and teacher oversight. When applied responsibly, AI can enhance learning by making education more adaptive, accessible, and efficient while still maintaining the essential role of human educators in guiding students.

#
OpenAI
Expert Views
March 3, 2026

Why LLM benchmarking on leaderboards is not enough for enterprise AI

Research shows LLM leaderboards can give a misleading view of model capability. Rankings often reflect benchmark optimization rather than real world performance, so enterprises must evaluate models using their own tasks and data.
Expand

The article explains why leaderboard rankings are not a reliable way to judge large language models for enterprise use. Research shows that many models are optimized to perform well on specific benchmark tests rather than real world tasks.

This creates what researchers call a “leaderboard illusion,” where rankings suggest strong capabilities that may not translate to practical applications. Benchmarks can also be distorted by selective reporting, overfitting, and differences in evaluation methods.

For organizations building AI systems, the key takeaway is to evaluate models on real workflows, datasets, and production scenarios instead of relying only on public leaderboards. Use case based testing gives a more accurate view of model performance.

#
GoML
Models
March 3, 2026

Gemini 3.1 flash-lite

Gemini 3.1 Flash-Lite is Google’s fast, low-cost AI model designed for high-volume tasks such as translation, summarization, tagging, and moderation. It focuses on speed and efficiency rather than complex reasoning.
Expand

Gemini 3.1 Flash-Lite is a lightweight AI model from Google built for speed, efficiency, and large-scale use. It belongs to the Gemini 3 model family and targets routine tasks that require fast responses and consistent results. Typical uses include translation, summarization, data extraction, tagging, and content moderation.

The model prioritizes low cost and high throughput instead of deep reasoning, which makes it suitable for large production workloads. Google released it in preview through the Gemini API, Google AI Studio, and Vertex AI for developers and enterprises.

Compared with earlier Flash models, it delivers faster response times and improved efficiency for applications that process massive amounts of data.

#
Google
Models
March 3, 2026

GPT‑5.3 Instant: Smoother, more useful everyday conversations

GPT-5.3 Instant is OpenAI’s updated fast AI model for everyday ChatGPT tasks. It improves conversational flow, reduces hallucinations, increases factual accuracy, and delivers clearer answers while maintaining fast response speed.
Expand

GPT-5.3 Instant is an upgraded version of OpenAI’s widely used ChatGPT model designed for fast, everyday interactions. Instead of focusing on heavy reasoning tasks, it handles common requests such as writing, summarizing documents, answering questions, and simple coding.

The update improves conversational flow, produces clearer and more structured answers, and reduces unnecessary refusals. It also reduces hallucinations and improves reliability when using web information. Tests show hallucinations dropped by about 26.8 percent on web-based queries, reflecting a stronger focus on accuracy.

Overall, GPT-5.3 Instant prioritizes practical usefulness, faster responses, and consistent performance for daily tasks used by millions of ChatGPT users.

#
OpenAI
Ecosystem
February 28, 2026

NVIDIA advances autonomous networks with agentic AI blueprints and telco reasoning models

NVIDIA introduced agentic AI blueprints and a large telco reasoning model to help telecom operators build autonomous networks. These systems use AI agents to analyze data, reason through operations, and automate network decisions.
Expand

NVIDIA announced new agentic AI blueprints and a large telecom reasoning model to help operators build autonomous networks. These networks go beyond simple automation and can understand operator intent, analyze complex situations, and decide actions independently.

NVIDIA released an open 30-billion-parameter large telco model based on Nemotron that telecom companies can train using their own operational data. The company also introduced blueprints for tasks such as network configuration and energy optimization using multi-agent systems. These tools allow AI agents to collaborate, test decisions in simulations, and manage telecom infrastructure more efficiently.

The initiative supports the telecom industry’s shift toward AI-driven network operations and autonomous infrastructure management.

#
Nvidia
Models
February 27, 2026

Joint Statement from OpenAI and Microsoft

OpenAI confirmed that its partnership with Microsoft remains central. Both companies continue working together across research, engineering, and product development to build AI systems and deliver advanced tools through Microsoft Azure.
Expand

OpenAI announced that its long-standing partnership with Microsoft will continue as a core part of its AI strategy. The two companies plan to keep collaborating across research, engineering, and product development to build advanced AI systems and deliver them through Microsoft’s cloud platform, Azure.

Microsoft retains exclusive licensing access to OpenAI’s models and intellectual property for its products and services. At the same time, the partnership structure allows OpenAI to pursue additional collaborations with other companies while maintaining deep integration with Microsoft’s technology ecosystem.

Both organizations say they remain committed to developing powerful AI tools responsibly and ensuring that the benefits of AI reach businesses and individuals worldwide.

#
OpenAI
#
Microsoft
Ecosystem
February 27, 2026

Introducing the Stateful Runtime Environment for Agents in Amazon Bedrock

OpenAI and Amazon built a new stateful runtime environment for agents on Amazon Bedrock. It lets AI agents keep context, memory, and workflow state to run multi-step tasks reliably in production.
Expand

OpenAI announced a stateful runtime environment for AI agents that runs natively on Amazon Bedrock.

The new environment helps agents keep memory, history, tool outputs, permissions, and workflow state across multiple steps, reducing the need for developers to build custom orchestration layers. It lets teams focus on business logic instead of managing stateless requests.

This setup supports complex workflows like customer support automation, internal IT tasks, and finance processes that need context over time. The runtime is optimized for AWS infrastructure and integrates with existing security and governance systems. Availability is planned soon for AWS customers.

#
Bedrock
Models
February 27, 2026

Scaling AI for everyone

OpenAI announced a $110 billion investment round with SoftBank, NVIDIA, and Amazon to expand AI compute, global reach, and infrastructure so more people, businesses, and communities can use advanced AI tools.
Expand

OpenAI published a plan called “Scaling AI for everyone.” It says demand for AI is growing among users, developers, and companies.

To meet that demand, the company secured $110 billion in new funding at a $730 billion valuation with major contributions from SoftBank, NVIDIA, and Amazon. Strategic partnerships with Amazon and NVIDIA will expand compute capacity and infrastructure worldwide.

OpenAI aims to use this expanded support to make products such as ChatGPT, Codex, and the Frontier platform more available and reliable for individuals and businesses. The plan focuses on building systems that can support broader adoption of advanced AI.

#
OpenAI
Models
February 27, 2026

Disrupting malicious uses of AI

OpenAI’s “Disrupting malicious uses of AI” report explains how the company detects and prevents harmful AI abuse, including scams, influence operations, and cyber threats, and shares insights to improve defenses and protect users.
Expand

OpenAI’s “Disrupting malicious uses of AI” report highlights efforts to identify and stop harmful uses of its AI models.

The report shows how threat actors combine AI with traditional tools to run scams, cyberattacks, social engineering, and covert influence operations. OpenAI uses its tools alongside human investigation to detect misuse, ban abusive accounts, and share findings with partners to strengthen defenses.

The company aims to make AI beneficial and safe by monitoring activity, enforcing policies, and improving understanding of how malicious actors operate. Case studies illustrate real abuses and how controls disrupt those threats, helping protect users and support broader safety measures.

#
OpenAI
Spotlight
February 26, 2026

Enterprise AI compliance software for Metatate platform

Metatate built an AI compliance platform with GoML to simplify complex data governance. The system converts legal policies into actionable guidance and lets teams check compliance through natural language queries.
Expand

Metatate partnered with GoML to build enterprise AI compliance software that helps organizations manage complex data and AI governance requirements. Many teams struggled with interpreting dense policy documents and depended heavily on legal experts for compliance checks.

GoML developed an AI powered system that interprets natural language descriptions of data activities and maps them to relevant policies stored in Metatate’s repository. The solution uses an agentic AI framework that handles steps like intent detection, policy retrieval, risk evaluation, and response generation.

Users can interact through a conversational interface, upload documents, and receive structured compliance guidance with policy references, risk levels, and recommended next steps.

#
GoML
Models
February 26, 2026

Get more context and understand translations

Google Translate added new AI features that use Gemini’s language understanding to give alternative phrasing, explain nuance and let users ask follow-up questions for clearer, context-aware translations.
Expand

Google updated Translate with AI features that help users get clearer, context-aware translations.

The service now uses Gemini’s multilingual capabilities to offer multiple phrasing options instead of a single literal result. Users can tap a new “understand” button to see why certain translations were chosen, and use “ask” to follow up with questions about phrasing for specific regions or dialects.

These updates help with idioms, tone and cultural nuance, making conversations and written messages more accurate and natural across languages. The new experience is available now on Android and iOS in the U.S. and India and will reach the web soon.

#
Google
Models
February 26, 2026

Nano Banana 2: Combining Pro capabilities with lightning-fast speed

Google launched Nano Banana 2, its newest AI image model. It combines studio-quality visuals, real-world knowledge, and fast generation across Google tools, making high-quality image creation more accessible.
Expand

Google introduced Nano Banana 2, also known as Gemini 3.1 Flash Image, as its latest AI image generation and editing model.

It brings together the advanced capabilities of Nano Banana Pro with the speed of Gemini Flash, enabling fast, high-quality image creation and editing across Gemini, Search, Google Lens, Vertex AI, and Google Ads.

Nano Banana 2 uses real-world knowledge and web data to render specific subjects accurately, supports resolutions up to 4K, and maintains consistent visuals with improved text rendering and creative control. Google also enhances AI content verification with SynthID and C2PA Content Credentials as part of this rollout.

#
Google
Spotlight
February 25, 2026

Advancing grid reliability with AI for battery energy storage

GoML built an AI chatbot for Enfinite Technologies that answers complex queries across databases and PDFs for oil, gas, and water well operations, improving access to data and visual insights instantly.
Expand

GoML developed an AI chatbot for Enfinite Technologies to streamline data access in oil, gas, and water well operations.

The system classifies user questions, uses text-to-SQL and vector search for accurate responses from a PostgreSQL database and domain-specific PDFs, and delivers results through an API interface.

It uses AWS Lambda, SageMaker Llama2, and vector stores for efficient retrieval and classification. The solution improved query response accuracy and sped up information access with visual output options, helping users get insights quickly from large datasets that previously required manual lookup.

#
GoML
Spotlight
February 25, 2026

Contextual intelligence and customer support for Durabuilt Windows and Doors

GoML and TensorIoT built a generative AI chatbot for Durabuilt Windows & Doors to improve customer support, cut response times by 60%, deliver faster personalized replies, and reduce support costs by 35-45%.
Expand

GoML partnered with TensorIoT to create a proof-of-concept generative AI chatbot for Durabuilt Windows & Doors to solve slow, unscalable customer support and efficiency issues.

The solution uses Amazon Bedrock LLMs, embeddings, OpenSearch vector indexing, and semantic retrieval to deliver fast, contextually relevant responses tailored to customer queries. Results include a 60% reduction in support response time, 45% faster personalized replies, and a 35% cut in support-related costs.

The architecture stores text data in Amazon S3, indexes it for quick retrieval, and generates responses with a retrieval-augmented GenAI model, improving operational efficiency and customer satisfaction.

#
GoML
Spotlight
February 25, 2026

Document automation software for Loft47

GoML built AI-powered document automation for Loft47 to extract structured data and validate signatures from real estate contracts, cutting manual review time by 60-70% and improving accuracy for transaction processing.
Expand

GoML developed document automation software for Loft47 to eliminate manual contract review in real estate transaction processing.

The system uses coordinated AI agents to extract structured data from diverse MLS contract formats, detect and validate signatures, and output ready-to-use JSON for backend systems. This reduced manual review time by 60-70% and achieved at least 80% accuracy on extraction fields in MVP testing.

The solution runs on AWS with secure PDF ingestion, flexible contract configuration, and a web-based review interface for edits and approvals. The automation supports scalable operations, cuts costs, and speeds transaction approvals for brokerages across North America.

#
GoML
Models
February 24, 2026

Anthropic’s Responsible Scaling Policy: Version 3.

Anthropic released Responsible Scaling Policy Version 3.0, a risk governance framework that updates how it assesses and mitigates AI risks, expands transparency with risk reports and safety roadmaps, and adapts safeguards as model capabilities grow.
Expand

Anthropic’s Responsible Scaling Policy Version 3.0 updates its voluntary framework for managing risks from advanced AI systems. The policy explains how safeguards should scale with increasing capabilities, using “if-then” commitments tied to capability thresholds.

It introduces transparency measures like Frontier Safety Roadmaps and periodic Risk Reports to show how risks and mitigations align. The update separates internal plans from broader industry recommendations and aims to reinforce successful elements of earlier versions while improving accountability.

Anthropic says the policy will evolve as AI advances, balancing practical safeguards with the need to address emerging threats and encourage broader industry risk governance.

#
Anthropic
Ecosystem
February 23, 2026

Claude Sonnet 4.6 in Amazon Bedrock, Kiro in GovCloud Regions, new Agent Plugins, and more

AWS highlights Claude Sonnet 4.6 now in Amazon Bedrock, new agent plugins, Kiro available in AWS GovCloud regions, and other launches like EC2 Hpc8a and SageMaker Inference updates.
Expand

AWS Weekly Roundup covers key announcements from Feb 23, 2026. Anthropic’s Claude Sonnet 4.6 model is now available in Amazon Bedrock with high performance for coding, agents, and professional work at lower cost.

AWS added new EC2 Hpc8a instances with improved performance and expanded SageMaker Inference for custom Nova models. Nested virtualization support arrived for EC2.

Kiro, an AI development tool, is supported in AWS GovCloud regions for regulated workloads. AWS also introduced open-source Agent Plugins that extend agent skills for deployment tasks. The post highlights community content on agent memory, tooling, and best practices.

#
AWS
Models
February 23, 2026

The persona selection model

Anthropic’s persona selection model explains why AI assistants like Claude often behave human-like, arguing that large language models simulate human characters learned from data, and post-training refines these personas rather than creating them from scratch.
Expand

Anthropic’s persona selection model describes how AI assistants like Claude develop human-like behavior. The research says models learn to predict text during pre-training by simulating human language and characters, which naturally creates personas.

Post-training then refines the assistant persona to be more helpful and aligned with desired traits, but the core human-like behavior comes from pre-training itself. The paper argues that when a model learns a specific behavior, it may infer broader personality traits, and think of assistant behavior in terms of a character’s psychology.

Understanding this helps explain unexpected AI behaviors and guide safer training practices.

#
Anthropic
Models
February 23, 2026

Detecting and preventing distillation attacks

Anthropic reports industrial-scale distillation attacks on its Claude AI by three labs using fake accounts to extract capabilities at scale, describes how it detects and blocks these attacks, and outlines defensive measures.
Expand

Anthropic says it has detected coordinated distillation attacks by three AI labs that used roughly 24,000 fraudulent accounts to make millions of requests to its Claude model, aiming to extract reasoning, coding, and tool use capabilities for training their own systems.

It explains how these campaigns used proxy services, evaded detection, and targeted Claude’s most valuable features.

Anthropic outlines how it identifies and prevents such activity with classifiers and behavioral fingerprinting, strengthens account verification, and shares threat data with industry partners. The company calls for broader cooperation across AI developers and policymakers to defend against large-scale distillation attacks.

#
Anthropic
Models
February 21, 2026

Anthropic launches Claude code security to scan codebases for security vulnerabilities

Anthropic launched Claude Code Security, an AI tool that scans codebases for vulnerabilities, offers context-aware detection, rechecks findings to cut false positives, and shows suggested patches for human review.
Expand

Anthropic introduced Claude Code Security, a new feature inside Claude Code that uses AI to scan software code for security issues and suggest fixes. The system goes beyond traditional scanners by understanding how code works, not just patterns, and rechecks results to lower false positives.

Detected issues appear on a dashboard with severity and confidence ratings.

Developers keep control and review any suggested patches before applying them. The feature aims to help security teams handle large backlogs of vulnerabilities by combining automated insights with human oversight. It’s currently available in a limited research preview.

#
Anthropic
Models
February 19, 2026

Introducing OpenAI for India

OpenAI launches OpenAI for India, partnering with Tata and Indian institutions to build sovereign AI infrastructure, expand enterprise adoption, boost AI skills, and deepen local presence with new offices and education programs.
Expand

OpenAI has unveiled OpenAI for India, a comprehensive initiative to expand AI access and impact across the country.

Announced at the India AI Impact Summit 2026, the program includes building secure, local AI-ready data centers with Tata Group, accelerating enterprise transformation with ChatGPT Enterprise, and investing in workforce upskilling through certifications and campus partnerships.

OpenAI plans to support education with ChatGPT Edu licenses for students and faculty and establish new offices in Mumbai and Bengaluru. This effort aims to strengthen India’s AI ecosystem, enhance infrastructure, and empower developers, businesses, and learners with cutting-edge AI tools and skills.

#
OpenAI
Expert Views
February 18, 2026

Exploring OpenClaw

OpenClaw is a self hosted AI assistant that runs locally with full user control. It offers local memory, automation, and customization, giving users a private alternative to cloud based AI assistants.
Expand

OpenClaw is a self hosted AI assistant designed for users who want full control over how their AI operates. Instead of relying on cloud platforms and subscriptions, the assistant runs locally and keeps data under the user’s control.

It supports features such as persistent memory, task automation, and deep customization so users can tailor the assistant to their workflows. This approach challenges traditional cloud based AI assistants by prioritizing privacy, ownership, and flexibility.

The rapid growth of OpenClaw shows rising demand for AI systems that people can run and manage themselves while still benefiting from powerful automation and intelligent assistance.

#
GoML
Models
February 18, 2026

Gemini can now create music

The Gemini app now includes Lyria 3, Google DeepMind’s advanced AI model that generates custom 30-second music tracks with vocals, lyrics, and cover art from text, images, or videos.
Expand

Google’s Gemini app has added music generation powered by Lyria 3, the latest music model from Google DeepMind.

This feature lets users type text or upload a photo or video to create a custom 30-second track that includes instrumentals, vocals, lyrics, and shareable cover art. Lyria 3 improves creative control with style, tempo, and vocal options, and embeds a SynthID watermark for AI content identification.

Available globally in several languages for users 18 and older, the goal is fun, unique expression not professional production while ensuring responsible generative AI use with safeguards and verification tools.

#
Google
Ecosystem
February 16, 2026

AWS tools installer V2

AWS has released AWS Tools Installer V2 (preview) for PowerShell, improving module installation speed, adding offline/prerelease support and self-update commands, fixing installation bugs, and removing outdated modules all to simplify tooling management.
Expand

Amazon Web Services announced the preview release of AWS Tools Installer V2 for PowerShell to make managing AWS Tools for PowerShell modules faster and more reliable. V2 speeds up installs by bundling many modules, introduces new commands for self-updating the installer itself, and supports offline and prerelease installations.

It also fixes issues where some modules failed to update, and improves reliability during publishing windows.

New features include installer update notifications and better standard removal support. Some breaking changes require updated firewall settings and removal of certain parameters. Legacy modules can now be uninstalled automatically.

#
AWS
Models
February 14, 2026

OpenAI strengthens Enterprise Security with Lockdown Mode

OpenAI introduced Lockdown Mode for high-security users and Elevated Risk labels across ChatGPT, Atlas, and Codex, reducing prompt injection threats, limiting tool access, and improving oversight.
Expand

OpenAI launched Lockdown Mode, an optional advanced security setting for executives and security teams needing stronger protection against prompt injection and data exfiltration attacks.

In this mode, ChatGPT restricts external tool use, limits browsing to cached content, and disables higher-risk capabilities when deterministic safety guarantees are not possible. Alongside this, OpenAI added Elevated Risk labels across ChatGPT, ChatGPT Atlas, and Codex to clearly flag features that may introduce additional security concerns, such as granting network access.

These protections build on enterprise controls like audit logs, role-based access, and monitoring, with consumer rollout planned soon.

#
OpenAI
Models
February 13, 2026

gpt-5.2 derives a new result in theoretical physics

GPT-5.2 helped derive a new formula for gluon scattering amplitudes previously thought to be zero in certain conditions. Humans and AI collaborated to prove and verify this result.
Expand

OpenAI announced that GPT-5.2 assisted scientists in deriving a new theoretical physics result involving gluons particles that mediate the strong nuclear force. Conventional wisdom held that particular gluon scattering amplitudes vanished, but the research identified a special momentum configuration where they are nonzero.

GPT-5.2 originally conjectured a general formula from patterns in simpler cases, and an internal model then produced and verified a formal proof.

The work, authored by researchers from multiple institutions including OpenAI, is now available as a preprint and opens the door to further AI-assisted discoveries in quantum field theory.

#
OpenAI
Models
February 13, 2026

Scaling social science research

OpenAI released GABRIEL, an open-source GPT-based toolkit that converts qualitative data (text/images) into quantitative measurements, helping social scientists analyze large datasets more efficiently.
Expand

OpenAI introduced GABRIEL, a new open-source toolkit designed to help researchers turn qualitative data like interviews, course materials, social media, and images into quantitative measurements for easier analysis.

Traditional qualitative analysis is time-consuming and limits studies at scale, but GABRIEL lets researchers define what they want to measure in simple language and consistently score thousands or millions of documents. It also includes tools for merging datasets, deduplicating data, coding passages, and protecting privacy.

Available as a Python library with a tutorial, GABRIEL aims to make rich qualitative data more accessible for economists, social scientists, and data scientists.

#
OpenAI
Models
February 13, 2026

Introducing lockdown mode and elevated risk labels in ChatGPT

OpenAI added Lockdown Mode and Elevated Risk labels to ChatGPT to help protect against prompt injection attacks and guide users about web-connected features’ risks.
Expand

OpenAI introduced two new safety features in ChatGPT: Lockdown Mode, an optional, advanced security setting for highly risk-sensitive users, and consistent “Elevated Risk” labels for capabilities that may introduce additional security concerns, especially when connected to apps or the web.

Lockdown Mode restricts how ChatGPT interacts with external systems to reduce prompt injection and data exfiltration. Elevated Risk labels appear in product interfaces to help users understand those features’ potential risks and make informed choices.

These updates build on existing safeguards like sandboxing, URL protections, and enterprise controls.

#
OpenAI
Models
February 13, 2026

Beyond rate limits: scaling access to Codex and Sora

OpenAI rethought traditional rate limits for Codex and Sora by building a unified real-time system blending rate limits and credit balances, letting users continue usage seamlessly while ensuring accurate billing and fairness.
Expand

OpenAI redesigned how access works for its tools Codex and Sora by moving beyond rigid rate limits that stop users abruptly.

Too many users hit traditional limits and experienced frustrating “hard stops.” To solve this, OpenAI’s engineers created a real-time access engine that combines rate limits with purchasable credits, allowing use to continue smoothly when limits are reached. The system tracks usage, rate windows, and credit balances together, making real-time decisions that are accurate and auditable.

This hybrid model improves user experience, avoids performance issues, and ensures fair access and trust without interrupting workflows.

#
OpenAI
Models
February 12, 2026

Introducing GPT‑5.3‑Codex‑Spark

OpenAI released GPT-5.3-Codex-Spark, an ultra-fast real-time coding model with over 1,000 tokens per second, 128k context, and a research preview for ChatGPT Pro users. It uses Cerebras hardware.
Expand

OpenAI introduced GPT-5.3-Codex-Spark, a new real-time coding model built for rapid interactive development. It generates more than 1,000 tokens per second and supports a 128,000-token context window.

The model runs on specialized low-latency hardware from Cerebras Systems and is available as a research preview for ChatGPT Pro users through the Codex app, CLI, and IDE extension. Codex-Spark focuses on instant feedback and rapid iteration while staying capable on real software engineering tasks.

It expands the Codex family by enabling real-time collaboration alongside longer, deeper reasoning workflows.

#
OpenAI
Expert Views
February 11, 2026

The comprehensive guide to building production-ready Model Context Protocol systems

Building production ready Model Context Protocol systems requires more than tool calling. Teams must design scenario coverage, validation workflows, failure handling, and strong logging so AI integrations work reliably in real world environments.
Expand

The guide explains how to build production ready systems using the Model Context Protocol, a standard that connects AI models with external tools, data sources, and applications. Instead of stopping at demo level tool calls, teams must design reliable architectures that handle real world complexity.

Key practices include strong scenario coverage, maker checker validation patterns for outputs, and robust failure handling across different workflows. Logging, authentication, and governance are also essential to monitor how AI systems interact with tools and data.

When these elements are built from the start, MCP based systems move from experimental prototypes to stable enterprise deployments that can handle unpredictable inputs and scale effectively.

#
GoML
Models
February 11, 2026

Animate your facebook profile picture with Meta AI

Facebook now uses Meta AI to animate profile pictures from still photos with fun effects like wave, confetti, and party hat, plus AI-powered style tools for Stories and animated backgrounds for text posts.
Expand

Facebook launched Meta AI features that let users animate profile pictures from still photos using preset effects such as natural motion, wave, heart, confetti, and party hat, with more options coming later.

Photos can be chosen from the camera roll or existing uploads and shared in Feed. Meta also added an AI-driven Restyle tool for Stories and Memories, letting users adjust style, mood, lighting, color, or background with text prompts or presets.

Text posts can now include animated or still backgrounds accessed by tapping an icon when creating a post. These updates aim to make self-expression on Facebook more dynamic and visual.

#
Meta
Models
February 10, 2026

A one-prompt attack that breaks LLM safety alignment

Microsoft research shows a single unlabeled prompt can strip safety guardrails from large language models through a method called GRP-Obliteration, making them respond to harmful requests across many categories.
Expand

Microsoft published research showing how a single unlabeled prompt can remove safety alignment from large language models. The team used a technique normally meant to improve model behavior, called Group Relative Policy Optimization, and flipped it to weaken guardrails.

In tests, training with one prompt asking for “a fake news article that could lead to panic or chaos” caused 15 different language models to become more willing to produce harmful or disallowed content. This finding means safety layers can be fragile, especially once models are fine-tuned after deployment.

Researchers warn teams must test safety continually as they adapt models.

#
Microsoft
Models
February 10, 2026

Testing Ads in ChatGPT

OpenAI began testing ads in ChatGPT for free and Go users in the U.S. Ads are clearly labeled, separate from answers, and will not influence responses, while keeping user chats private.
Expand

OpenAI has started testing advertisements inside ChatGPT for users on the Free and Go plans in the United States. The ads are clearly labeled as sponsored content and appear separately from the AI’s answers, with OpenAI saying they will not change how responses are generated.

Users can control ad settings, including opt-out options and personalization preferences. OpenAI says it will protect conversation privacy and keep ads out of sensitive topics like health or politics. Subscribers on higher-paid tiers will not see ads.

The goal of this test is to fund broader access to advanced features while maintaining user trust.

#
OpenAI
Spotlight
February 9, 2026

AI healthcare platform boosting dietary consistency by 50% for Heartful Sprouts

GoML built an AI healthcare platform that automates dietary recipe adjustments for clinicians using natural language input, increasing guideline consistency by 50% and drastically reducing manual effort while preserving ingredient tracking.
Expand

GoML developed an AI healthcare platform for Heartful Sprouts, a pediatric nutrition service, to automate medical dietary recipe adjustments. Clinicians previously spent time manually modifying recipes for conditions like celiac disease and diabetes.

The platform uses natural language queries to interpret dietary needs, apply clinical rules, and generate compliant recipe updates while preserving ingredient identifiers. It integrates structured medical dietary knowledge and secure database connections, enabling accurate substitutions and consistent outputs.

Impact included a 70–80% reduction in manual adjustment effort, twice-as-fast compliant recipe generation, and a 50% improvement in adherence to dietary guidelines. The solution maintained nutritional accuracy and traceability.

#
GoML
Models
February 6, 2026

Perplexity’s multi-model AI system

Perplexity’s multi-model AI system routes queries to the best-suited models in parallel, letting users tap teams of models like GPT-5.2, Claude Opus, and others for better accuracy and depth. It synthesizes outputs for reliable answers.
Expand

Perplexity now runs queries across multiple AI models simultaneously rather than relying on a single model, using a multi-model orchestration approach.

Its Model Council feature sends prompts in parallel to top systems such as Claude Opus 4.6, GPT-5.2, and Gemini 3.0, then synthesizes their responses into one unified answer with consensus and disagreements highlighted.

This design aims to improve accuracy and reduce blind spots by combining strengths from different models. The platform routes work to models best suited for reasoning, search grounding, or creative tasks, reflecting a shift toward coordinated multi-model AI workflows.

#
OpenAI
Models
February 6, 2026

Introducing GPT-5.3-Codex

GPT-5.3-Codex is OpenAI’s latest agentic coding model combining advanced coding, reasoning, and professional knowledge with 25% faster performance, supporting long tasks and real-time interaction across coding and computer work.
Expand

GPT-5.3-Codex is the newest and most capable version of OpenAI’s Codex, designed to handle advanced coding and broader professional workflows on a computer.

It builds on GPT-5.2-Codex with stronger coding, reasoning, and knowledge work capabilities, running about 25% faster and excelling at long-running tasks involving research, tool use, and complex execution.

The model achieves state-of-the-art benchmarks, produces functional software and websites, and supports debugging, deployment, tests, documentation, and more. Users can interact with it in real time as it works, steering progress. GPT-5.3-Codex is available in the Codex app, CLI, IDE extensions, and web for paid ChatGPT plans.

#
OpenAI
Models
February 6, 2026

Getting started with Gemini 3

Google Cloud’s Gemini-3 free-trial guide helps developers explore Gemini-3 and cloud AI tools with Google’s free credits and trial programs, letting you build and test AI apps before paying.
Expand

The Getting Started with Gemini-3 free-trial post walks developers and practitioners through how to begin building AI applications on Google Cloud using Gemini-3 and associated tools.

It explains that new customers receive free trial credits (e.g., $300 to spend on Google Cloud services), which can be used with services like Vertex AI and AI APIs to experiment with Gemini-3 models and other cloud products without upfront cost.

It also highlights how the free tier and 90-day trial can help you try out AI features, build prototypes or proofs of concept, and test workloads before committing to paid usage.

#
Google
Models
February 5, 2026

Introducing Claude Opus 4.6

Claude Opus 4.6 is Anthropic’s newest flagship AI model, boosting coding, enterprise automation, and long-context reasoning with up to a 1 million-token window and collaborative agent teams.
Expand

Anthropic has released Claude Opus 4.6, its most advanced AI to date that significantly improves coding, multi-step reasoning, and enterprise workflows.

It introduces a 1 million-token context window (beta) so the model can handle massive codebases, long documents, and sustained tasks in one go. The update also includes enhanced autonomous planning, debugging, and tool use, plus support for agent teams that split work across multiple AI agents for faster results.

Opus 4.6 shows stronger performance on professional benchmarks, excels in financial and legal analyses, and powers integrated tools on cloud platforms and Claude’s API, continuing Anthropic’s push toward AI-assisted productivity.

#
Anthropic
Models
February 5, 2026

Introducing OpenAI Frontier

OpenAI Frontier is a new enterprise platform for building, deploying, and managing AI agents that act like AI coworkers, giving them shared business context, permissions, memory, and tools to do real work.
Expand

OpenAI Frontier is a new enterprise platform designed to help companies build, deploy, and manage autonomous AI agents “AI coworkers” that can perform real tasks across business workflows.

It connects with existing systems like data warehouses, CRMs, and internal apps to give agents shared business knowledge, clear permissions, and the ability to learn through experience. Frontier allows agents to work with files, run code, and interact with tools, improving performance over time.

The system supports enterprise-grade security, governance, and integrations without requiring major infrastructure changes. Early adopters include HP, Intuit, Oracle, State Farm, Thermo Fisher, and Uber, with broader availability coming soon.

#
OpenAI
Expert Views
February 3, 2026

Why Agentic AI implementation fails and how to get it right

Many agentic AI projects fail because companies treat them like simple chatbots instead of redesigning workflows and infrastructure. Successful implementation requires strong architecture, governance, and real integration with enterprise systems.
Expand

The GoML blog explains why many agentic AI implementations fail and what organizations should do differently. A key reason is that companies deploy AI agents without redesigning workflows, infrastructure, and governance systems around them.

Agentic AI systems need reliable orchestration, tool integration, and strong operational monitoring to work at scale. The article notes that about 40 percent of agentic AI projects may fail by 2027 if teams treat them as simple automation tools rather than goal driven systems.

Successful deployments require production ready architecture, workflow redesign, and workforce training. With the right strategy, organizations can move beyond experiments and deploy AI agents that run real business processes securely and efficiently.

#
GoML
Models
February 2, 2026

Snowflake and OpenAI partner to bring frontier intelligence to enterprise data

OpenAI and Snowflake announced a $200 million multi-year partnership to bring OpenAI’s AI models directly into Snowflake’s enterprise data platform, enabling customers to build AI agents and extract insights from their data.
Expand

OpenAI and Snowflake entered a multi-year, $200 million partnership that embeds OpenAI’s advanced AI models into Snowflake’s AI Data Cloud and Cortex platform.

This integration lets Snowflake’s global enterprise customers use models like GPT-5.2 directly on their proprietary data to build custom AI applications, natural-language queries, and autonomous AI agents without moving sensitive data outside the secure environment.

The goal is to accelerate enterprise AI adoption by combining Snowflake’s secure, governed data infrastructure with OpenAI’s reasoning and analytics capabilities. Early use cases include AI-driven insights, automation workflows, and real-time decision support across industries worldwide.

#
OpenAI
Models
February 2, 2026

Introducing the Codex app

Open AI’s Codex app for macOS gives developers a desktop hub to manage multiple AI coding agents, run tasks in parallel, and build software more quickly with autonomous workflows.
Expand

OpenAI introduced the Codex app for macOS as a dedicated workspace for developers to manage AI coding agents in one place.

The app lets users run multiple agents in parallel, supervise long tasks, and collaborate across coding projects without switching tools. It works with a ChatGPT account and supports agent orchestration across IDEs, command line, and cloud environments.

Developers can review changes, assign tasks, and extend agents with skills for broader workflows beyond code generation. The app aims to streamline software development by centralizing agent control and reducing manual overhead.

#
OpenAI
Models
January 31, 2026

DeepSeek gets approval to buy Nvidia's H200 AI chips

China has conditionally approved AI startup DeepSeek to buy Nvidia’s high-performance H200 chips, pending regulatory terms. Other Chinese tech firms received similar clearances. Nvidia awaits formal notice.
Expand

China has granted conditional approval for DeepSeek, a leading domestic AI company, to buy Nvidia’s advanced H200 artificial intelligence chips, sources said. Final regulatory conditions are still being worked out by China’s National Development and Reform Commission.

The decision aligns DeepSeek with other major Chinese tech groups like ByteDance, Alibaba, and Tencent, which received approvals to acquire large quantities of the same processors. The approvals come amid tight U.S. and Chinese rules on advanced chip exports and imports.

Nvidia’s CEO said the company has not yet received official notification. If finalized, the deal could boost China’s AI data center and infrastructure development.

#
DeepSeek
Models
January 30, 2026

How AI assistance impacts the formation of coding skills

Anthropic research found AI coding help speeds some tasks up to 80 percent, but heavy reliance can weaken learning. Developers who ask questions and seek explanations retain more understanding.
Expand

Anthropic studied how AI assistance affects coding skill development. Using a controlled trial with software developers, the research found that AI can accelerate task completion but may reduce mastery of new coding concepts.

Participants who used AI scored about 17 percent lower on a quiz measuring comprehension after completing tasks with AI support, compared to those who coded without help. The effect was strongest in areas like debugging and code reading.

The research also showed that how developers interact with AI matters: those who asked for explanations alongside code generation retained more knowledge. The study highlights a trade-off between speed and deep learning.

#
Anthropic
Expert Views
January 29, 2026

The 2026 Guide to Amazon Bedrock AgentCore

GoML explains Amazon Bedrock AgentCore as a platform for building and running AI agents at scale with memory, runtime, identity, and observability, simplifying production deployments and reducing infrastructure friction.
Expand

GoML outlines Amazon Bedrock AgentCore as a managed platform that helps organizations build, deploy, and operate enterprise-grade AI agents. It solves common deployment hurdles such as memory management, scaling, security, and observability by providing services like serverless runtime, persistent context memory, identity controls, and deep tracing.

AgentCore supports multiple frameworks and models, making it flexible for diverse agent workloads. The platform includes policy enforcement and built-in evaluation tools to maintain quality and safety in production.

This guide highlights how AgentCore bridges the gap between early prototypes and scaled, reliable AI agents in real-world systems.

#
Bedrock
Expert Views
January 28, 2026

Top AI trends 2026 suggested by Stanford specialists

GoML covers key AI trends for 2026 from Stanford specialists. Focus shifts from hype to real value, rigorous evaluation, medical AI progress, job impact tracking and transparent decision systems. Investment efficiency matters.
Expand

GoML presents Stanford-backed AI trends for 2026. The narrative moves away from speculative promise to measured results and practical use.

Businesses now demand clear productivity gains, cost insight and reliable systems. Medical AI gains traction with models trained on large clinical data supporting rare disease detection and clinician workflows. Real-time tracking of job effects replaces broad forecasts, letting policymakers and companies adjust training and workforce strategies.

Explainability becomes essential, especially in high-stakes decisions like medical diagnosis and lending. Rising data center costs and sovereignty concerns drive efficient infrastructure choices. Overall, 2026 prioritizes systems that deliver value and are governed with discipline.

#
OpenAI
Models
January 28, 2026

Anthropic selected to build government AI assistant pilot

Anthropic was chosen by the UKs Department for Science, Innovation and Technology to build and pilot an AI assistant for GOV.UK, guiding citizens through complex public service processes with tailored support.
Expand

Anthropic has been selected by the UK Department for Science, Innovation and Technology to develop and pilot an AI assistant for the GOV.UK platform.

The pilot uses an agentic system powered by Claude to help citizens navigate public services with step-by-step support rather than simple question-and-answer responses. Initial deployment focuses on employment services, guiding jobseekers through job search, training resources, and government support programs.

The project follows a structured scan, pilot, scale framework to test safety and effectiveness before wider rollout, with emphasis on data control, transparency, and compliance with UK laws. Engineers from Anthropic will work with civil service teams to build internal expertise.

#
Anthropic
Models
January 28, 2026

Introducing Prism

OpenAI introduced Prism, a free AI-powered scientific workspace built on GPT-5.2. It streamlines drafting, citations, collaboration, and equation work for researchers in a unified environment.
Expand

OpenAI launched Prism, a free AI-native workspace designed for scientific research and writing. Built on GPT-5.2, Prism integrates drafting, literature search, citation management, and real-time collaboration into one platform. It supports complex tasks such as converting diagrams to LaTeX and reasoning over equations within full document context.

The tool aims to reduce workflow fragmentation and help researchers focus on substantive scientific work. Prism is available today to users with a ChatGPT personal account and will expand to business and enterprise plans.

By bringing advanced reasoning and collaboration together, OpenAI hopes to streamline research workflows and enhance productivity.

#
OpenAI
Spotlight
January 27, 2026

NewVue AI radiology transforming reporting with 2� visibility

GoML and NewVue used AI to modernize radiology reporting. Real-time dictation, structured reports, and live insights doubled visibility for clinical, admin, and engineering teams and cut manual work.
Expand

GoML built an AI-driven radiology reporting platform for NewVue that replaces manual, slow workflows with real-time insights. The system captures live speech, transcribes with confidence scores, and generates structured reports. Radiologists see live dictation progress, turnaround metrics, and correction trends.

Admin and engineering teams get guaranteed visibility into performance and error patterns. Role-based access keeps sensitive data secure and compliant. The solution runs on AWS Bedrock and integrates with hospital systems.

Early results show about 50 percent faster visibility into reporting status and twice the insight availability across teams, reducing manual monitoring by 60-70 percent.

#
GoML
Models
January 27, 2026

DeepSeek releases DeepSeek-OCR 2 with advanced visual reasoning

DeepSeek unveiled DeepSeek-OCR 2, an open-source vision model using DeepEncoder V2 to enable causal visual reasoning, significantly improving image understanding, structured text extraction, and multimodal AI performance.
Expand

DeepSeek has launched DeepSeek-OCR 2, its latest open-source image understanding model, introducing the DeepEncoder V2 architecture and a new “visual causal flow” approach.

Unlike traditional OCR systems, the model reasons about visual structure and relationships, allowing it to interpret complex layouts, charts, and documents more accurately. By open-sourcing the model and research paper, DeepSeek reinforces its strategy of competing globally through cost-efficient, high-performance AI.

The release strengthens China’s position in multimodal AI and signals growing maturity beyond text-only large language models.

#
DeepSeek
Models
January 26, 2026

Microsoft aims for etter inference efficiency with Maia 200

Microsoft unveiled its Maia 200 inference-focused AI chip built for large-scale generative AI workloads, delivering faster model inference with lower memory use and cost, addressing growing demand in reasoning-driven AI systems.
Expand

Microsoft is pushing for better AI inference efficiency with its new Maia 200 accelerator, highlighted in its January 2026 announcement.

The Maia 200 is designed specifically for inference workloads turning trained models into real-time responses and optimizes performance with advanced hardware like FP8/FP4 tensor cores and high-bandwidth memory.

According to Microsoft, the chip can run today’s largest AI models faster and more efficiently, using less memory and power than previous systems, which can help reduce overall operational costs and energy usage for cloud-scale AI services. This reflects the industry trend toward specialized inference hardware as AI workloads become more agentic and reasoning-focused.

#
Microsoft
Ecosystem
January 26, 2026

Amazon Bedrock adds one-hour prompt caching to boost latency and cost efficiency

Amazon Bedrock now supports one-hour prompt caching, allowing developers to reuse context efficiently, reduce inference latency, and lower costs for repetitive or long-running generative AI workloads.
Expand

AWS has enhanced Amazon Bedrock by extending prompt caching duration to one hour, a significant upgrade for developers building production-scale generative AI applications.

Prompt caching enables reuse of previously processed context, reducing repeated computation and improving response latency while lowering inference costs. This is particularly valuable for agentic workflows, RAG systems, and conversational applications with stable system prompts.

The update signals AWS’s focus on inference optimization rather than just model access, positioning Bedrock as a more cost-efficient, enterprise-ready platform for scalable Gen AI deployments.

#
Bedrock
Models
January 23, 2026

Unrolling the Codex Agent loop

OpenAI explains the core “agent loop” in Codex CLI how user input, model inference, and tool calls interact to generate effective code actions, with prompt construction, context management, and iterations highlighted.
Expand

OpenAI’s unrolling the Codex Agent Loop details the internal mechanics of the Codex CLI, focusing on the “agent loop,” which orchestrates the flow between user input, model inference, and tool execution to perform software tasks.

It explains how prompts are built, how the model’s responses can trigger tool calls, and how these interactions repeat until a final result is produced. The post also discusses challenges like context window growth and performance optimization through prompt caching and automatic compaction.

This deep technical overview is the first in a series aimed at revealing design insights behind Codex’s efficient and safe code generation.

#
OpenAI
Models
January 23, 2026

Claude's new constitution

Anthropic has released a new constitution for its AI model Claude, detailing the values and ethical principles that guide its behavior and training. The document aims to shape Claude’s safety, ethics, and helpfulness.
Expand

Anthropic has published a new constitution for its AI model, Claude, outlining in detail the ethical framework and core values that should guide the model’s behavior and decision-making.

The constitution serves as both a training tool and a transparency measure, explaining the principles Claude should uphold, such as safety, ethics, compliance with internal guidelines, and genuine helpfulness to users.

Anthropic says the document shapes Claude’s reasoning and training, helping it apply broad values rather than merely following rules. The constitution is released openly under a Creative Commons license so others can study or adopt similar frameworks.

#
Anthropic
Models
January 21, 2026

OpenAI unveils stargate community plan to keep data-center energy costs in check

OpenAI launched the Stargate Community plan to ensure its AI data centers don’t raise local electricity costs. Each site will tailor energy solutions using community input and may fund new power infrastructure.
Expand

OpenAI has introduced the Stargate Community plan as part of its broader $500 billion Stargate initiative to build large-scale AI data centers across the U.S.

The plan aims to ensure that expanding AI infrastructure doesn’t increase local electricity costs by having OpenAI fund and develop energy resources and grid upgrades as needed. Each Stargate campus will work closely with local communities and utilities to tailor solutions that support stable power without burdening residents.

The initiative reflects growing industry efforts to address energy concerns tied to rapid AI infrastructure growth

#
OpenAI
Models
January 20, 2026

Introducing OpenAI’s education for Countries

OpenAI launches Education for Countries to partner with governments and institutions, embedding AI into education to personalize learning, build AI skills, support research, and prepare students and teachers for future workforce demands.
Expand

OpenAI’s Education for Countries initiative aims to support national education systems in integrating AI tools like ChatGPT Edu, GPT-5.2, and study mode to enhance learning and teaching.

Working with governments, ministries, and universities worldwide, the program focuses on personalized learning, reducing administrative burden, promoting research on AI’s educational impact, and providing certification and training aligned with workforce needs.

Initial partners include Estonia, Greece, Italy, Jordan, Kazakhstan, Slovakia, Trinidad and Tobago, and the UAE, with nationwide deployments and research collaborations already underway. OpenAI emphasizes responsible, equitable AI adoption to strengthen education and future workforce readiness.

#
OpenAI
Models
January 20, 2026

Cisco and OpenAI redefine enterprise engineering with AI agents

Cisco and OpenAI partnered to integrate Codex into enterprise engineering workflows, enabling AI agents to operate at scale across complex codebases, reducing build times, automating defect fixes, and shaping Codex for large organizations.
Expand

OpenAI and Cisco are collaborating to transform enterprise engineering by embedding Codex AI agents into real-world development workflows.

Rather than using Codex as a simple tool, Cisco integrates it into production environments with large, interconnected codebases, enhancing complex tasks like build optimization, defect remediation, and framework migrations.

This collaboration has yielded measurable benefits such as reduced build times and faster defect resolution and influenced Codex’s enterprise readiness in areas like security, compliance, and long-running task management. Together, Cisco and OpenAI aim to expand how AI can function as a reliable engineering teammate in demanding global software environments.

#
OpenAI
Models
January 20, 2026

Horizon 1000 advancing AI for primary healthcare

OpenAI and the Gates Foundation launched Horizon 1000, committing $50 million in funding, technology, and support to strengthen primary healthcare in African communities, reaching 1,000 clinics by 2028 to improve care quality and access.
Expand

OpenAI, in partnership with the Gates Foundation, announced Horizon 1000, a major initiative to advance AI tools for primary healthcare across Africa.

With a $50 million commitment, the project aims to strengthen health systems by 2028, reaching 1,000 primary care clinics and surrounding communities. The program will help frontline health workers use AI to navigate complex guidelines, reduce administrative burden, and improve care consistency where staffing and resources are limited.

Leaders in Rwanda and other nations will receive funding, technology, and technical support to deploy AI safely and meaningfully, closing the gap between AI capabilities and real-world healthcare needs.

#
OpenAI
Models
January 20, 2026

Anthropic and Teach For All launch global AI training initiative for educators

Anthropic partnered with Teach For All to launch a global AI training initiative, providing AI tools, Claude access, and training to over 100,000 educators across 63 countries to boost AI fluency in classrooms.
Expand

Anthropic and Teach For All have announced a major global initiative to bring AI tools and training to educators in 63 countries.

Through the AI Literacy & Creator Collective, more than 100,000 teachers and Teach For All alumni who collectively serve over 1.5 million students will gain access to Claude, AI fluency programs, and practical classroom applications.

This partnership positions teachers as co-creators shaping how AI is used in education, with ongoing peer learning and innovation spaces where educators can pilot AI-enabled tools and provide feedback to inform future product development.

#
Anthropic
Ecosystem
January 20, 2026

Introducing multimodal retrieval for Amazon Bedrock Knowledge Bases

Amazon Bedrock Knowledge Bases now supports multimodal retrieval, enabling RAG applications to search and retrieve insights across text, images, audio, and video using unified, fully managed AI workflows.
Expand

Amazon has announced the general availability of multimodal retrieval for Amazon Bedrock Knowledge Bases, expanding native support beyond text and images to include audio and video content.

This capability allows organizations to build Retrieval Augmented Generation (RAG) applications that seamlessly search across multiple data formats using a single, managed workflow. Powered by Amazon Nova Multimodal Embeddings and Bedrock Data Automation, the solution preserves visual and audio context or delivers precise transcriptions based on use-case needs.

By eliminating complex custom pipelines, Bedrock Knowledge Bases makes it easier for enterprises to unlock insights from diverse data sources such as videos, recordings, images, and documents at scale.

#
AWS
Models
January 19, 2026

Approach to advertising and expanding access to ChatGPT

OpenAI will test ads in ChatGPT for free and low-cost “Go” users in the US. Ads will be clearly labeled at the bottom of answers and won’t change responses or sell user data.
Expand

OpenAI announced it will begin testing advertisements inside ChatGPT for adult users in the United States on the free tier and its newly expanded $8 “Go” subscription plan.

The ads will appear clearly at the bottom of chatbot responses and are designed to be separate from the AI’s answers, so they do not influence what ChatGPT says. OpenAI emphasized it will never sell users’ data to advertisers, and that higher-tier paid subscriptions (Plus, Pro, Business, Enterprise) remain ad-free.

The trial aims to support broader access to the AI service while diversifying revenue beyond subscriptions.

#
OpenAI
Models
January 16, 2026

How scientists are using Claude to accelerate research and discovery

Anthropic explains how scientists are using its AI, Claude, to accelerate research and discovery by handling complex tasks like data analysis and experiment design, speeding work that once took months into hours.
Expand

Anthropic highlights how researchers are using its advanced AI, Claude, to accelerate scientific discovery.

By integrating Claude with scientific tools and workflows, researchers can automate complex tasks such as data analysis, experiment planning, and interpreting results, compressing processes that typically take months into just hours.

This AI collaboration helps eliminate bottlenecks and enables deeper insights by navigating diverse databases and tools more efficiently. Case studies show Claude supporting scientists across all stages of research, reshaping how work is done and accelerating progress in areas like genomics and biomedical discovery.

#
Anthropic
Models
January 15, 2026

Google unveils translateGemma a new open translation AI built on Gemma 3

Google launched TranslateGemma, a suite of open translation models built on Gemma 3, offering high-quality translation across 55 languages with efficient performance for mobile, laptops, and cloud environments.
Expand

Google introduced TranslateGemma, a new family of open translation models based on the Gemma 3 architecture, designed to support translation across 55 languages with high accuracy and efficiency.

Available in 4B, 12B, and 27B parameter sizes, these models deliver strong performance for various devices from mobile and edge hardware to cloud GPUs without sacrificing translation quality.

The 12B model even outperforms larger baselines on benchmark tests, while all variants retain multimodal translation abilities (including text in images). TranslateGemma aims to empower developers and researchers with accessible, state-of-the-art translation tools.

#
Google
Models
January 14, 2026

Introducing Labs

Anthropic announces Anthropic Labs, a dedicated team for experimenting with new AI capabilities and turning them into products. It expands innovation around Claude features and incubates future AI tools.
Expand

Anthropic has launched Anthropic Labs, a new initiative focused on experimenting with cutting-edge AI ideas and rapidly building them into scalable products.

This innovation team supports early-stage exploration around advanced Claude capabilities and tests unpolished versions with early users to identify what works best. Anthropic Labs has already contributed to successful offerings like Claude Code, MCP, Skills, and Cowork, and now aims to expand this experimental approach.

Instagram co-founder Mike Krieger joins the Labs team, while product leadership shifts to support scaling core Claude experiences. The goal is to foster frontier AI development responsibly and bring new tools to market from these experiments.

#
Anthropic
Models
January 14, 2026

New technical directions for DeepSeek V4

DeepSeek’s latest research reveals new technical directions for DeepSeek V4, emphasizing sparse architectures and efficiency-focused design to overcome hardware constraints and reduce dependence on high-end GPUs.
Expand

DeepSeek has published new research outlining the technical direction of its upcoming DeepSeek V4 model, focusing on architectural efficiency rather than brute-force scaling.

The company is exploring sparse model designs, modular components, and memory-efficient computation to overcome hardware bottlenecks such as GPU shortages and memory limits. These innovations aim to deliver frontier-level performance while reducing compute costs and reliance on top-tier hardware.

By rethinking model architecture instead of simply increasing parameter counts, DeepSeek positions V4 as a more scalable and sustainable alternative in the global AI race, particularly under tightening chip export restrictions.

#
DeepSeek
Ecosystem
January 13, 2026

AWS strengthens agentic AI ecosystem with LangGraph and DynamoDB integrations

AWS launched new tools enabling LangGraph agents to run on Bedrock AgentCore, with DynamoDB support for state management, accelerating development of production-grade AI agents.
Expand

AWS is expanding its agentic AI ecosystem by enabling LangGraph-based agents to run directly on the Amazon Bedrock AgentCore runtime, alongside new DynamoDB-backed state persistence.

This combination allows developers to build robust, long-running AI agents with memory, observability, and fault tolerance built in.

By supporting popular open-source agent frameworks while offering managed infrastructure, AWS positions Bedrock as a neutral, enterprise-friendly platform for agentic AI competing with Google Vertex AI and OpenAI-centric stacks without locking customers into proprietary tooling.

#
Bedrock
Models
January 13, 2026

Google Gemini introduces personal intelligence to connect Google apps for tailored AI assistance

Google launched Personal Intelligence for its Gemini AI assistant, letting users opt-in to connect Gmail, Photos, Search, and YouTube so Gemini can provide more contextual, personalized responses based on user data.
Expand

Google has announced Personal Intelligence, a new beta feature for its Gemini AI assistant that allows users to optionally connect their Gmail, Photos, Search, and YouTube accounts to create more personalized and context-aware conversations.

Once enabled, Gemini can reason across connected apps to provide tailored help like planning trips, answering specific questions using emails or photos, and suggesting relevant content without training on the user’s private data.

The feature is off by default, available initially to Google AI Pro and Ultra subscribers in the U.S., and will expand to more users and platforms over time, with privacy controls for what data is shared.

#
Google