Latest AI News & Updates

#ai

The next big trend in AI providers appears to be "studio" environments on the web that allow users to spin up agents and AI applications within minutes. Case in point, today the well-funded French AI startup Mistral launched its own Mistral AI Studio, a new production platform designed to help enterprises build, observe, and operationalize AI applications at scale atop Mistral's growing family of proprietary and open source large language models (LLMs) and multimodal models.It's an evolution of its legacy API and AI building platorm, "Le Platforme," initially launched in late 2023, and that brand name is being retired for now. The move comes just days after U.S. rival Google updated its AI Studio, also launched in late 2023, to be easier for non-developers to use and build and deploy apps with natural language, aka "vibe coding."But while Google's update appears to target novices who want to tinker around, Mistral appears more fully focused on building an easy-to-use enterprise AI app development and launchpad, which may require some technical knowledge or familiarity with LLMs, but far less than that of a seasoned developer. In other words, those outside the tech team at your enterprise could potentially use this to build and test simple apps, tools, and workflows — all powered by E.U.-native AI models operating on E.U.-based infrastructure. That may be a welcome change for companies concerned about the political situation in the U.S., or who have large operations in Europe and prefer to give their business to homegrown alternatives to U.S. and Chinese tech giants.In addition, Mistral AI Studio appears to offer an easier way for users to customize and fine-tune AI models for use at specific tasks.Branded as “The Production AI Platform,” Mistral's AI Studio extends its internal infrastructure, bringing enterprise-grade observability, orchestration, and governance to teams running AI in production.The platform unifies tools for building, evaluating, and deploying AI systems, while giving enterprises flexible control over where and how their models run — in the cloud, on-premise, or self-hosted. Mistral says AI Studio brings the same production discipline that supports its own large-scale systems to external customers, closing the gap between AI prototyping and reliable deployment. It's available here with developer documentation here.Extensive Model CatalogAI Studio’s model selector reveals one of the platform’s strongest features: a comprehensive and versioned catalog of Mistral models spanning open-weight, code, multimodal, and transcription domains.Available models include the following, though note that even for the open source ones, users will still be running a Mistral-based inference and paying Mistral for access through its API.ModelLicense TypeNotes / SourceMistral LargeProprietaryMistral’s top-tier closed-weight commercial model (available via API and AI Studio only).Mistral MediumProprietaryMid-range performance, offered via hosted API; no public weights released.Mistral SmallProprietaryLightweight API model; no open weights.Mistral TinyProprietaryCompact hosted model optimized for latency; closed-weight.Open Mistral 7BOpenFully open-weight model (Apache 2.0 license), downloadable on Hugging Face.Open Mixtral 8×7BOpenReleased under Apache 2.0; mixture-of-experts architecture.Open Mixtral 8×22BOpenLarger open-weight MoE model; Apache 2.0 license.Magistral MediumProprietaryNot publicly released; appears only in AI Studio catalog.Magistral SmallProprietarySame; internal or enterprise-only release.Devstral MediumProprietary / LegacyOlder internal development models, no open weights.Devstral SmallProprietary / LegacySame; used for internal evaluation.Ministral 8BOpenOpen-weight model available under Apache 2.0; basis for Mistral Moderation model.Pixtral 12BProprietaryMultimodal (text-image) model; closed-weight, API-only.Pixtral LargeProprietaryLarger multimodal variant; closed-weight.Voxtral SmallProprietarySpeech-to-text/audio model; closed-weight.Voxtral MiniProprietaryLightweight version; closed-weight.Voxtral Mini Transcribe 2507ProprietarySpecialized transcription model; API-only.Codestral 2501OpenOpen-weight code-generation model (Apache 2.0 license, available on Hugging Face).Mistral OCR 2503ProprietaryDocument-text extraction model; closed-weight.This extensive model lineup confirms that AI Studio is both model-rich and model-agnostic, allowing enterprises to test and deploy different configurations according to task complexity, cost targets, or compute environments.Bridging the Prototype-to-Production DivideMistral’s release highlights a common problem in enterprise AI adoption: while organizations are building more prototypes than ever before, few transition into dependable, observable systems. Many teams lack the infrastructure to track model versions, explain regressions, or ensure compliance as models evolve.AI Studio aims to solve that. The platform provides what Mistral calls the “production fabric” for AI — a unified environment that connects creation, observability, and governance into a single operational loop. Its architecture is organized around three core pillars: Observability, Agent Runtime, and AI Registry.1. ObservabilityAI Studio’s Observability layer provides transparency into AI system behavior. Teams can filter and inspect traffic through the Explorer, identify regressions, and build datasets directly from real-world usage. Judges let teams define evaluation logic and score outputs at scale, while Campaigns and Datasets automatically transform production interactions into curated evaluation sets.Metrics and dashboards quantify performance improvements, while lineage tracking connects model outcomes to the exact prompt and dataset versions that produced them. Mistral describes Observability as a way to move AI improvement from intuition to measurement.2. Agent Runtime and RAG supportThe Agent Runtime serves as the execution backbone of AI Studio. Each agent — whether it’s handling a single task or orchestrating a complex multi-step business process — runs within a stateful, fault-tolerant runtime built on Temporal. This architecture ensures reproducibility across long-running or retry-prone tasks and automatically captures execution graphs for auditing and sharing.Every run emits telemetry and evaluation data that feed directly into the Observability layer. The runtime supports hybrid, dedicated, and self-hosted deployments, allowing enterprises to run AI close to their existing systems while maintaining durability and control.While Mistral's blog post doesn’t explicitly reference retrieval-augmented generation (RAG), Mistral AI Studio clearly supports it under the hood. Screenshots of the interface show built-in workflows such as RAGWorkflow, RetrievalWorkflow, and IngestionWorkflow, revealing that document ingestion, retrieval, and augmentation are first-class capabilities within the Agent Runtime system. These components allow enterprises to pair Mistral’s language models with their own proprietary or internal data sources, enabling contextualized responses grounded in up-to-date information. By integrating RAG directly into its orchestration and observability stack—but leaving it out of marketing language—Mistral signals that it views retrieval not as a buzzword but as a production primitive: measurable, governed, and auditable like any other AI process.3. AI RegistryThe AI Registry is the system of record for all AI assets — models, datasets, judges, tools, and workflows. It manages lineage, access control, and versioning, enforcing promotion gates and audit trails before deployments.Integrated directly with the Runtime and Observability layers, the Registry provides a unified governance view so teams can trace any output back to its source components.Interface and User ExperienceThe screenshots of Mistral AI Studio show a clean, developer-oriented interface organized around a left-hand navigation bar and a central Playground environment.The Home dashboard features three core action areas — Create, Observe, and Improve — guiding users through model building, monitoring, and fine-tuning workflows.Under Create, users can open the Playground to test prompts or build agents.Observe and Improve link to observability and evaluation modules, some labeled “coming soon,” suggesting staged rollout.The left navigation also includes quick access to API Keys, Batches, Evaluate, Fine-tune, Files, and Documentation, positioning Studio as a full workspace for both development and operations.Inside the Playground, users can select a model, customize parameters such as temperature and max tokens, and enable integrated tools that extend model capabilities.Users can try the Playground for free, but will need to sign up with their phone number to receive an access code.Integrated Tools and CapabilitiesMistral AI Studio includes a growing suite of built-in tools that can be toggled for any session:Code Interpreter — lets the model execute Python code directly within the environment, useful for data analysis, chart generation, or computational reasoning tasks.Image Generation — enables the model to generate images based on user prompts.Web Search — allows real-time information retrieval from the web to supplement model responses.Premium News — provides access to verified news sources via integrated provider partnerships, offering fact-checked context for information retrieval.These tools can be combined with Mistral’s function calling capabilities, letting models call APIs or external functions defined by developers. This means a single agent could, for example, search the web, retrieve verified financial data, run calculations in Python, and generate a chart — all within the same workflow.Beyond Text: Multimodal and Programmatic AIWith the inclusion of Code Interpreter and Image Generation, Mistral AI Studio moves beyond traditional text-based LLM workflows. Developers can use the platform to create agents that write and execute code, analyze uploaded files, or generate visual content — all directly within the same conversational environment.The Web Search and Premium News integrations also extend the model’s reach beyond static data, enabling real-time information retrieval with verified sources. This combination positions AI Studio not just as a playground for experimentation but as a full-stack environment for production AI systems capable of reasoning, coding, and multimodal output.Deployment FlexibilityMistral supports four main deployment models for AI Studio users:Hosted Access via AI Studio — pay-as-you-go APIs for Mistral’s latest models, managed through Studio workspaces.Third-Party Cloud Integration — availability through major cloud providers.Self-Deployment — open-weight models can be deployed on private infrastructure under the Apache 2.0 license, using frameworks such as TensorRT-LLM, vLLM, llama.cpp, or Ollama.Enterprise-Supported Self-Deployment — adds official support for both open and proprietary models, including security and compliance configuration assistance.These options allow enterprises to balance operational control with convenience, running AI wherever their data and governance requirements demand.Safety, Guardrailing, and ModerationAI Studio builds safety features directly into its stack. Enterprises can apply guardrails and moderation filters at both the model and API levels.The Mistral Moderation model, based on Ministral 8B (24.10), classifies text across policy categories such as sexual content, hate and discrimination, violence, self-harm, and PII. A separate system prompt guardrail can be activated to enforce responsible AI behavior, instructing models to “assist with care, respect, and truth” while avoiding harmful or unethical content.Developers can also employ self-reflection prompts, a technique where the model itself classifies outputs against enterprise-defined safety categories like physical harm or fraud. This layered approach gives organizations flexibility in enforcing safety policies while retaining creative or operational control.From Experimentation to Dependable OperationsMistral positions AI Studio as the next phase in enterprise AI maturity. As large language models become more capable and accessible, the company argues, the differentiator will no longer be model performance but the ability to operate AI reliably, safely, and measurably.AI Studio is designed to support that shift. By integrating evaluation, telemetry, version control, and governance into one workspace, it enables teams to manage AI with the same discipline as modern software systems — tracking every change, measuring every improvement, and maintaining full ownership of data and outcomes.In the company’s words, “This is how AI moves from experimentation to dependable operations — secure, observable, and under your control.”Mistral AI Studio is available starting October 24, 2025, as part of a private beta program. Enterprises can sign up on Mistral’s website to access the platform, explore its model catalog, and test observability, runtime, and governance features before general release.

#ai

China’s Ant Group, an affiliate of Alibaba, detailed technical information around its new model, Ring-1T, which the company said is “the first open-source reasoning model with one trillion total parameters.”Ring-1T aims to compete with other reasoning models like GPT-5 and the o-series from OpenAI, as well as Google’s Gemini 2.5. With the new release of the latest model, Ant extends the geopolitical debate over who will dominate the AI race: China or the US. Ant Group said Ring-1T is optimized for mathematical and logical problems, code generation and scientific problem-solving. “With approximately 50 billion activated parameters per token, Ring-1T achieves state-of-the-art performance across multiple challenging benchmarks — despite relying solely on natural language reasoning capabilities,” Ant said in a paper.Ring-1T, which was first released on preview in September, adopts the same architecture as Ling 2.0 and trained on the Ling-1T-base model the company released earlier this month. Ant said this allows the model to support up to 128,000 tokens.To train a model as large as Ring-1T, researchers had to develop new methods to scale reinforcement learning (RL).New methods of training
Ant Group developed three “interconnected innovations” to support the RL and training of Ring-1T, a challenge given the model's size and the typically large compute requirements it entails. These three are IcePop, C3PO++ and ASystem.IcePop removes noisy gradient updates to stabilize training without slowing inference. It helps eliminate catastrophic training-inference misalignment in RL. The researchers noted that when training models, particularly those using a mixture-of-experts (MoE) architecture like Ring-1T, there can often be a discrepancy in probability calculations. “This problem is particularly pronounced in the training of MoE models with RL due to the inherent usage of the dynamic routing mechanism. Additionally, in long CoT settings, these discrepancies can gradually accumulate across iterations and become further amplified,” the researchers said. IcePop “suppresses unstable training updates through double-sided masking calibration.”The next new method the researchers had to develop is C3PO++, an improved version of the C3PO system that Ant previously established. The method manages how Ring-1T and other extra-large parameter models generate and process training examples, or what they call rollouts, so GPUs don’t sit idle. The way it works would break work in rollouts into pieces to process in parallel. One group is the inference pool, which generates new data, and the other is the training pool, which collects results to update the model. C3PO++ creates a token budget to control how much data is processed, ensuring GPUs are used efficiently.The last new method, ASystem, adopts a SingleController+SPMD (Single Program, Multiple Data) architecture to enable asynchronous operations.  Benchmark resultsAnt pointed Ring-1T to benchmarks measuring performance in mathematics, coding, logical reasoning and general tasks. They tested it against models such as DeepSeek-V3.1-Terminus-Thinking, Qwen-35B-A22B-Thinking-2507, Gemini 2.5 Pro and GPT-5 Thinking. In benchmark testing, Ring-1T performed strongly, coming in second to OpenAI’s GPT-5 across most benchmarks. Ant said that Ring-1T showed the best performance among all the open-weight models it tested. The model posted a 93.4% score on the AIME 25 leaderboard, second only to GPT-5. In coding, Ring-1T outperformed both DeepSeek and Qwen.“It indicates that our carefully synthesized dataset shapes Ring-1T’s robust performance on programming applications, which forms a strong foundation for future endeavors on agentic applications,” the company said. Ring-1T shows how much Chinese companies are investing in models Ring-1T is just the latest model from China aiming to dethrone GPT-5 and Gemini. Chinese companies have been releasing impressive models at a quick pace since the surprise launch of DeepSeek in January. Ant's parent company, Alibaba, recently released Qwen3-Omni, a multimodal model that natively unifies text, image, audio and video. DeepSeek has also continued to improve its models and earlier this month, launched DeepSeek-OCR. This new model reimagines how models process information. With Ring-1T and Ant’s development of new methods to train and scale extra-large models, the battle for AI dominance between the US and China continues to heat up.   

New experimental AI tool helps people explore the context and origin of images seen online.

The International Mathematical Olympiad (“IMO”) is the world’s most prestigious competition for young mathematicians, and has been held annually since 1959. Each country taking part is represented by six elite, pre-university mathematicians who compete to solve six exceptionally difficult problems in algebra, combinatorics, geometry, and number theory.

Introducing the first model for contextualizing ancient inscriptions, designed to help historians better interpret, attribute and restore fragmentary texts.

Genie 3 can generate dynamic worlds that you can navigate in real time at 24 frames per second, retaining consistency for a few minutes at a resolution of 720p.

TAAFT shop live, Claude memory, the 2026 AI Playbook, Norton Neo, and more...

Our new Perch model helps conservationists analyze audio faster to protect endangered species, from Hawaiian honeycreepers to coral reefs.

Using AI to perceive the universe in greater depth

Gemini 2.5 Deep Think achieves breakthrough performance at the world’s most prestigious computer programming competition, demonstrating a profound leap in abstract problem solving.

Our new method could help mathematicians leverage AI techniques to tackle long-standing challenges in mathematics, physics and engineering.

#school of engineering #dmse #nuclear science and engineering #computer science and technology #artificial intelligence #energy #energy efficiency #sustainability #research #science communications #community #students #graduate, postdoctoral #profile

PhD student Miranda Schwacke explores how computing inspired by the human brain can fuel energy-efficient artificial intelligence.

We’re strengthening the Frontier Safety Framework (FSF) to help identify and mitigate severe risks from advanced AI models.

We’re powering an era of physical agents — enabling robots to perceive, plan, think, use tools and act to better solve complex, multi-step tasks.

Using advanced AI to fix critical software vulnerabilities

#ai

A new framework developed by researchers at Google Cloud and DeepMind aims to address one of the key challenges of developing computer use agents (CUAs): Gathering high-quality training examples at scale.The framework, dubbed Watch & Learn (W&L), addresses the problem of training data generation in a way that doesn’t require human annotation and can automatically extract demonstrations from raw videos.Their experiments show that data generated W&L can be used to train or fine-tune existing computer use and foundation models to improve their performance on computer-use tasks. But equally important, the same approach can be used to create in-context learning (ICL) examples for computer use agents, enabling companies to create CUAs for bespoke internal tasks without the need for costly training of specialized models.The data bottleneck of CUAThe web is rich with video tutorials and screencasts that describe complex workflows for using applications. These videos are a gold mine that can provide computer use agents with domain knowledge and instructions for accomplishing different tasks through user interface interactions.However, before they can be used to train CUA agents, these videos need to be transformed into annotated trajectories (that is, a set of task descriptions, screenshots and actions), a process that is prohibitively expensive and time-consuming when done manually.Existing approaches to address this data bottleneck rely on annotating these videos through the use of multimodal language models, which usually result in low precision and faulty examples. A different approach uses self-play agents that autonomously explore user interfaces to collect trajectories. However, techniques using this approach usually create simple examples that are not useful in unpredictable real-world situations.As the researchers note in their paper, “Overall, these approaches either rely on brittle heuristics, are costly as they rely on explorations in real environments or generate low-complexity demonstrations misaligned with human intent.”Watch & LearnThe Watch & Learn framework tries to address the challenges of creating CUA demonstrations by rethinking the problem formulation.Instead of directly generating trajectories or depending on complex multi-stage pipelines, the researchers frame the problem as an “inverse dynamics objective”: Given two consecutive observations, predict the intermediate action that produced the transition.According to the researchers, this formulation is “easier to learn, avoids hand-crafted heuristics and generalizes robustly across applications.”The W&L framework can be broken down into three key stages: Training an inverse dynamics model (IDM), retrieving raw videos, and training CUA agents.In the first phase, the researchers used agents to interact with live web pages to create a large corpus of 500,000 state transitions (two consecutive observations and the action that resulted in the transition). They then used this data (along with 132,000 human-annotated transitions from existing open datasets) to train an inverse dynamics model (IDM) that takes in two consecutive observations and predicts the transition action. Their trained IDM, which is a small transformer model, outperformed off-the-shelf foundation models in predicting transition actions.The researchers then designed a pipeline that retrieves videos from platforms such as YouTube and runs them through IDM to generate high-quality trajectories. The IDM takes in consecutive video frames and determines the actions (scroll, click) that caused the changes in the environment, which are then packaged into annotated trajectories. Using this method, they generated 53,125 trajectories with high-accuracy action labels.These examples can be used to train effective computer use models for specific tasks. But the researchers also found that trajectories extracted through IDM can serve as in-context learning examples to improve the performance of CUAs on bespoke tasks at inference time. For ICL, they use Gemini 2.5 Flash to add additional reasoning annotations to the observation/action examples in the trajectories, which can then be inserted into the CUA agent’s prompt (usually 3-5 examples) during inference.“This dual role (training and in-context guidance) enables flexible integration with both open-source models and general-purpose agents,” the researchers write.W&L in actionTo test the usefulness of W&L, the researchers ran a series of experiments with closed and open source models on the OSWorld benchmark, which evaluates agents in real desktop and operating system environments across different tasks, including productivity, programming and design.For fine-tuning, they used their corpus of 53,000 trajectories to train two open source models: UI-TARS-1.5, a strong, open source vision-language-action model designed specifically for computer use, and Qwen 2.5-VL, an open-weight multimodal LLM. For in-context learning tests, they applied W&L examples to general-purpose multimodal models such as Gemini 2.5 Flash, OpenAI o3 and Claude Sonnet 4. W&L resulted in improvements on OSWorld in all model categories, including up to 3 points for ICL on general-purpose models and up to 11 points for fine-tuned open-source models.More importantly, these benefits were achieved without any manual annotation, “demonstrating that web-scale human workflows can serve as a practical and scalable foundation for advancing CUAs towards real-world deployment,” the researchers write.This could have important implications for real-world applications, enabling enterprises to turn their existing corpora of videos and conference recordings into training data for CUAs. It also makes it easier to generate new training trajectories. All you will need to do is record videos of performing different tasks and have them annotated by an IDM. And with frontier models constantly improving and becoming cheaper, you can expect to get more from your existing data and the field continues to progress.

#ai

Is the Google Search for internal enterprise knowledge finally here...but from OpenAI? It certainly seems that way. Today, OpenAI has launched company knowledge in ChatGPT, a major new capability for subscribers to ChatGPT's paid Business, Enterprise, and Edu plans that lets them call up their company's data directly from third-party workplace apps including Slack, SharePoint, Google Drive, Gmail, GitHub, HubSpot and combine it in ChatGPT outputs to them. As OpenAI's CEO of Applications Fidji Simo put it in a post on the social network X: "it brings all the context from your apps (Slack, Google Drive, GitHub, etc) together in ChatGPT so you can get answers that are specific to your business."Intriguingly, OpenAI's blog post on the feature states that is "powered by a version of GPT‑5 that’s trained to look across multiple sources to give more comprehensive and accurate answers," which sounds to me like a new fine-tuned version of the model family the company released back in August, though there are no additional details on how it was trained or its size, techniques, etc.OpenAI tells VentureBeat it's a version of GPT-5 that specifically powers company knowledge in ChatGPT Business, Enterprise, and Edu. Nonetheless, company knowledge in ChatGPT is rolling out globally and is designed to make ChatGPT a central point of access for verified organizational information, supported by secure integrations and enterprise-grade compliance controls, and give employees way faster access to their company's information while working.Now, instead of toggling over to Slack to find the assignment you were given and instructions, or tabbing over to Google Drive and opening up specific files to find the names and numbers you need to call, ChatGPT can deliver all that type of information directly into your chat session — if your company enables the proper connections.As OpenAI Chief Operating Officer Brad Lightcap wrote in a post on the social network X: "company knowledge has changed how i use chatgpt at work more than anything we have built so far - let us know what you think!"It builds upon the third-party app connectors unveiled back in August 2025, though those were only for individual users on the ChatGPT Plus plans.Connecting ChatGPT to Workplace SystemsEnterprise teams often face the challenge of fragmented data across various internal tools—email, chat, file storage, project management, and customer platforms. Company knowledge bridges those silos by enabling ChatGPT to connect to approved systems like, and other supported apps through enterprise-managed connectors.Each response generated with company knowledge includes citations and direct links to the original sources, allowing teams to verify where specific details originated. This transparency helps organizations maintain data trustworthiness while increasing productivity.The sidebar shows a live view of the sources being examined and what it is getting from them. When it’s done, you’ll see exactly the sources used, along with the specific snippets it drew from. You can then click on any citation to open the original source for more details.Built for Enterprise Control and SecurityCompany knowledge was designed from the ground up for enterprise governance and compliance. It respects existing permissions within connected apps — ChatGPT can only access what a user is already authorized to view— and never trains on company data by default.Security features include industry-standard encryption, support for SSO and SCIM for account provisioning, and IP allowlisting to restrict access to approved corporate networks. Enterprise administrators can also define role-based access control (RBAC) policies and manage permissions at a group or department level.OpenAI’s Enterprise Compliance API provides a full audit trail, allowing administrators to review conversation logs for reporting and regulatory purposes. This capability helps enterprises meet internal governance standards and industry-specific requirements such as SOC 2 and ISO 27001 compliance.Admin Configuration and Connector ManagementFor enterprise deployment, administrators must enable company knowledge and its connectors within the ChatGPT workspace. Once connectors are active, users can authenticate their own accounts for each work app they need to access.In Enterprise and Edu plans, connectors are off by default and require explicit admin approval before employees can use them. Admins can selectively enable connectors, manage access by role, and require SSO-based authentication for enhanced control.Business plan users, by contrast, have connectors enabled automatically if available in their workspace. Admins can still oversee which connectors are approved, ensuring alignment with internal IT and data policies.Company knowledge becomes available to any user with at least one active connector, and admins can configure group-level permissions for different teams — such as restricting GitHub access to engineering while enabling Google Drive or HubSpot for marketing and sales.Organizations who turn on the feature can also elect to turn it off just as easily. Once you disconnect a connector, ChatGPT does not have access to that data.How Company Knowledge Works in PracticeActivating company knowledge is straightforward. Users can start a new or existing conversation in ChatGPT and select “Company knowledge” under the message composer or from the tools menu. It must be turned on proactively for each new conversation or chat session, even from the same user.After authenticating their connected apps, they can ask questions as usual—such as “Summarize this account’s latest feedback and risks” or “Compile a Q4 performance summary from project trackers.”ChatGPT searches across the connected tools, retrieves relevant context, and produces an answer with full citations and source links. The system can combine data across apps — for instance, blending Slack updates, Google Docs notes, and HubSpot CRM records — to create an integrated view of a project, client, or initiative.When company knowledge is not selected, ChatGPT may still use connectors in a limited capacity as part of the default experience, but responses will not include detailed citations or multi-source synthesis.Advanced Use Cases for Enterprise TeamsFor development and operations leaders, company knowledge can act as a centralized intelligence layer that surfaces real-time updates and dependencies across complex workflows. ChatGPT can, for example, summarize open GitHub pull requests, highlight unresolved Linear tickets, and cross-reference Slack engineering discussions—all in a single output.Technical teams can also use it for incident retrospectives or release planning by pulling relevant information from issue trackers, logs, and meeting notes. Procurement or finance leaders can use it to consolidate purchase requests or budget updates across shared drives and internal communications.Because the model can reference structured and unstructured data simultaneously, it supports wide-ranging scenarios—from compliance documentation reviews to cross-departmental performance summaries.Privacy, Data Residency, and ComplianceEnterprise data protection is a central design element of company knowledge. ChatGPT processes data in line with OpenAI’s enterprise-grade security model, ensuring that no connected app data leaves the secure boundary of the organization’s authorized environment.Data residency policies vary by connector. Certain integrations, such as Slack, support region-specific data storage, while others—like Google Drive and SharePoint—are available for U.S.-based customers with or without at-rest data residency. Organizations with regional compliance obligations can review connector-specific security documentation for details.No geo restrictions apply to company knowledge, making it suitable for multinational organizations operating across multiple jurisdictions.Limitations and Future EnhancementsAt present, users must manually enable company knowledge in each new ChatGPT conversation. OpenAI is developing a unified interface that will automatically integrate company knowledge with other ChatGPT tools—such as browsing and chart generation—so that users won’t need to toggle between modes.When enabled, company knowledge temporarily disables web browsing and visual output generation, though users can switch modes within the same conversation to re-enable those features.OpenAI also continues to expand the network of supported tools. Recent updates have added connectors for Asana, GitLab Issues, and ClickUp, and OpenAI plans to support future MCP (Model Context Protocol) connectors to enable custom, developer-built integrations.Availability and Getting StartedCompany knowledge is now available to all ChatGPT Business, Enterprise, and Edu users. Organizations can begin by enabling the feature under the ChatGPT message composer and connecting approved work apps.For enterprise rollouts, OpenAI recommends a phased deployment: first enabling core connectors (such as Google Drive and Slack), configuring RBAC and SSO, then expanding to specialized systems once data access policies are verified.Procurement and security leaders evaluating the feature should note that company knowledge is covered under existing ChatGPT Enterprise terms and uses the same encryption, compliance, and service-level guarantees.With company knowledge, OpenAI aims to make ChatGPT not just a conversational assistant but an intelligent interface to enterprise data—delivering secure, context-aware insights that help technical and business leaders act with confidence.

We’re partnering with Commonwealth Fusion Systems (CFS) to bring clean, safe, limitless fusion energy closer to reality.

#amazon bedrock #amazon machine learning #amazon neptune #amazon nova #artificial intelligence #database

In this post, we explore an innovative approach that converts natural language to Gremlin queries using Amazon Bedrock models such as Amazon Nova Pro, helping business analysts and data scientists access graph databases without requiring deep technical expertise. The methodology involves three key steps: extracting graph knowledge, structuring the graph similar to text-to-SQL processing, and generating executable Gremlin queries through an iterative refinement process that achieved 74.17% overall accuracy in testing.

#best practices #generative ai #responsible ai

In this post, we explore how companies can systematically incorporate responsible AI practices into their generative AI project prioritization methodology to better evaluate business value against costs while addressing novel risks like hallucination and regulatory compliance. The post demonstrates through a practical example how conducting upfront responsible AI risk assessments can significantly change project rankings by revealing substantial mitigation work that affects overall project complexity and timeline.

#ai

Microsoft today held a live announcement event online for its Copilot AI digital assistant, with Mustafa Suleyman, CEO of Microsoft's AI division, and other presenters unveiling a new generation of features that deepen integration across Windows, Edge, and Microsoft 365, positioning the platform as a practical assistant for people during work and off-time, while allowing them to preserve control and safety of their data.The new Copilot 2025 Fall Update features also up the ante in terms of capabilities and the accessibility of generative AI assistance from Microsoft to users, so businesses relying on Microsoft products, and those who seek to offer complimentary or competing products, would do well to review them.Suleyman emphasized that the updates reflect a shift from hype to usefulness. “Technology should work in service of people, not the other way around,” he said. “Copilot is not just a product—it’s a promise that AI can be helpful, supportive, and deeply personal.”Intriguingly, the announcement also sought to shine a greater spotlight on Microsoft's own homegrown AI models, as opposed to those of its partner and investment OpenAI, which previously powered the entire Copilot experience. Instead, Suleyman wrote today in a blog post: “At the foundation of it all is our strategy to put the best models to work for you – both those we build and those we don’t. Over the past few months, we have released in-house models like MAI-Voice-1, MAI-1-Preview and MAI-Vision-1, and are rapidly iterating.”12 Features That Redefine CopilotThe Fall Release consolidates Copilot’s identity around twelve key capabilities—each with potential to streamline organizational knowledge work, development, or support operations.Groups – Shared Copilot sessions where up to 32 participants can brainstorm, co-author, or plan simultaneously. For distributed teams, it effectively merges a meeting chat, task board, and generative workspace. Copilot maintains context, summarizes decisions, and tracks open actions. Imagine – A collaborative hub for creating and remixing AI-generated content. In an enterprise setting, Imagine enables rapid prototyping of visuals, marketing drafts, or training materials.Mico – A new character identity for Copilot that introduces expressive feedback and emotional expression in the form of a cute, amorphous blob. Echoing Microsoft’s historic character interfaces like Clippy (Office 97) or Cortana (2014), Mico serves as a unifying UX layer across modalities.Real Talk – A conversational mode that adapts to a user’s communication style and offers calibrated pushback — ending the sycophancy that some users have complained about with other AI models such as prior versions of OpenAI's ChatGPT. For professionals, it allows Socratic problem-solving rather than passive answer generation, making Copilot more credible in technical collaboration.Memory & Personalization – Long-term contextual memory that lets Copilot recall key details—training plans, dates, goals—at the user’s direction.Connectors – Integration with OneDrive, Outlook, Gmail, Google Drive, and Google Calendar for natural-language search across accounts.Proactive Actions (Preview) – Context-based prompts and next-step suggestions derived from recent activity.Copilot for Health – Health information grounded in credible medical sources such as Harvard Health, with tools allowing users to locate and compare doctors.Learn Live – A Socratic, voice-driven tutoring experience using questions, visuals, and whiteboards.Copilot Mode in Edge – Converts Microsoft Edge into an “AI browser” that summarizes, compares, and executes web actions by voice.Copilot on Windows – Deep integration across Windows 11 PCs with “Hey Copilot” activation, Copilot Vision guidance, and quick access to files and apps.Copilot Pages and Copilot Search – A collaborative file canvas plus a unified search experience combining AI-generated, cited answers with standard web results.The Fall Release is immediately available in the United States, with rollout to the UK, Canada, and other markets in progress. Some functions—such as Groups, Journeys, and Copilot for Health—remain U.S.-only for now. Proactive Actions requires a Microsoft 365 Personal, Family, or Premium subscription.Together these updates illustrate Microsoft’s pivot from static productivity suites to contextual AI infrastructure, with the Copilot brand acting as the connective tissue across user roles.From Clippy to Mico: The Return of a Guided InterfaceOne of the most notable introductions is Mico, a small animated companion that is available within Copilot’s voice-enabled experiences, including the Copilot app on Windows, iOS, and Android, as well as in Study Mode and other conversational contexts. It serves as an optional visual companion that appears during interactive or voice-based sessions, rather than across all Copilot interfaces.Mico listens, reacts with expressions, and changes color to reflect tone and emotion — bringing a visual warmth to an AI assistant experience that has traditionally been text-heavy.Mico’s design recalls earlier eras of Microsoft’s history with character-based assistants. In the mid-1990s, Microsoft experimented with Microsoft Bob (1995), a software interface that used cartoon characters like a dog named Rover to guide users through everyday computing tasks. While innovative for its time, Bob was discontinued after a year due to performance and usability issues.A few years later came Clippy, the Office Assistant introduced in Microsoft Office 97. Officially known as “Clippit,” the animated paperclip would pop up to offer help and tips within Word and other Office applications. Clippy became widely recognized—sometimes humorously so—for interrupting users with unsolicited advice. Microsoft retired Clippy from Office in 2001, though the character remains a nostalgic symbol of early AI-driven assistance.More recently, Cortana, launched in 2014 as Microsoft’s digital voice assistant for Windows and mobile devices, aimed to provide natural-language interaction similar to Apple’s Siri or Amazon’s Alexa. Despite positive early reception, Cortana’s role diminished as Microsoft refocused on enterprise productivity and AI integration. The service was officially discontinued on Windows in 2023.Mico, by contrast, represents a modern reimagining of that tradition—combining the personality of early assistants with the intelligence and adaptability of contemporary AI models. Where Clippy offered canned responses, Mico listens, learns, and reflects a user’s mood in real time. The goal, as Suleyman framed it, is to create an AI that feels “helpful, supportive, and deeply personal.”Groups Are Microsoft's Version of Claude and ChatGPT ProjectsDuring Microsoft’s launch video, product researcher Wendy described Groups as a transformative shift: “You can finally bring in other people directly to the conversation that you’re having with Copilot,” she said. “It’s the only place you can do this.”Up to 32 users can join a shared Copilot session, brainstorming, editing, or planning together while the AI manages logistics such as summarizing discussion threads, tallying votes, and splitting tasks. Participants can enter or exit sessions using a link, maintaining full visibility into ongoing work.Instead of a single user prompting an AI and later sharing results, Groups lets teams prompt and iterate together in one unified conversation. In some ways, it's an answer to Anthropic’s Claude Projects and OpenAI’s ChatGPT Projects, both launched within the last year as tools to centralize team workspaces and shared AI context. Where Claude and ChatGPT Projects allow users to aggregate files, prompts, and conversations into a single container, Groups extends that model into real-time, multi-participant collaboration. Unlike Anthropic’s and OpenAI’s implementations, Groups is deeply embedded within Microsoft’s productivity environment. Like other Copilot experiences connected to Outlook and OneDrive, Groups operates within Microsoft’s enterprise identity framework, governed by Microsoft 365 and Entra ID (formerly Azure Active Directory) authentication and consent modelsThis means conversations, shared artifacts, and generated summaries are governed under the same compliance policies that already protect Outlook, Teams, and SharePoint data.Hours after the unveiling, OpenAI hit back against its own investor in the escalating AI competition between the "frenemies" by expanding its Shared Projects feature beyond its current Enterprise, Team, and Edu subscriber availability to users of its free, Plus, and Pro subscription tiers. Operational Impact for AI and Data TeamsMemory & Personalization and Connectors effectively extend a lightweight orchestration layer across Microsoft’s ecosystem. Instead of building separate context-stores or retrieval APIs, teams can leverage Copilot’s secure integration with OneDrive or SharePoint as a governed data backbone. A presenter explained that Copilot’s memory “naturally picks up on important details and remembers them long after you’ve had the conversation,” yet remains editable. For data engineers, Copilot Search and Connectors reduce friction in data discovery across multiple systems. Natural-language retrieval from internal and cloud repositories may lower the cost of knowledge management initiatives by consolidating search endpoints.For security directors, Copilot’s explicit consent requirements and on/off toggles in Edge and Windows help maintain data residency standards. The company reiterated during the livestream that Copilot “acts only with user permission and within organizational privacy controls.”Copilot Mode in Edge: The AI Browser for Research and AutomationCopilot Mode in Edge stands out for offering AI-assisted information workflows. The browser can now parse open tabs, summarize differences, and perform transactional steps.“Historically, browsers have been static—just endless clicking and tab-hopping,” said a presenter during Microsoft’s livestream. “We asked not how browsers should work, but how people work.”In practice, an analyst could prompt Edge to compare supplier documentation, extract structured data, and auto-fill procurement forms—all with consistent citation. Voice-only navigation enables accessibility and multitasking, while Journeys, a companion feature, organizes browsing sessions into storylines for later review.Copilot on Windows: The Operating System as an AI SurfaceIn Windows 11, Copilot now functions as an embedded assistant. With the wake-word “Hey Copilot,” users can initiate context-aware commands without leaving the desktop—drafting documentation, troubleshooting configuration issues, or summarizing system logs.A presenter described it as a “super assistant plugged into all your files and applications.” For enterprises standardizing on Windows 11, this positions Copilot as a native productivity layer rather than an add-on, reducing training friction and promoting secure, on-device reasoning.Copilot Vision, now in early deployment, adds visual comprehension. IT staff can capture a screen region and ask Copilot to interpret error messages, explain configuration options, or generate support tickets automatically.Combined with Copilot Pages, which supports up to twenty concurrent file uploads, this enables more efficient cross-document analysis for audits, RFPs, or code reviews.Leveraging MAI Models for Multimodal WorkflowsAt the foundation of these capabilities are Microsoft’s proprietary MAI-Voice-1, MAI-1 Preview, and MAI-Vision-1 models—trained in-house to handle text, voice, and visual inputs cohesively.For engineering teams managing LLM orchestration, this architecture introduces several potential efficiencies:Unified multimodal reasoning – Reduces the need for separate ASR (speech-to-text) and image-parsing services.Fine-tuning continuity – Because Microsoft owns the model stack, updates propagate across Copilot experiences without re-integration.Predictable latency and governance – In-house hosting under Azure compliance frameworks simplifies security certification for regulated industries.A presenter described the new stack as “the foundation for immersive, creative, and dynamic experiences that still respect enterprise boundaries.”A Strategic Pivot Toward Contextual AIFor years, Microsoft positioned Copilot primarily as a productivity companion. With the Fall 2025 release, it crosses into operational AI infrastructure—a set of extensible services for reasoning over data and processes.Suleyman described this evolution succinctly: “Judge an AI by how much it elevates human potential, not just by its own smarts.” For CIOs and technical leads, the elevation comes from efficiency and interoperability.Copilot now acts as:A connective interface linking files, communications, and cloud data.A reasoning agent capable of understanding context across sessions and modalities.A secure orchestration layer compatible with Microsoft’s compliance and identity framework.Suleyman’s insistence that “technology should work in service of people” now extends to organizations as well: technology that serves teams, not workloads; systems that adapt to enterprise context rather than demand it.

We're rolling out Deep Think in the Gemini app for Google AI Ultra subscribers, and we're giving select mathematicians access to the full version of the Gemini 2.5 Deep Think model entered into the IMO competition.

Game Arena is a new, open-source platform for rigorous evaluation of AI models. It allows for head-to-head comparison of frontier systems in environments with clear winning conditions.

Today, we're adding a new, highly specialized tool to the Gemma 3 toolkit: Gemma 3 270M, a compact, 270-million parameter model.

Transform images in amazing new ways with updated native image editing in the Gemini app.

We introduce VaultGemma, the most capable model trained from scratch with differential privacy.

Available in preview via the API, our Computer Use model is a specialized model built on Gemini 2.5 Pro’s capabilities to power agents that can interact with user interfaces.

We’re rolling out significant updates to Veo that give people even more creative control.

We’re launching a new 27 billion parameter foundation model for single-cell analysis built on the Gemma family of open models.

#artificial intelligence #data science #deep learning #machine learning #vision transformer #fast fourier transform

Exploring the frequency fingerprints of Transformers to guide smarter knowledge distillation
The post When Transformers Sing: Adapting SpectralKD for Text-Based Knowledge Distillation appeared first on Towards Data Science.

« 1...2627282930...191»
×