Published on: March 19, 2026 7 min read GitLab 18.10: Agentic AI now open to even more teams on GitLab Free GitLab.com teams can purchase GitLab Credits and start using AI agents and workflows, including flat-rate automated code review. features product Agentic AI is changing ho…
Transform live video for mobile audiences with AWS Elemental Inference Today, we’re announcing AWS Elemental Inference, a fully managed AI service that automatically transforms and maximizes live and on-demand video broadcasts to engage audiences at scale. At launch, you’ll be a…
GitLab 18.9 introduces critical updates designed to provide regulated enterprises with governed, agentic AI capabilities through self-hosted infrastructure and model flexibility. By combining the Duo Agent Platform with Bring Your Own Model (BYOM) support, organizations in sectors like finance and government can now automate complex DevSecOps workflows while maintaining total control over data residency. This release transforms GitLab into a high-security AI control plane that balances the need for advanced automation with the rigid sovereignty requirements of high-compliance environments.
## Self-Hosted Duo Agent Platform for Online Cloud Licenses
The Duo Agent Platform allows engineering teams to automate sequences of tasks, such as hardening CI/CD pipelines and triaging vulnerabilities, but was previously difficult to deploy for customers under strict online cloud licensing. This update makes the platform generally available for these environments, bridging the gap between cloud-based licensing and self-hosted security needs.
* **Usage-Based Billing:** The platform now utilizes GitLab Credits to provide transparent, per-request metering, which is essential for internal chargeback and regulatory reporting.
* **Infrastructure Control:** Enterprises can host models on their own internal infrastructure or within approved cloud environments, ensuring that inference traffic is routed according to internal security policies.
* **Deployment Readiness:** By removing the requirement to route data through external AI vendors, the platform is now a viable option for critical infrastructure and government agencies.
## Bring Your Own Model (BYOM) Integration
Recognizing that many enterprises have already invested in domain-tuned LLMs or air-gapped deployments, GitLab now allows customers to integrate their existing models directly into the Duo Agent Platform. This ensures that organizations are not locked into a specific vendor and can leverage models that have already passed internal risk assessments.
* **AI Gateway Connectivity:** Administrators can connect third-party or internal models via the GitLab AI Gateway, allowing these models to function as enterprise-ready options within the GitLab ecosystem.
* **Granular Model Mapping:** The system provides the ability to map specific models to individual Duo Agent Platform flows or features, giving admins fine-grained control over which agent uses which model.
* **Administrative Ownership:** While GitLab provides the orchestration layer, administrators retain full responsibility for model validation, performance tuning, and risk evaluation for the models they choose to bring.
For organizations operating in high-compliance sectors, these updates offer a path to consolidate fragmented AI tools into a single, governed platform. Engineering leaders should evaluate their current model investments and leverage the GitLab AI Gateway to unify their automation workflows under one secure DevSecOps umbrella.
Our Multi-Agent Architecture for Smarter Advertising Introduction When we kicked this off, we weren’t trying to ship an “AI feature.” We were trying to fix a structural problem in how our ads business actually runs in software. On the business side, we have multiple ways of buyi…
Google’s AI advantage: why crawler separation is the only path to a fair Internet 2026-01-30 Maria Palmieri Sebastian Hufnagel Earlier this week, the UK’s Competition and Markets Authority (CMA) opened its consultation on a package of proposed conduct requirements for Google. Th…
While generative AI focuses on creating content like text and images through prompt-based prediction, agentic AI represents a shift toward autonomous goal achievement and execution. By combining the creative output of large language models with a continuous loop of perception and action, these technologies allow users to move from simply generating drafts to managing complex, multi-step workflows. Ultimately, the two systems are most effective when used together, with one providing the ideas and the other handling the coordination and follow-through.
### Distinguishing Creative Output from Autonomous Agency
* Generative AI functions as a responder that produces new content—such as text, code, or visuals—by predicting the most likely next "token" or piece of data based on a user’s prompt.
* Agentic AI possesses "agency," meaning it can take a high-level goal (e.g., "prepare a client kickoff") and determine the necessary steps to achieve it with minimal guidance.
* While tools like Midjourney or GitHub Copilot focus on the immediate delivery of a specific creative asset, agentic systems act as proactive partners that can use external tools, manage schedules, and make independent decisions.
### The Underlying Mechanics of Prediction and Action
* Generative models rely on Large Language Models (LLMs) trained on massive datasets to identify patterns and chain together original sequences of information.
* Agentic systems operate on a "perceive, plan, act, and learn" loop, where the AI gathers context from its environment, executes tasks across different applications, and adjusts its strategy based on the results.
* The generative process is typically a direct path from input to output, whereas the agentic process is iterative, allowing the system to adapt to changes and feedback in real-time.
### Practical Applications in Content and Workflow Management
* Generative use cases include transforming rough bullet points into polished emails, summarizing long documents into flashcards, and adjusting the tone of a message to be more professional.
* Agentic use cases involve higher-level orchestration, such as monitoring document revisions, consolidating feedback from multiple stakeholders, and automatically sending follow-up reminders.
* In a project management context, an agentic system can draft a project plan, identify owners for specific tasks, and update timelines as milestones are met or missed.
### Navigating Technical and Operational Limitations
* Generative AI is susceptible to "hallucinations" because it prioritizes probabilistic output over factual reasoning or logic.
* Agentic AI introduces complexity regarding security and permissions, as the system needs authorized access to various apps and tools to perform actions on a user's behalf.
* Current agentic systems still require human oversight for critical decision-making to ensure that autonomous actions align with the user's intent and organizational standards.
To maximize efficiency, you should utilize generative AI for the creative phases of a project—such as brainstorming and drafting—while delegating administrative overhead and coordination to agentic AI. As these technologies continue to converge, the focus of AI utility is shifting from the volume of content produced to the successful execution of complex, real-world results.
GitLab is transitioning from seat-based pricing to a usage-based model with the introduction of GitLab Credits, a virtual currency designed for the GitLab Duo Agent Platform. This shift addresses the limitations of traditional licensing, which often creates "AI haves and have-nots" by making access too expensive for light or occasional users. By pooling resources across an entire organization, GitLab aims to provide equitable access to agentic AI for every developer while ensuring costs align with actual consumption.
## The Shift from Seat-Based to Usage-Based AI
* Traditional seat-based models are poorly suited for agentic AI, which can be triggered by background SDLC events rather than just direct user interaction.
* The credit model allows every member of a Premium or Ultimate organization to use AI capabilities without requiring an individual "AI seat."
* Usage-based pricing automatically offsets the costs of power users against lighter users, lowering the total cost of ownership for the organization.
## Mechanics of GitLab Credits
* Credits function as a pooled resource consumed by both synchronous interactions (like Agentic Chat in the IDE) and asynchronous background tasks.
* Supported capabilities include foundational agents (Security, Planner, Data Analyst) and specific workflows such as Code Review and CI/CD pipeline fixing.
* The system integrates with external models like Anthropic Claude Code and OpenAI Codex, as well as custom agents published in the GitLab AI Catalog.
* Each credit has an on-demand list price of $1, with volume discounts available for enterprise customers who sign up for annual commitments.
## Governance and Usage Controls
* Administrators can monitor consumption through two dedicated dashboards: a financial oversight portal for billing managers and an operational monitoring view for administrators.
* Granular controls allow organizations to enable or disable Duo Agent Platform access for specific teams or projects to prevent unexpected credit depletion.
* Proactive email alerts are triggered when consumption reaches 50%, 80%, and 100% of committed monthly credits.
* A sizing calculator is available to help organizations estimate their monthly credit requirements based on patterns observed during the platform's beta period.
## Transitioning and Promotional Access
* Existing GitLab Duo Pro and Duo Enterprise customers can roll over their current seat investments into GitLab Credits with volume-based discounts.
* As part of a limited-time promotion, GitLab is providing $12 in monthly credits per user for Premium subscribers and $24 per user for Ultimate subscribers.
* Self-managed and GitLab Dedicated customers will gain access to these credit-based features starting with the 18.8 and 18.9 releases.
For organizations looking to scale AI across the software development lifecycle, the credit-based model offers a more flexible and cost-effective path than rigid seat licenses. Current Premium and Ultimate subscribers should leverage their monthly promotional credits to baseline their usage before committing to larger annual credit bundles.
Kakao’s development of the Kanana-2 model family represents a strategic shift toward Agentic AI, prioritizing complex reasoning and execution capabilities over simple conversational fluency. By implementing a sophisticated post-training pipeline—including a specialized Mid-training stage and refined reinforcement learning—the team successfully enhanced the model's instruction-following and tool-calling performance. This methodology ensures that the 30B parameter models excel in logical tasks and real-world agentic environments while maintaining high linguistic stability in both English and Korean.
## Mid-training and Catastrophic Forgetting Prevention
* A 250B token Mid-training stage was introduced between Pre-training and Post-training to bridge the gap in reasoning, coding, and tool-calling capabilities.
* The dataset comprised 200B tokens of high-quality reasoning data (Chain-of-Thought math and code) and 50B tokens of "replay" data from the original pre-training set.
* This replay strategy specifically targeted "Catastrophic Forgetting," preventing the model from losing its Korean linguistic nuances and performance on benchmarks like KoMT-bench while it gained English-heavy reasoning skills.
* Experimental results indicated that Mid-training serves as a foundational "force multiplier," leading to faster convergence and higher performance ceilings during subsequent Supervised Fine-Tuning (SFT) and Reinforcement Learning (RL) stages.
## Enhanced Instruction Following and Tool Calling
* To optimize for Agentic AI, the developers focused on Instruction Following (IFEval) by synthesizing high-quality, long-form responses that strictly adhere to complex constraints.
* Tool-calling capabilities were improved using "Rejection Sampling" (Iterative SFT), where model-generated trajectories are validated in a real execution environment; only successful outcomes are retained for training.
* The training data was categorized into distinct buckets—such as Chat, Math, Code, and Tool Calling—allowing for a more balanced recipe compared to previous Kanana versions.
* This approach specifically addressed multi-turn and multi-tool scenarios, ensuring the model can handle the recursive logic required for autonomous agents.
## Parallel Reinforcement Learning and Calibration Tuning
* A "Parallel RL" framework was adopted to optimize different capabilities simultaneously: the "Chat" track focused on helpfulness and safety, while the "Logic" track focused on accuracy in math and programming.
* The pipeline moved beyond standard SFT to include Reinforcement Learning from Human Feedback (RLHF), utilizing DPO and PPO-style methods to align the model with human preferences.
* A final "Calibration Tuning" step was implemented to ensure the model’s internal confidence levels match its actual accuracy, effectively reducing hallucinations and improving reliability in technical tasks.
* Comparative benchmarks show that the Kanana-2 Instruct and Thinking models significantly outperform earlier versions and rival larger open-source models in reasoning and coding benchmarks like HumanEval and GSM8K.
The Kanana-2 development cycle demonstrates that achieving "Agentic" performance requires more than just scaling data; it requires a structured transition from general language understanding to execution-verified reasoning. For organizations building AI agents, the Kanana-2 post-training recipe suggests that integrating environment-validated feedback and balancing reasoning data with foundational language "replays" is critical for creating reliable, multi-functional models.
The GitLab Duo Agent Platform has reached general availability, marking a shift from basic AI code assistance to comprehensive agentic automation across the entire software development lifecycle. By orchestrating intelligent agents to handle complex tasks like security analysis and planning, the platform aims to resolve the "AI paradox" where faster code generation often creates downstream bottlenecks in review and deployment.
### Usage-Based Economy via GitLab Credits
* GitLab is introducing "GitLab Credits," a virtual currency used to power the platform’s usage-based AI features.
* Premium and Ultimate subscribers receive monthly credits ($12 and $24 respectively) at no additional cost to facilitate immediate adoption.
* Organizations can manage a shared pool of credits or opt for on-demand monthly billing, with existing Duo Enterprise contracts eligible for conversion into credits.
### Agentic Chat and Contextual Orchestration
* The Duo Agentic Chat provides a unified experience across the GitLab Web UI and various IDEs, including VS Code, JetBrains, Cursor, and Windsurf.
* The chat utilizes multi-step reasoning to perform actions autonomously, drawing from the context of issues, merge requests, pipelines, and security findings.
* Capabilities extend beyond code generation to include infrastructure-as-code (IaC) creation, pipeline troubleshooting, and explaining vulnerability reachability.
### Specialized Foundational and Custom Agents
* **Foundational Agents:** Pre-built specialists designed for specific roles, such as the Planner Agent for breaking down work and the Security Analyst Agent for triaging vulnerabilities.
* **Custom Agents:** Developed through a central AI Catalog, these allow teams to build and share agents that adhere to organization-specific engineering standards and guardrails.
* **External Agents:** Native integration of third-party AI tools, such as Anthropic’s Claude Code and OpenAI’s Codex CLI, provides access to external LLM capabilities within the governed GitLab environment.
### Automated End-to-End Flows
* The platform introduces "Flows," which are multi-step agentic sequences designed to automate repeatable transitions in the development cycle.
* The "Issue to Merge Request" flow builds structured code changes directly from defined requirements to jumpstart development.
* Specialized CI/CD flows help teams modernize pipeline configurations and automatically analyze and suggest fixes for failed pipeline runs.
* The Code Review flow streamlines the feedback loop by providing AI-native analysis of merge request comments and code changes.
To maximize the impact of agentic AI, organizations should move beyond basic chat interactions and begin integrating these specialized agents into their broader orchestration workflows to eliminate manual handoffs between planning, coding, and security.
The first AWS Weekly Roundup of 2026 highlights a strategic focus on community-driven AI innovation and significant performance upgrades to the EC2 instance lineup. By combining high-stakes competitions like the 10,000 AIdeas challenge with technical releases such as Graviton4-powered instances, AWS is positioning itself to lead in both "Agentic AI" development and high-performance cloud infrastructure.
**AI Innovation and Professional Mentorship**
* The "Become a Solutions Architect" (BeSA) program is launching a new six-week cohort on February 21, 2026, specifically focused on Agentic AI on AWS.
* The Global 10,000 AIdeas Competition offers a $250,000 prize pool and recognition at re:Invent 2026, with a submission deadline of January 21, 2026.
* Competition participants are required to utilize the "Kiro" development tool and must ensure their applications remain within AWS Free Tier limits.
**Next-Generation EC2 Instances and Hardware**
* New M8gn and M8gb instances utilize AWS Graviton4 processors, providing a 30% compute performance boost over the previous Graviton3 generation.
* The M8gn variant features 6th generation AWS Nitro Cards, delivering up to 600 Gbps of network bandwidth, the highest available for network-optimized instances.
* The M8gb variant is optimized for storage-heavy workloads, offering up to 150 Gbps of dedicated Amazon EBS bandwidth.
**Resilience Testing and Governance**
* AWS Direct Connect now integrates with the AWS Fault Injection Service (FIS), allowing engineers to simulate Border Gateway Protocol (BGP) failovers to validate redundant pathing.
* AWS Control Tower has expanded its governance capabilities by supporting 176 additional Security Hub controls within the Control Catalog.
* These controls address a broad spectrum of requirements across security, cost optimization, operations, and data durability.
**Hybrid Cloud and Windows Support**
* Amazon ECS Managed Instances now support Windows Server for on-premises and remote environment management.
* The service uses AWS Systems Manager (SSM) to register external instances, which can then be managed as part of an ECS cluster using Windows-based ECS-optimized AMIs.
Developers and infrastructure architects should prioritize the January 21 deadline for AI project submissions while evaluating the M8gn instances for high-throughput networking requirements. Additionally, organizations running hybrid Windows workloads should explore the new ECS Managed Instances support to unify their container orchestration across on-premises and cloud environments.
Kakao has released Kanana-2, a high-performance open-source language model specifically engineered to power Agentic AI by enhancing tool-calling and instruction-following capabilities. Surpassing its predecessors and rivaling global frontier models like Qwen3, Kanana-2 offers a versatile suite of variants designed for practical, high-efficiency application in complex service environments.
### Optimized Model Lineup: Base, Instruct, and Thinking
* **Kanana-2-30b-a3b-base:** Provided as a foundational model with pre-training weights, allowing researchers to fine-tune the model using their own datasets.
* **Kanana-2-30b-a3b-instruct:** A version optimized through post-training to maximize the model's ability to follow complex user instructions accurately.
* **Kanana-2-30b-a3b-thinking:** Kakao’s first reasoning-specialized model, designed for tasks requiring high-level logical thinking, such as mathematics and coding.
### Strengthening Agentic AI Capabilities
* **Tool Calling:** Multi-turn tool-calling performance has improved more than threefold compared to Kanana-1.5, significantly enhancing its utility with the Model Context Protocol (MCP).
* **Instruction Following:** The model's ability to understand and execute multi-step, complex user requirements has been refined to ensure reliable task completion.
* **Reasoning-Tool Integration:** Unlike many reasoning models that lose instruction-following quality during deep thought, the "Thinking" variant maintains high performance in both logical deduction and tool use.
### High-Efficiency Architecture for Scale
* **MLA (Multi-head Latent Attention):** Compresses memory usage to handle long contexts more efficiently, reducing the resources needed for extensive data processing.
* **MoE (Mixture of Experts):** Activates only the necessary parameters during inference, maintaining high performance while drastically reducing computational costs and response times.
* **Improved Tokenization:** A newly trained tokenizer has improved Korean language token efficiency by 30%, enabling faster throughput and lower latency in high-traffic environments like KakaoTalk.
### Expanded Multilingual Support
* **Broad Linguistic Reach:** The model has expanded its support from just Korean and English to include six languages: Korean, English, Japanese, Chinese, Thai, and Vietnamese.
By open-sourcing Kanana-2, Kakao provides a robust foundation for developers seeking to build responsive, tool-integrated AI services. Its focus on practical efficiency and advanced reasoning makes it an ideal choice for implementing agentic workflows in real-world applications where speed and accuracy are critical.
How Dash uses context engineering for smarter AI When we first built Dash, it looked like most enterprise search systems: a traditional RAG pipeline that combined semantic and keyword search across indexed documents. It worked well for retrieving information and generating conci…
Introducing our MCP server: Bringing Figma into your workflow Inside Figma Product updates Dev Mode AI Engineering News Today we’re announcing the beta release of the Figma MCP server, which brings Figma directly into the developer workflow to help LLMs achieve design-informed c…
Figma's 2025 AI report: Perspectives from designers and developers Insights AI Research Report Thought leadership Figma’s AI report tells us how designers and developers are navigating the changing landscape. Artwork by Saiman Chow Agentic AI is the fastest growing product categ…
Double click: Human takes on agentic AI There’s a lot of buzz about AI agents. Robots that do more with less supervision—what could go wrong? We asked our community how this might shake up how we think about UX. Insights AI UI/UX Design Thought leadership