MoreRSS

site iconJimmy Song | 宋净超修改

Tetrate 布道师,云原生社区 创始人,CNCF Ambassador,云原生技术专家。
请复制 RSS 到你的阅读器,或快速订阅到 :

Inoreader Feedly Follow Feedbin Local Reader

Jimmy Song | 宋净超的 RSS 预览

Bun Acquired by Anthropic: A Structural Signal for AI-Native Runtimes

2025-12-03 13:21:28

The shifting ownership of runtimes is reshaping the underlying logic of AI programming and infrastructure.

After the announcement of Bun’s acquisition by Anthropic , my focus was not on the deal itself, but on the structural signal it revealed: general-purpose language runtimes are now being drawn into the path dependencies of AI programming systems. This is not just “a JS project finding a home,” but “the first time a language runtime has been actively integrated into the unified engineering system of a leading large model company.”

This event deserves a deeper analysis.

Bun’s Engineering Features and Current Status

Before examining Bun ’s industry significance, let’s outline its runtime characteristics. The following list summarizes Bun’s main engineering capabilities:

  • High-performance JavaScript/TypeScript runtime
  • Built-in bundler, test framework, and package manager
  • Single-file executable
  • Extremely fast cold start
  • Node compatibility without Node’s legacy dependencies

These capabilities have formed measurable performance barriers.

However, it should be noted that Bun currently lacks the core attributes of an AI Runtime, including:

  • Permission model
  • Tool isolation
  • Capability declaration protocol
  • Execution semantics understandable by models
  • Sandbox execution environment

Therefore, Bun’s “AI Native” properties have not yet been established, but Anthropic’s acquisition provides an opportunity for it to evolve in this direction.

The Significance of a Leading Model Company Acquiring a General-Purpose Runtime

Historically, it is not uncommon for model companies to acquire editors, plugins, or IDEs, but in known public cases, mainstream large model vendors have never directly acquired a mature general-purpose language runtime. Bun × Anthropic is the first clear event pulling the runtime into the AI programming system landscape. This move sends two engineering-level signals:

  • The speed of AI code generation continues to increase, amplifying the need for deterministic execution environments. The generate→execute→validate→destroy cycle intensifies the problem of environment non-repeatability.
  • Models require a “controllable execution substrate” rather than a traditional operating system. Agents are not suited to run tools in an uncontrollable, unpredictable OS layer.
  • The runtime needs to be embedded into the model’s internal engineering pipeline. Future IDEs, agents, and auto-repair pipelines may directly invoke the runtime’s API.

This is not a short-term business integration, but a manifestation of the trend toward compressed engineering pipelines.

Runtime Requirements Differentiation in the AI Coding Era

Based on observations of agentic runtimes over the past year, runtime requirements in the AI coding era are diverging. The following list summarizes the main engineering abstractions trending in this space:

  • Determinism: AI-generated code is not reviewed line by line; execution results must be consistent across machines and over time.
  • Minimal distribution unit: Users no longer install language environments and numerous dependencies. Verifiable, replicable, and portable single execution units are becoming the norm.
  • Tool isolation: Models cannot directly access all OS capabilities; the context and permissions visible to tools must be strictly defined.
  • Short-lived execution: Agent invocation patterns resemble “batch jobs” rather than long-running services.
  • Capability declaration: The runtime must expose “what I can do,” rather than the entire OS interface.
  • Embeddable self-testing pipeline: After generating code, models need to immediately execute tests, collect errors, and iterate. The runtime must provide observability and diagnostic primitives.

These requirements are not unique to Bun, nor did Bun originate them, but Bun’s “monolithic and controllable” runtime structure is more conducive to evolving in this direction.

Figure 1: Minimal execution loop of an AI-native runtime
Figure 1: Minimal execution loop of an AI-native runtime

Bun’s Potential Role Within Anthropic’s System

If Bun is seen merely as a Node.js replacement, the acquisition is of limited significance. But if it is viewed as the execution foundation for future AI coding systems, the logic becomes clearer:

  • Code is generated by models
  • Building is handled by the runtime’s built-in toolchain
  • Testing, validation, and repair are performed by models repeatedly invoking the runtime
  • All execution behaviors are defined by the runtime’s semantics
  • The runtime forms Anthropic’s internal “minimal stable layer”

This model is similar to the relationship between Chrome and V8: the execution engine and upper-layer system co-evolve over time, with performance and semantics advancing in sync.

Whether Bun can fulfill this role depends on Anthropic’s architectural choices, but the event itself has opened up possibilities in this direction.

Industry Trends and Future Evolution

Combining facts, signals, and engineering trends, the following directions can be anticipated:

  • The “Agent Runtime” category will gradually become more defined
  • The boundaries between bundler, runtime, and test runner will continue to blur
  • Cloud vendors will launch controllable runtimes with capability declarations
  • Permission models and secure sandboxes will move down to the language runtime layer
  • Runtimes will become part of the model toolchain, rather than an external environment

These trends will not all materialize in the short term, but they represent the inevitable path of engineering evolution.

Summary

The combination of Bun × Anthropic is not about “an open-source project being absorbed,” but about a language runtime being actively integrated into the engineering pipeline of a large model system for the first time. Competition at the model layer will continue, but what truly reshapes software is the structural transformation of AI-native runtimes. This is a foundational change worth long-term attention.

Agentic Runtime Realism: Insights from McKinsey Ark on 2026 Infrastructure Trends

2025-12-02 20:07:45

Statement
ArkSphere has no affiliation or association with McKinsey Ark.

The value of Agentic Runtime lies not in unified interfaces, but in semantic governance and the transformation of engineering paradigms. Ark is just a reflection of the trend; the future belongs to governable Agentic Workloads.

Recently, the ArkSphere community has been focusing on McKinsey’s open-source Ark (Agentic Runtime for Kubernetes). Although the project is still in technical preview, its architecture and semantic model have already become key indicators for the direction of AI Infra in 2026.

This article analyzes the engineering paradigm and semantic model of Ark, highlighting its industry implications. It avoids repeating the reasons for the failure of unified model APIs and generic infrastructure logic, instead focusing on the unique perspective of the ArkSphere community.

Ark’s Semantic Model and Engineering Paradigm

Ark’s greatest value is in making Agents first-class citizens in Kubernetes, achieving closed-loop tasks through CRD (Custom Resource Definition) and controllers (Reconcilers). This semantic abstraction not only enhances governance capabilities but also aligns closely with the Agentic Runtime strategies of major cloud providers.

Ark’s main resources include:

  • Agent (inference entity)
  • Model (model selection and configuration)
  • Tools (capability plugins/MCP, Model Capability Plugin)
  • Team (multi-agent collaboration)
  • Query (task lifecycle)
  • Evaluation (assessment)

The diagram below illustrates the semantic relationships in Agentic Runtime:

Figure 1: Agentic Runtime Semantic Relationships
Figure 1: Agentic Runtime Semantic Relationships

Architecture and Community Activity

Ark’s architecture adopts a standard control plane system, emphasizing unified runtime semantics. The community is highly active, engineer-driven, and the codebase is well-structured, though production readiness is still being improved.

Figure 2: Ark Architecture and Control Plane Flow
Figure 2: Ark Architecture and Control Plane Flow

ArkSphere’s Boundaries and Inspirations

The emergence of Ark has clarified the boundaries of ArkSphere. ArkSphere does not aim for unified model interfaces, multi-cloud abstraction, a collection of miscellaneous tools, or a comprehensive framework layer. Instead, it focuses on:

  • The semantic system of Agentic Runtime (tasks, states, tool invocation, collaboration graphs, etc.)
  • Enterprise-grade runtime governance models (permissions, auditing, isolation, multi-tenancy, compliance, cost tracking)
  • Integration capabilities for domestic ecosystem tools
  • Engineering paradigms from a runtime perspective

ArkSphere is an ecosystem and engineering system at the runtime level, not a “model abstraction layer” or an “agent development framework.”

Key Changes in 2026

2026 will usher in the era of Agentic Runtime, where Agents are no longer just classes but workloads that require governance rather than mere importation. Ark is just one example of this trend, and the direction is clear:

  • Semantic models and governability become highlights
  • Closed-loop tasks are the core value

Summary

Ark’s realism teaches us that the future belongs to runtime, semantics, governability, and workload-level Agents. The industry will no longer pursue unified APIs or framework implementations, but will focus on governable runtime semantics and engineering paradigms.

In-Depth Analysis of Ark: Kubernetes for the AI Era or a New Engineering Paradigm Shift?

2025-12-02 18:54:34

Statement
ArkSphere has no affiliation or association with McKinsey Ark.

The greatest value of Ark lies in reshaping engineering paradigms, not just its features. It points the way for AI Infra and leaves vast space for community ecosystems.

Recently, many members in our ArkSphere community have started exploring McKinsey’s open-source Ark (Agentic Runtime for Kubernetes) .

Some see it as radical, some think it’s just a consulting firm’s experiment, and others quote a realistic maxim:

What we need now is “agentic runtime realism,” not “unified model romanticism.”

I strongly agree with this sentiment.

I’ve spent some time analyzing Ark’s source code, architecture, and design philosophy, combined with our community discussions. My conclusion is:

Ark’s significance is not in its features, but in its paradigm.

It’s not the answer, but it points toward the future of AI Infra.

Below is my interpretation of Ark, focusing on engineering, architecture, trends, and its inspiration for ArkSphere.

What Exactly Is Ark?

Ark’s core positioning is: A runtime that treats Agents as Kubernetes Workloads.

It’s not a framework, not an SDK, not an AutoGen-style multi-agent tool, but a complete system including:

  • Control plane (Controller)
  • Custom resource models (CRD, Custom Resource Definition)
  • API service
  • Dashboard
  • CLI
  • Python SDK

Essentially, Ark is the control plane for Agents.

Ark defines seven core CRDs in Kubernetes. The following flowchart shows the relationships among these resources:

Figure 1: Ark CRD Resource Relationships
Figure 1: Ark CRD Resource Relationships

Through this set of CRDs, Ark makes Agent systems resource-oriented and declarative, enabling capabilities such as:

  • Lifecycle management
  • Multi-tenant isolation
  • RBAC (Role-Based Access Control)
  • Observability
  • Upgradability
  • Extensibility (tools, models, MCP)

In other words, Ark is not about “how to write Agents,” but “how to operate Agents in enterprise-grade systems.”

Three-Layer Architecture: Mixed Languages and Components, but a Complete System

Ark’s overall architecture is divided into three layers, each with different tech stacks and responsibilities. The following flowchart illustrates the relationships among components in each layer:

Figure 2: Ark Three-Layer Architecture Components
Figure 2: Ark Three-Layer Architecture Components

This is not a “wrapper project,” but a fully operational AI Runtime system, with a level of engineering far beyond most agent frameworks on the market.

Is It the Kubernetes of the AI Era?

Let’s revisit Kubernetes’ core value:

Kubernetes was never about “unifying cloud APIs”; it unified the “application runtime model.”

Cloud provider APIs aren’t unified, nor are networking or storage. What is unified: Pod, Deployment, Service—these application models.

Kubernetes succeeded because:

It provides a stable application abstraction on top of diversity.

Ark’s goal is not to unify all large language models (LLMs), MCPs, or tool formats, but rather:

Agent resource model (CRD) + control plane (Reconciler) + lifecycle.

From this perspective, Ark offers a prototype of a “declarative application model” for the AI era.

Whether it will become “Kubernetes for AI” is still too early to say, but it has already planted a seed.

Comparison with Other Frameworks: Not on the Same Level

Current mainstream agent frameworks like LangChain, CrewAI, AutoGen, MetaGPT, etc., address problems fundamentally different from Ark.

The table below compares the positioning and limitations of each framework:

Name What Problem Does It Solve Core Limitation
LangChain Agent/Tool composition Doesn’t address deployment or governance
AutoGen Multi-agent conversations Lacks control plane and lifecycle
CrewAI Workflow-style multi-agent Missing scheduling, RBAC, resource model
MetaGPT Agent SOP Just execution logic, not a platform
OpenDevin AI IDE/Dev Assistant Not an Agent Runtime
Ark Agent control plane + resource system Functionality not yet mature
Table 1: Mainstream Agent Frameworks vs. Ark

In short:

  • Other tools focus on “how to write Agents.”
  • Ark focuses on “how Agents run, schedule, govern, observe, and extend.”

That’s an architectural difference.

Execution Flow: Agents Scheduled Like Pods

Ark’s execution flow closely resembles the Kubernetes controller model. The following sequence diagram shows the core process:

Figure 3: Ark Agent Execution Flow
Figure 3: Ark Agent Execution Flow

You can see Ark’s process logic is transparent, with a clear engineering path, bringing agent systems into a “controllable” state.

Production Readiness: Right Direction, Still a Tech Preview

According to official notes and code maturity, Ark currently offers:

  • Runnable
  • Learnable
  • Extensible
  • But not recommended for large-scale production use yet

Main reasons include:

  • CRD structures may change
  • APIs are not yet stable
  • MCP ecosystem is still forming
  • Memory service is still basic
  • Multi-agent team execution strategies are primitive

But the engineering system is already taking shape, which is crucial.

Community Activity: Small but Elite, Strong McKinsey Drive

From GitHub data:

  • Stars: 222
  • Forks: 50
  • Contributors: 48
  • Commit frequency is steady
  • The vast majority of contributions come from within McKinsey

Note: Data as of December 2, 2025.

High stability, but limited openness.

This is also ArkSphere’s opportunity:

The paradigm is right, but the ecosystem needs community-driven growth.

Trends for 2026: From Framework Era to Runtime Era

After deep analysis, I’m increasingly convinced:

  • 2023–2024: Large model API call era
  • 2024–2025: Agent framework era
  • 2025–2027: Agent Runtime / Control Plane era (Ark’s direction)

While everyone is writing Python scripts for agents, the real value lies in:

  • Multi-agent task scheduling
  • Tool registration and governance
  • Session/Memory lifecycle
  • Result reproducibility
  • RBAC, auditing, tenant isolation
  • Observability
  • Enterprise internal personalized agent systems

Ark is providing a practical path forward.

Inspiration for ArkSphere

Ark’s inspiration for ArkSphere is both critical and direct:

ArkSphere Should Focus on “Paradigm Building,” Not “Feature Stacking”

Ark offers a prototype for future Agentic Runtime:

  • Resource model
  • Control plane
  • Tool registration
  • Multi-agent collaboration
  • Evaluation and governance

ArkSphere’s role should be:

Aggregate paradigms, produce standards, incubate ecosystems, not rewrite Ark itself.

This is the “CNCF (Cloud Native Computing Foundation) for the AI-native era.”

Huge Potential for Localization in China

Localization opportunities include but are not limited to:

  • Integration with domestic large language models (e.g., Qwen, DeepSeek, Zhipu)
  • Enterprise privatization scenarios
  • Local tool/MCP discovery ecosystem
  • Multi-cluster/edge inference
  • Enterprise-grade RBAC, auditing, data isolation
  • AgentSpec enhancements for industrial scenarios
  • Enhanced versions of Runtime/Controller

In other words:

Ark solves the “model,” while ArkSphere can solve the “ecosystem.”

What We Need Is Not “Kubernetes for the LLM Era,” But an “Industry-Grade Cognition System for AI Runtime”

The biggest takeaway from dissecting Ark:

The future of AI-native is not a pile of tools, but an engineering system.

ArkSphere can be the initiator of this system.

Summary

Ark is not a “universal runtime,” nor is it the “ultimate Kubernetes for the AI era.”

But it has done one crucial thing right:

It abstracts all the pain points people faced when writing Python agent scripts into Kubernetes resources and controllers.

It represents engineering, not just a demo.

It’s not mature yet, but it’s heading in the right direction.

It’s not the end, but it gives us a clear roadmap.

For the ArkSphere community I’m running, Ark provides a clear inspiration:

The future belongs to Runtime, to Control Plane, to governable agent systems.

And the ones who can truly scale this system are not McKinsey, but the community.

From Using AI to Relying on AI: Why the Era of AI Engineering Has Yet to Begin

2025-11-29 20:40:54

The real inflection point for AI engineering is not “how many people use it,” but “how many people cannot do without it.” Only when not using AI leads to direct loss of opportunity and efficiency, can we say the era of AI engineering has truly arrived.

Starting Point: Predictions for AI in 2026

Recently, I came across two predictions for 2026 from Amazon CTO Werner Vogels that struck me the most:

  • Renaissance Developer: Developers must span code, product, business, and social impact.
  • Personalized Learning: AI will reshape education, focusing on differentiated paths rather than a unified curriculum.

Both point to the same trend: AI is not just a tool, but is redefining how people grow and how they are defined.

There is a gap between prediction and reality, and it is worth exploring.

Correction: Will AI Really Be “Saturated” by 2026?

My initial prediction was that AI usage would reach saturation by 2026. Reality has shown me this is too optimistic.

By the end of 2025, even among internet professionals, most people’s use of AI remains at the “heard of it” or “tried it a few times” stage. It is still far from being a daily workflow necessity.

More importantly, this judgment is conditional: infrastructure supply, regulation, and compute costs must not reverse in the next 3–6 years. If any variable breaks down (costs double, models go offline, policy shifts), the adoption curve will be disrupted.

The Truth About the Inflection Point: From “Using” to “Relying On”

“Relying on” is a vague term. A more precise definition requires measurable indicators.

Here is a diagram that visualizes the metrics for being truly dependent on AI:

Figure 1: Quantitative Definition of AI Dependency
Figure 1: Quantitative Definition of AI Dependency

Most industries have not reached the “cannot operate without” stage, unlike the internet, mobile, or payment inflection points. Most metrics are still far below the threshold, which is why the most likely outcome for 2026 is: more people will use AI, but those who truly rely on it will remain a minority.

Using ≠ Building: The Five-Level Capability Ladder

This difference is not binary, but a clear progression.

The following table shows the five-level model of AI capability maturity.

Level Name Description Scarcity
1 Tool User ChatGPT/Claude, Coding, Copywriting, Accelerator, Optional Low
2 Integrator LLM API + Vector DB, AI layered on existing systems, Usable, not critical Low
3 Settler Restructuring data flow, business decisions, AI becomes critical path Rising
4 Engineering Abstraction Extracting frameworks, runtimes, providing infra for ecosystem Extremely High
5 Autonomous System Self-feedback, self-optimizing, redefining human-AI relationship Future
Table 1: Five-Level Model of AI Capability Maturity

Currently, the biggest gap is at Level 3 and Level 4. Most people are stuck at Level 1 or 2, with very few reaching Level 4. This means high-value scarcity will not disappear, but will continue to rise.

Why the Era of AI Engineering Has Not Arrived: Three-Dimensional Delaying Factors

It is not technology alone that is holding things back, but constraints in three dimensions.

The following diagram illustrates the three main constraints delaying AI engineering maturity:

Figure 2: Three-Dimensional Constraints on AI Engineering Maturity
Figure 2: Three-Dimensional Constraints on AI Engineering Maturity

The key observation: If any one dimension is stuck, the entire ecosystem’s maturity will be delayed. Currently, none of the three dimensions have fully mature solutions.

The Realistic Window: Three Paths for Capability Advancement

The next three years will not be “winner takes all,” but rather a period where multiple capability levels appreciate simultaneously.

Below is a table comparing the value and bottlenecks of different capability advancement paths:

Capability Path Short-Term Value Long-Term Outlook Bottleneck
Level 1→2 (Tool→Integration) ⭐⭐ Rapid Depreciation ⭐ Saturation Low barrier, fierce competition
Level 2→3 (Integration→Settlement) ⭐⭐⭐⭐ Scarce ⭐⭐⭐⭐ Continual Appreciation Requires industry depth, long-term iteration
Level 3→4 (Settlement→Abstraction) ⭐⭐⭐⭐⭐ Extremely Scarce ⭐⭐⭐⭐⭐ Defines Ecosystem Large cognitive leap, needs community influence
Table 2: AI Capability Advancement Paths and Value Comparison

Key conclusion: While the number of “AI users” is rapidly increasing (depressing Level 1 value), due to the three-dimensional delaying factors, scarcity at Level 3 and 4 will only rise.

What I’m Doing on arksphere.dev

Based on the above judgment, I focus on exploring the architectural evolution of AI Native infrastructure. The goal is not to catalog model usage, but to study the foundational capability stack supporting scalable intelligent systems: scheduling, storage, inference, Agent Runtime, autonomous control, observability, and reliability.

The content is no longer a collection of courses or tips, but a continuous record of evolution around Infra → Runtime → System Abstraction. arksphere.dev is the site for this experiment and settlement.

Summary

The inflection point for the era of AI engineering is not “how many people use it,” but “how many people cannot do without it.” The latter requires five measurable indicators to reach their thresholds, and we are still far from that.

“Using ≠ Building” is not a binary, but a five-level progression. Scarcity at Level 3 and 4 will rise as the number of Level 1 users increases—this is the biggest opportunity window in the next three years.

But the width of this window depends largely on how technology, institutions, and organizations evolve together. I hope more people working on AI engineering will not only focus on technical innovation, but also invest equal thought into institutional development, talent growth, and risk governance—these “invisible engineering” challenges.

Turning Antigravity Into a VS Code-Style AI IDE

2025-11-20 11:55:30

The biggest pain point when switching IDEs is user habits. By installing a series of plugins and tweaking configurations, you can make Antigravity feel much more like VS Code—preserving familiar workflows while adding Open Agent Manager capabilities.

I installed Antigravity on day one of its release. After a few days of use, my main impression is: it feels more like an “Agent console” than a traditional integrated development environment (IDE). Still, I’m used to the VS Code interface and plugin ecosystem, so I spent some time tuning Antigravity to become a “VS Code-style AI IDE”.

Figure 1: Antigravity IDE UI
Figure 1: Antigravity IDE UI

Below are the configurations and steps I actually use. Feel free to follow along.

First Impressions of Antigravity

A few subjective observations:

  • The interface is split between agent management and editor views, somewhat like AgentHQ + VS Code.
  • Agents modify code very quickly, with a much higher completion rate than typical “chat-based” assistants.
  • The editor and context windows are large, ideal for long diffs and logs.
  • By default, it uses OpenVSX / OpenVSCode Gallery, so the extension ecosystem isn’t identical to my VS Code setup.

All subsequent steps focus on one goal: keep Antigravity’s agent features while maintaining my VS Code workflow.

Switching the Extension Marketplace to VS Code Official

Antigravity is essentially a VS Code fork, so you can directly change the Marketplace configuration.

In Antigravity:

Go to Settings -> Antigravity Settings -> Editor, and update the following URLs to point to VS Code:

Marketplace Item URL:

https://marketplace.visualstudio.com/items

Marketplace Gallery URL:

https://marketplace.visualstudio.com/_apis/public/gallery
Figure 2: VSCode Marketplace Configuration
Figure 2: VSCode Marketplace Configuration

Restart Antigravity.

After this change, searching and installing extensions works just like the official VS Code Marketplace. Installing AMP, GitHub Theme, VS Code Icon, etc., all follow this process.

Installing the AMP Extension

AMP isn’t officially supported on Antigravity yet, but you can install it directly via the VS Code Marketplace.

Steps:

  1. Open the Extensions panel (the same icon as in VS Code).
  2. Search for the AMP extension and install it as usual.
  3. Log in using your AMP API Key.

Currently, Antigravity doesn’t support one-click account login like VS Code; you have to use the API key.

Summary
Once installed, AMP works almost identically in Antigravity as in VS Code—completion and refactoring features are available. The only difference is manual login configuration.

I recommend AMP because it offers a free mode. In my experience, it’s great for writing documentation, running scripts, and as a daily command-line tool. It’s fast, and especially useful for optimizing prompts.

Importing the CodeX Extension

CodeX doesn’t provide a direct VSIX download link on the web. My approach is to export it from VS Code and then import it into Antigravity.

Figure 3: Exporting Codex Extension in VS Code
Figure 3: Exporting Codex Extension in VS Code

Steps:

  1. Install the CodeX extension in VS Code (if you haven’t already).
  2. In VS Code’s extension manager, find CodeX and export it as a .vsix file.
  3. Switch to Antigravity, open the Extensions panel, and select “Install from VSIX”.
  4. Choose the exported codex-x.x.x.vsix file to complete installation.
Tip
Since my local VS Code is already logged into CodeX, importing it into Antigravity automatically reuses the login state—I didn’t need to log in again.

Optimizing Editor Settings

Beyond the marketplace and plugins, a few tweaks make the experience even closer to VS Code:

  • Theme: Choose the same color scheme as VS Code to minimize visual switching. I use GitHub Theme and vscode-icons.
  • Editor Settings: In “Open Editor Settings”, set indentation, formatting, line width, etc., to match your VS Code preferences. I define these in the workspace’s settings.json, so no migration is needed.

After these changes, the editing area is essentially “VS Code with an agent console”.

Remaining Issues

To fully migrate from VS Code/GitHub Copilot to Antigravity, I think there are still several key challenges:

  • Limited Customization: Antigravity can’t support custom prompts and agents like Copilot Chat. Currently, only “rules” configuration is available, which limits workflow flexibility.
  • Model Ecosystem Needs Improvement: Antigravity hasn’t natively integrated the latest models from major vendors (OpenAI, Anthropic, Microsoft, xAI, etc.), whereas GitHub Copilot excels here.
  • Cost Considerations:
    • Future pricing may start at $20/month.
    • No free models are supported, unlike GitHub Copilot (even Copilot Pro users have free model options).
  • Stability Issues: Agents often encounter “Agent terminated due to error” during operation, requiring manual retries or new sessions. This affects workflow smoothness, though I expect improvements in the future.

GitHub Copilot VS. Antigravity

Although Antigravity excels in several areas, there is still significant room for improvement compared to the combination of GitHub Copilot and VS Code.

The large language models (LLMs, Large Language Models) I frequently use are all supported in VS Code:

Figure 4: Copilot-supported LLMs (partial)
Figure 4: Copilot-supported LLMs (partial)

My long-accumulated custom prompts:

Figure 5: Copilot Chat enables quick access to custom prompts
Figure 5: Copilot Chat enables quick access to custom prompts

My collection of agents:

Figure 6: Copilot Chat allows selection of custom agents
Figure 6: Copilot Chat allows selection of custom agents

Here are some personal experiences using VS Code and Copilot that, for now, are hard to replace with other IDEs:

  • The Ask/Edit/Agent/Plan workflow perfectly fits my working habits.
  • Support for custom prompts and agents is essential. Many of my prompts and agents have been refined over time and are deeply integrated into my daily workflow—it’s hard to find alternatives elsewhere.
  • New models are integrated at lightning speed. Whenever a new model is released, GitHub Copilot is among the first to support it.
  • The integration with VS Code is seamless—no extra configuration required, making it extremely convenient.
  • Frequent updates: just a few days ago, a bug I reported to VS Code was fixed the same night.
  • Copilot Chat’s keyboard shortcuts make it easy to quickly access various features.
  • GitHub has granted me a free Pro account. Although the monthly premium quota is only 300 calls, combining Copilot with other plugins like AMP, Codex, Droid, and Qwen enables a highly efficient workflow. Even if I upgrade to a paid account in the future, the $10/month fee is very cost-effective compared to similar products.

Practical Experience

A few subjective tips from my actual usage—take them as reference:

  • Don’t treat Antigravity as “VS Code + chat box”. Use its agent features for complete tasks: let the agent propose a plan, then execute changes.
  • For major changes, always create a new Git branch and restrict agent actions to that branch. Handle all diffs via standard Pull Request (PR) workflows.
  • Ask agents to produce “artifacts” (plans, proposals, test descriptions), not just final code. This makes it easier to review and track changes.
  • Plugins you’re already comfortable with in VS Code (like AMP, CodeX) can be migrated directly, reducing cognitive load and letting you focus on new agent workflows.

Summary

My current experience: Antigravity delivers powerful agent capabilities and multi-view consoles. By following these steps to align the interface and plugin ecosystem with VS Code, you can smoothly transition your daily development workflow.

Cloudflare November 18 Global Outage: The Dangers of Implicit Assumptions in Modern Infrastructure

2025-11-19 18:56:34

The greatest risks to modern internet infrastructure often aren’t in the code itself, but in those implicit assumptions and automated configuration pipelines that go undefined. Cloudflare’s outage is a wake-up call every Infra/AI engineer must heed.

Yesterday (November 18), Cloudflare experienced its largest global outage since 2019. As this site is hosted on Cloudflare, it was also affected—one of the rare times in eight years that the site was inaccessible due to an outage (the last time was a GitHub Pages failure, which happened the year Microsoft acquired GitHub).

Figure 1: jimmysong.io was down for 27 minutes due to the Cloudflare outage
Figure 1: jimmysong.io was down for 27 minutes due to the Cloudflare outage

This incident was not caused by an attack or a traditional software bug, but by a seemingly “safe” permissions update that triggered the weakest link in modern infrastructure: implicit assumptions (Implicit Assumption) and automated configuration pipelines (Automated Configuration Pipeline). Cloudflare has published a blog post Cloudflare outage on November 18, 2025 explaining the cause.

Here is the chain reaction process of the outage:

  • A permissions adjustment led to metadata changes;
  • The metadata change doubled the lines in the feature file;
  • The doubled lines triggered the proxy module’s memory limit;
  • The memory limit caused the core proxy to panic;
  • The proxy panic led to a cascade failure in downstream systems.

This kind of chain reaction is the most typical—and dangerous—systemic failure mode at today’s internet scale.

Root Cause: Implicit Assumptions Are Not Contracts

Let’s first look at the core hidden risk in this incident. The Bot Management feature file is automatically generated every five minutes, relying on a default premise:

The system.columns query result contains only the default database.

This assumption was not documented or validated in configuration—it existed only in the engineer’s mental model.

After a ClickHouse permissions update, the underlying r0 tables were exposed, instantly doubling the query results. The file size exceeded the FL2 preset of 200 features in memory, ultimately causing a panic.

Once an implicit assumption is broken, the system lacks a buffer and is highly prone to cascading failures.

Configuration Pipelines Are Riskier Than Code Pipelines

This incident was not caused by code changes, but by data-plane changes:

  • SQL query behavior changed;
  • Feature files were automatically generated;
  • The files were broadcast across the network.

A typical phenomenon in modern infrastructure: data, schema, and metadata are far more likely to destabilize systems than code.

Cloudflare’s feature file is a “supply chain input,” not a regular configuration. Anything entering the automated broadcast path is equivalent to a system-level command.

Language Safety Can’t Eliminate Boundary Layer Complexity

A former Cloudflare engineer summarized it well:

Rust can prevent a class of errors, but the complexity of boundary layers, data contracts, and configuration pipelines does not disappear.

The FL2 panic stemmed from a single unwrap(). This isn’t a language issue, but a lack of system contracts:

  • No upper-bound validation for feature count;
  • File schema lacked version constraints;
  • Feature generation logic depended on implicit behavior;
  • Core proxy error mode was panic, not graceful degradation.

Most incidents in modern distributed systems (Distributed System) come from “bad input,” not “bad memory.”

Core Proxies Need Controllable Failure Paths

FL/FL2 are Cloudflare’s core proxies; all requests must pass through them. Such components should not fail with a panic, but have the following capabilities:

  • Ignore abnormal features;
  • Truncate over-limit fields;
  • Roll back to previous versions;
  • Fail-open or fail-close;
  • Skip the Bot module and continue processing traffic.

As long as the proxy “stays alive,” the entire network won’t be completely paralyzed.

Data Changes Are More Uncontrollable Than Code Changes

The essence of this incident:

  • Subtle permission changes;
  • ClickHouse default behavior changed;
  • Query results propagated to distributed systems;
  • Automated publishing amplified the error;
  • Edge proxies crashed due to uncontrolled input.

Future AI Infra (AI Infrastructure) will be even more complex: models, tokenizers, adapters, RAG indexes, and KV snapshots all require frequent updates.

In future AI infrastructure, data-plane risks will far exceed those of the code-plane.

Recovery Process Shows Engineering Maturity

During the incident, Cloudflare took several measures:

  • Stopped generating erroneous feature files;
  • Force-distributed the previous version of the file;
  • Rolled back Bot module configuration;
  • Ran Workers KV and Access outside the core proxy;
  • Restored traffic in stages.

Restoring hundreds of PoPs worldwide simultaneously demonstrates a high level of engineering maturity.

Lessons for Infra/AI/Cloud Native Engineers

The Cloudflare event highlights four common risks in large-scale systems:

  • Implicit assumptions fail;
  • Configuration supply chain contamination;
  • Automated publishing amplifies errors;
  • Core proxies lack graceful degradation paths.

For AI Infra practitioners, these risks are even more relevant:

  • Model weight updates without schema validation;
  • Adapter merges may be contaminated;
  • RAG index incremental builds are unstable;
  • Inference graph configuration may be broken by bad data;
  • Automatically rolled-out models may propagate errors network-wide.

AI engineering is replaying Cloudflare’s infrastructure dilemmas—just at greater speed and scale.

Summary of Former Cloudflare Engineer’s Views

His insights pinpoint the hardest problems in distributed systems:

  • The issue isn’t code, but missing contracts;
  • Not the language, but undefined input boundaries;
  • Not modules, but lack of validation in the configuration supply chain;
  • Not bugs, but absence of fail-safe mechanisms.

This incident proves: The real fragility in modern infrastructure lies in “behavioral boundaries,” not “memory boundaries.”

Summary

The Cloudflare November 18 outage was not a coincidence, but an inevitable result of modern internet infrastructure evolving to large-scale, highly automated stages.

Key takeaways from this event:

  • System assumptions must be made explicit;
  • Configuration pipelines must be validated;
  • Automated publishing needs “dead-end” mechanisms;
  • Core proxies must be designed with controllable failure paths;
  • Data-plane contracts must be stricter than code-plane contracts.

In the AI-native Infra era, these requirements will only become more stringent.

References