News listOpenAI teaches you how to use Codex safely: sandbox boundaries, automated approval, security classification, and a complete enterprise deployment framework
動區 BlockTempo2026-05-10 07:10:27

OpenAI teaches you how to use Codex safely: sandbox boundaries, automated approval, security classification, and a complete enterprise deployment framework

ORIGINALOpenAI 教你如何安全使用 Codex:沙盒邊界、自動審批、安全分類,完整企業部署框架
AI Impact AnalysisGrok analyzing...
📄Full Article· Automatically extracted by trafilaturaGemini 翻譯1959 words
OpenAI has disclosed how it securely deploys its internal AI coding agent, Codex. The core strategy involves "sandboxed execution boundaries + automated approval for low-risk actions + AI security classification agents for alert handling," enabling development efficiency and enterprise security control to operate in sync. (Context: OpenAI Codex major upgrade: background Mac control, built-in browser, image generation, 111 new plugins launched) (Background: OpenAI launches new engineer agent Codex! AI can write features, fix bugs, and run tests) This report explains how the internal security team executes Codex in production environments. It is a practical operational record, from sandbox configuration to alert classification, revealing the layers of security control required when AI agents are adopted by large organizations. OpenAI published an internal enterprise deployment report this week. In its official announcement, OpenAI stated that there is only one core principle for deploying Codex: keep the agent efficient within clear technical boundaries; low-risk actions do not require interrupting the user, while high-risk actions must be paused for human review. Once this principle was implemented, it was broken down into two complementary mechanisms: sandboxing and approval policies. The sandbox defines the execution space for Codex, including which paths can be written to, whether external network access is allowed, and which system directories are protected. Actions outside the sandbox must enter the approval process. Users can approve a specific operation once, or approve that type of operation to be automatically permitted throughout the entire session. For routine daily operations, OpenAI enabled "Auto-review mode." This feature sends the actions Codex plans to execute and the recent operational context to an "automated approval sub-agent." If the sub-agent determines the risk is low, it proceeds directly without interrupting the user's workflow; if it determines the risk is high or has potential unintended consequences, it escalates to human confirmation. The logic for network control is the same. Codex does not have open external access; OpenAI maintains an allowlist of target domains required for Codex's normal workflows. Domains outside this list are blocked by default, and encountering an unfamiliar domain triggers the approval process. Identity authentication is also included in the control scope. CLI and MCP OAuth credentials are stored in the operating system's secure keychain, and login is forced through the ChatGPT Enterprise workspace. Consequently, Codex operations are integrated into the ChatGPT Enterprise compliance logging platform, allowing the security team to review them centrally. OpenAI does not treat all Shell commands as having equal risk, but has instead established a layered rule set. Harmless commands common in daily engineering development are allowed to execute directly outside the sandbox without approval. Specific high-risk commands are directly blocked or forced into the approval process. This rule set is enforced through three overlapping layers: - Cloud management requirements (enforced by administrators, cannot be overridden by users) - macOS managed preferences - Local requirement configuration files This architecture allows OpenAI to maintain a unified baseline across the company while testing different configuration combinations according to team, user group, or environment needs. The same settings apply to all local interfaces, including the Codex desktop application, CLI, and IDE extensions. This stands in direct contrast to external research: studies show that AI-generated code has a 57% higher rate of security vulnerabilities compared to human-written code. GitHub Copilot was also revealed this year to have a critical CVSS 9.6 vulnerability (CVE-2025-53773), which allows for remote code execution via prompt injection. These data points indicate that when enterprises adopt AI coding agents, the attack surface will exceed expectations without corresponding layered controls. OpenAI's approach is to embed control logic into the configuration layer rather than relying on the AI agent's own judgment, making rule enforcement a technical fact rather than an operational habit. OpenAI emphasized in its official announcement that no matter how well security controls are implemented, visibility is still required after deployment. Traditional security logs can answer "what happened," such as a program starting, a file being modified, or a network connection being attempted. However, what security personnel truly need to know is "why Codex did this" and "was this the user's original intent." OpenAI enables Codex to support OpenTelemetry log output, with a recording scope that includes: user prompts, tool approval decisions, tool execution results, MCP server usage, and network proxy allow/deny events. Enterprise and educational institution customers can also access these logs through the OpenAI compliance platform. More crucially, OpenAI internally connects these logs to an "AI security classification agent." When endpoint detection tools discover suspicious Codex behavior and trigger an alert, this AI classification agent automatically retrieves relevant Codex logs, reconstructs the original request, tool activity, approval decisions, tool results, and network policy records, and generates an analysis report for the security team to review. This assists in determining whether it is normal agent behavior, a harmless user error, or an incident that truly requires escalation. The same telemetry data is also used for
Data Status✓ Full text extractedRead Original (動區 BlockTempo)
🔍Historical Similar Events· Keyword + Asset Matching6 items
💡 Currently matching via keywords + symbols (MVP) · Will be upgraded to embedding semantic search later
Raw Information
ID:4dc0db756e
Source:動區 BlockTempo
Published:2026-05-10 07:10:27
Category:zh_news · Export Category zh
Symbols:Unspecified
Community Votes:+0 /0 · ⭐ 0 Important · 💬 0 Comments
OpenAI teaches you how to use Codex safely: sandbox boundaries, automated approval, security classification, and a complete enterprise deployment framework | Feel.Trading