Did Build vs. Buy Just Flip?

Claude Opus 4.6 dropped on February 5th with 1M token context. GPT-5.3-Codex released last week running 25% faster than 5.2. Your vendors are calling with upgraded demos. Your engineers are asking what this means for the projects you shelved six months ago.

Here's what it means: the build vs buy calculus is shifting faster than most CTOs realize.

Companies are paying six figures annually for vendor AI platforms that only solve 80% of their requirements. Their engineers are building custom workarounds for the critical 20%—the part where actual business logic lives. Now with Opus 4.6's context window and Codex's speed improvements, those engineers are asking: "Why are we paying for the vendor platform?"

The Shift Happening Right Now

Claude Opus 4.6 released on February 5th with 1M token context. GPT-5.3-Codex dropped last week running 25% faster than 5.2. These aren't incremental improvements. They're changing the build vs buy calculus faster than most CTOs realize.

Your CFO approved a six-figure annual contract for an AI platform six months ago. The vendor demo looked perfect. Now your engineers are writing custom code to handle edge cases. They're asking a new question: "With Opus 4.6's context window and Codex's speed improvements, why are we paying for this platform?"

This is the 80/20 problem. AI vendors solve 80% of requirements. The missing 20% isn't random features. It's exactly where your business logic lives, where your competitive advantage matters, and where your operational reality differs from generic assumptions.

The build vs buy decision for enterprise AI now has three dimensions, not two: capability, data sovereignty, and build economics. Here's how to think through this decision with engineering rigor instead of vendor promises.


What Changed This Month

Context Windows That Actually Matter

Opus 4.6's 1M tokens means special-ops teams can load entire codebases, legacy system documentation, and business rule sets into a single session. Projects that needed six months in 2024 are getting scoped at four weeks in 2026. We're deploying teams for two Fortune 500 clients in the next ten days based on what this context window enables.

Speed Improvements That Compound

Codex’s 25% speed improvement changes the math.Codex's 25% speed improvement isn't incremental. What took six weeks takes four weeks. What took four weeks takes three weeks. The "buy because custom is too expensive" argument is collapsing. Custom builds for business-specific needs are suddenly cost-effective.

The Data Sovereignty Problem

Simultaneously, broad AI tools like Copilot and Cowork are delivering real productivity gains. Teams are integrating them for general tasks. But they're hitting data boundaries. Proprietary pricing algorithms, competitive intelligence, pre-launch product data—CIOs are grappling with how to capture AI’s benefits without exposing sensitive data to competitors using the same tools.CIOs are wrestling with how to get AI benefits without exposing what competitors using the same tools might see.

This creates a three-dimensional decision that most enterprise AI frameworks aren't addressing yet.


When Off-the-Shelf Tools Actually Work

Before dismissing vendor solutions, understand when they genuinely deliver value. Off-the-shelf AI tools work when your requirements match standard patterns that most other companies share.

Document processing with common formats. Invoice extraction, receipt processing, contract parsing. If your documents follow standard formats and your validation rules are straightforward, tools like Azure Form Recognizer or AWS Textract handle this reliably. You're not competing on invoice processing capability.

Standard chatbots and Q&A. Customer support bots answering FAQ-style questions from documented knowledge bases. If your differentiation comes from response time rather than AI sophistication, configurable platforms like Intercom or Zendesk AI work fine.

Basic recommendation engines. "Customers who bought X also bought Y" recommendations without complex business rules around margin, inventory, or strategic positioning.

The vendor advantage: infrastructure management, speed to baseline value, and handling commodity capabilities so your team can focus where you actually compete.


When Off-the-Shelf Tools Fail (And Why It Matters)

The limitations surface the moment your requirements include "except when" or "but in our case." Here's where generic tools break down.

Legacy System Integration

Enterprise reality: systems running for 15 years. Your ERP was customized in 2009. Your warehouse management system has undocumented business logic in stored procedures. Your pricing engine integrates with six different data sources using proprietary APIs.

Vendor solutions integrate with common platforms. They don't integrate with your specific implementations. Building integration layers around vendor tools often costs more than building the AI capability itself.

Domain-Specific Constraints

A logistics company evaluated route optimization tools. Every vendor demoed impressive algorithms. None understood their reality: time-dependent road restrictions changing by day of week, driver certifications limiting route assignments, individually-negotiated customer delivery windows, and vehicle capacity constraints varying by product type.

The vendors could optimize routes. They couldn't optimize routes given the actual constraints the business operates under. That's not a 20% gap. That's the entire problem.

Complex Business Rules

Your business logic reflects years of learned patterns, regulatory requirements, negotiated contracts, and strategic decisions. A pricing system might have hundreds of rules: volume discounts for specific customers, promotional pricing with expiration logic, competitive matching for strategic accounts, margin floors varying by product category, seasonal adjustments based on inventory levels.

Generic tools provide frameworks for configuration. That configuration work becomes a permanent development and maintenance burden.

Real Examples

Manufacturing client: "The inventory optimization tool looks great, but in our case we have minimum order quantities from suppliers that vary by season, contractual commitments creating inventory floors, and production line changeover costs making small batches uneconomical."

E-commerce client: "The recommendation engine works well, but in our case we need to balance gross margin, inventory age, strategic brand positioning, and vendor co-op marketing commitments."

Every "but in our case" represents a gap between generic capability and business-specific requirements. When these gaps are central to business operations, the vendor tool becomes expensive infrastructure that doesn't solve your actual problem.


The Data Sovereignty Dimension

Beyond capability gaps, February 2026 brings a consideration that didn't exist when most AI build vs buy frameworks were written: data boundaries.

When Broad AI Tools See Your Data

Microsoft Copilot, Claude Cowork, ChatGPT Enterprise deliver immediate value. Teams are analyzing data, generating reports, writing code, and automating workflows using AI assistants that understand natural language.

The tradeoff: these systems process your data through their models. For general productivity tasks, that's acceptable. For some use cases, it's problematic.

When Data Sovereignty Matters

You need to keep data inside your boundaries when working with:

  • Proprietary business logic representing competitive advantage (pricing algorithms, optimization parameters, strategic decision criteria)
  • Competitive intelligence that you don't want feeding back into models your competitors might use
  • Regulated data with compliance requirements (healthcare, financial services, government contracts)
  • Pre-public information that can't be exposed to external systems (M&A plans, product launches, strategic initiatives)

The Architecture Question

Most enterprises don't need custom-trained models. You can build effective AI systems using commercial models (GPT-4, Claude, Llama) with your data through RAG architectures, fine-tuning, or prompt engineering.

The architecture matters more than model ownership:

DEFINE DataSovereignArchitecture:

   // Commercial models run in YOUR infrastructure
   model_deployment: {
       location: your_private_cloud or on_premise
       data_flow: never_leaves_your_boundary
       model: commercial_llm_licensed_for_private_deployment
   }

   // Your proprietary data stays internal
   context_data: {
       stored_in: your_databases
       accessed_by: ai_system_in_your_infrastructure
       never_sent_to: external_apis
   }

   // Control plane separates general vs. sensitive
   routing_logic: {
       general_tasks: can_use_external_apis
       sensitive_tasks: routed_to_internal_systems
   }

You can use Claude or GPT-4 for general tasks while routing sensitive operations through models deployed in your infrastructure.


The Real Cost Comparison

The sticker price comparison misleads. A $50K/month SaaS platform looks cheaper than $300K in custom development. Until you account for the full cost.

Vendor Tool Total Cost

Annual platform fee plus implementation services plus ongoing configuration changes plus custom integration development plus workaround systems for unsupported requirements plus the opportunity cost of operating with suboptimal AI.

Client example: Manufacturing company committed to $600K annual contract for inventory optimization. Added $200K in integration development. Discovered the tool didn't handle their constraint-based optimization needs. Built a parallelBuilt parallel custom system for $400K. Now paying for both. Total three-year cost: $2.2M. Still doesn't have optimal inventory decisions.

Custom Development Total Cost

Initial development plus ongoing maintenance plus infrastructure costs. With Opus 4.6's context window and Codex's 25% speed improvement, these numbers are changing.

Same manufacturing client, alternative path evaluated this week: Custom development scoped at $380K initial build (down from $450K six months ago), $15K/month maintenance, owned infrastructure at $5K/month. Three-year total: $1.1M. Delivers AI that actually optimizes for their specific constraints.

The build economics shifted dramatically. What seemed too expensive six months ago is now the cost-effective path.


The Decision Framework

Four Paths Taking Shape

The decision now maps to four distinct paths based on capability specificity and data sensitivity:

Path 1: Generic Tasks, Non-Sensitive Data Teams are using broad AI tools (Copilot, Cowork, ChatGPT Enterprise). Document processing, data analysis, general coding assistance. Fast deployment, immediate value.

Path 2: Generic Tasks, Sensitive Data Companies are deploying commercial models in private clouds. Same capabilities as broad tools, but data stays inside boundaries. Azure OpenAI in your VPC, Claude in your infrastructure.

Path 3: Business-Specific Needs, Non-Sensitive Data Custom builds are suddenly cost-effective. With current AI coding tools, what took six months takes four weeks. Special-ops teams (2-4 engineers, 4-6 weeks) delivering purpose-built AI for department-level needs.

Path 4: Business-Specific Needs, Sensitive Data This is where competitive advantage lives. Custom systems with full data sovereignty. Proprietary business logic that can't be exposed to external models. We're deploying teams for this right now—inventory optimization for manufacturing, pricing engines for e-commerce, supply chain intelligence.

The Hybrid Approach

The smart path often combines all four. Use broad AI tools for general productivity. Deploy commercial models privately for sensitive general tasks. Build custom for business-specific capabilities.

DEFINE HybridAIArchitecture:

   // Generic, non-sensitive - use broad AI tools
   general_productivity: copilot_or_cowork

   // Generic, sensitive - deploy commercially
   document_analysis: azure_openai_in_private_cloud

   // Business-specific, non-sensitive - build custom
   workflow_automation: special_ops_team_build

   // Business-specific, sensitive - build custom with sovereignty
   pricing_optimization: custom_ai {
       runs_in: your_infrastructure
       encodes: proprietary_business_rules
       optimizes: actual_business_outcomes
       data: never_leaves_boundaries
   }


Real Client Decisions

Manufacturing: Built Custom

Client needed inventory optimization across 12 factories, 50,000 SKUs, complex supplier relationships. Evaluated three vendor platforms in Q4 2025. All provided generic optimization. None handled their constraint-based reality.

Decision: Build custom system integrating with their ERP, encoding their specific business rules.

Outcome: $4.2M reduction in excess inventory within first year. System paid for itself in six months.

E-Commerce: Hybrid Approach

Client needed product recommendations balancing conversion, margin, inventory age, and brand strategy. Used vendor tool (Algolia) for basic search. Built custom recommendation engine for personalized suggestions optimizing business outcomes, not just clicks.

Outcome: 18% increase in average order value while reducing aged inventory by 30%.

Logistics: Bought and Configured

Client needed route optimization for standard delivery operations. Requirements were common. No unusual constraints.

Decision: Bought Descartes routing platform, configured for their operations.

Outcome: 12% reduction in delivery costs, implemented in 6 weeks.


The Question Changed

The build vs buy decision for enterprise AI in February 2026 isn't a single question. It's three:

  1. Where does your competitive advantage live? If it's in generic capabilities, use existing tools. If it's in business-specific AI understanding your unique context, build custom.
  2. Does your data require sovereignty? If you're processing sensitive, proprietary, or regulated information, you need systems keeping data inside your boundaries.
  3. What can you actually build now? With Opus 4.6's context window and Codex's speed improvements, custom solutions that took six months last year take four weeks this month.

76% of enterprises are still buying vendor solutions. That number is about to shift. The calculus changed. Data sovereignty matters now. Build economics improved dramatically. The four-path evaluation reveals different answers than the old binary choice.

What matters: making the decision with engineering rigor, not vendor promises or outdated assumptions about what custom development costs.

Reflection: 
CREATE IT WITH INFORM.
Continue reading