Signal Hub logoSignal Hub

Tech News news and articles

Tech News
Updated just now
Date
Source

6 articles

VentureBeat
~7 min readMay 6, 2026

Market research is too slow for the AI era, so Brox built 60,000 identical 'digital twins' of real people you can survey instantly, repeatedly

In a world where a viral TikTok video can cause a brand to trend globally in mere hours, the traditional market research cycle — often spanning 12 weeks — is becoming a liability. The lag between a survey question and the answers from a wide (or targeted) pool of respondents has become a primary bottleneck for Fortune 500 decision-makers who are forced to navigate volatile geopolitical and economic shifts with data that is frequently outdated by the time it reaches a slide deck, as industry experts have observed. Brox, a predictive human intelligence startup, recently announced a strategic funding round following a year where they reported 10X revenue growth. Their proposition is as ambitious as it is technical: the creation of a "parallel universe" populated by 60,000 digital twins of real, living human beings and their entire demographic profiles and consumer preferences, allowing enterprises to run unlimited experiments in hours rather than months. “These digital twins are one-to-one replicas of actual, real individuals," said Brox CEO Hamish Brocklebank in a recent video call interview with VentureBeat. "We recruit real people like a normal panel company does, pay them to interview them, and capture all the data around them — fully consent-driven.” The company, currently a lean 14-person operation, is positioning itself as the antithesis of the "insane" research industry. By replacing statistical models with behavioral replicas, Brox aims to transform how the world’s largest banks and pharmaceutical giants anticipate human reactions to high-stakes global and market-shifting events, or narrow, targeted product releases and personnel news, and everything in between. The kinds of surveys and specific questions that Brox asks its digital twins are completely open-ended and can be customized to fit any conceivable business customer's use cases and goals. According to Brocklebank, examples of survey questions include: “What happens if America invades Iran or Greenland? Will depositors at Bank of America put more money into their account or take more money out? Or, in pharmaceuticals, if RFK Jr. says something next week, will that make people more likely to take vaccines or less likely?” Not synthetic people — AI copies of real ones The core differentiator of Brox’s technology lies in the fidelity of its input data. While many competitors in the "digital audience" space rely on purely synthetic identities — generic personas generated by Large Language Models (LLMs ) — Brocklebank argues that these methods inevitably produce "AI slop". Purely synthetic audiences often cluster around a tight distribution of answers, over-indexing for "correct" or "healthy" behaviors (such as eating broccoli) because of inherent biases in the underlying models. Brox’s "Digital Twins" are instead one-to-one behavioral replicas of real individuals who have been recruited and interviewed with exhaustive depth. The process is intensive: Deep Interviews: The company conducts hours of real and AI-driven interviews with each participant. Psychological Depth: The data collection seeks to understand fundamental "decision drivers," including upbringing, relationships, and even marital stability. Data Density: For some twins, Brox maintains up to 300 pages of text data, representing what Brocklebank calls "the deepest per person data set that exists". To solve the "black box" problem common in AI, Brox utilizes a "reasoning chain" for its predictive outputs. When a digital twin predicts a reaction — such as how a $2 billion net-worth individual might respond to a specific interest rate hike — the model introspects and provides a step-by-step explanation for that decision. This allows clients to understand not just what will happen, but the underlying psychology of why it is happening. Scaling the "unscalable" interview The product offering is currently live in the US, UK, Japan, and Turkey. Brox has successfully digitized specific, high-value cohorts that are traditionally difficult for researchers to access. This includes a panel of "high-net-worth" individuals (those worth over $5 million) and specialized medical professionals like dermatologists — including a multibillionaire. However, the largest value for customers is likely in the aggregate mass of all individuals that can be polled en masse and/or segmented across demographics, especially those of medium and lower income levels, whose purchasing power and decision-making is more constrained and whose market- One of the more unique aspects of the Brox platform is its incentive structure. To ensure twins remain up-to-date, real-world counterparts are re-contacted frequently. For high-value individuals who are not motivated by small cash payments, Brox has issued Stock Appreciation Rights (SARs), essentially making these participants "investors" in the company’s success to ensure they continue to provide high-fidelity personal updates. The platform’s use cases currently focus on two primary sectors: Pharmaceuticals: Predicting vaccine hesitancy or how physicians might react to new biologics based on shifting political climates. Finance: Simulating how depositors at major banks might move funds in response to geopolitical events, such as conflicts in the Middle East. As for why go to the trouble of interviewing and digitally cloning real people instead of just creating wholly fictitious, synthetic audience characters and personas using LLMs and other AI models, Brocklebank offered his perspective. “You can create 10,000 truly synthetic digital twins, but the answers will still normalize into a very tight distribution, which is not realistic when you’re actually asking real people," Brocklebank said. By maintaining a pre-built audience of 60,000 twins, the company enables clients to bypass the recruitment phase of research. A large US bank or a global pharma giant can now "query" the digital population and receive a validated analysis in a matter of hours. Pricing and accessibility Unlike traditional research firms that charge on a per-project or per-respondent basis, Brox operates as a high-end Software-as-a-Service (SaaS) platform with enterprise-level commercial licensing. The company avoids the "seat" or "usage" limits that often hinder rapid experimentation within large organizations. Pricing Tiers: Subscriptions are sold as blanket flat fees, starting at a minimum of $100,000 per year. Top-Tier Contracts: For larger deployments involving multiple teams and global data access, contracts scale up to $1.5 million per year. Usage Rights: Clients are granted unlimited usage during the contract period. This allows them to run thousands of simulations without worrying about incremental costs, encouraging a culture of "testing everything" before deployment. From a legal and privacy standpoint, the digital twins are built on a "fully consent-driven" framework. While the twins can be traced back to real human data for internal validation, the platform is designed to provide aggregated behavioral insights that protect the anonymity of the participants while maintaining the predictive power of their digital replicas. Rejecting the rise of Kalshi, Polymarket and 'prediction markets' The tech industry has recently seen a surge in valuations and interest in "prediction markets" like PolyMarket and Kalshi, which allow users to bet on the outcomes of various global events. However, the leadership at Brox maintains a distinct distance from these platforms, citing a "personal disdain" for betting markets from both a moral and intellectual perspective. Brocklebank argues that while betting markets can predict outcomes (e.g., who wins an election), they offer zero utility for business decision-makers because they fail to provide the "why". Knowing there is a 60% chance of a certain candidate winning does not help a company adjust its consumer strategy; knowing why a specific cohort of depositors is feeling anxious does. Investors including Scribble Ventures, Wonder Ventures, and Vela Partners have backed this "human-first" approach to AI, betting that the moat created by deep human data will prove more resilient than the commoditized models of synthetic data providers. As Brox prepares for launches in the Middle East and APAC, the company is moving toward its ultimate goal: simulating the entire world as a "parallel universe" for risk-free decision-making.

VentureBeat
~7 min readMay 6, 2026

Scaling AI into production is forcing a rethink of enterprise infrastructure

Presented by Nutanix Across industries, organizations are focused on how to move from AI pilots, proofs of concept, and cloud-based experimentation to deploying it at scale — across real workloads, for real users, in real business environments. VentureBeat spoke with Tarkan Maner, president and chief commercial officer at Nutanix, and Thomas Cornely, EVP of product management, about what that transition demands, and what it will take to get it right. “AI in general is shifting everything we do, not only in technology, but across all vertical industries, from regulated industries like banking, health care, government, education to non-regulated industries like manufacturing and retail,” Maner said. “As a complete platform company, we welcome this change. It’s creating more opportunities for us as a company to serve our customers in better ways as we move forward.” But there’s still a practical gap between experimentation and production, Cornely said. “It’s one thing to do an experiment, to do a prototype. It’s a different thing to take that prototype and deploy it for 10,000 employees,” he explained. “We went from people focusing on training models to chatbots to now doing agents, where the demand and pressures on AI infrastructure are growing exponentially.” Agentic AI introduces a new layer of enterprise complexity The rise of agentic AI is what makes this transition especially consequential. These systems introduce multi-step workflows across applications and data sources, along with a degree of autonomy that creates new operational demands. Enterprises now have to contend with multiple agents running simultaneously, unpredictable and real-time workloads, and the need to coordinate access to infrastructure across teams. “OpenClaw is making it very easy now for anybody to build agents and run with agents,” Cornely said. “You want those agents to be running on premises with your data. You need to have the right constructs around it to protect the enterprise from what an agent could do.” As these systems become more autonomous, the challenge extends beyond how they operate to how they interact with enterprise data, systems, and teams. AI is augmenting human work, not replacing it Agentic AI is fundamentally an amplifier of human capability rather than a substitute for it, Maner said. The goal for enterprises is not to eliminate human work but to find the right balance between human decision-making, AI-driven automation, and agent-based workflows. “We believe that there’s going to be love, peace, and harmony between AI, agentic tools, and robotics systems, and human capital,” Maner said. “That harmony can be optimized for better outcomes for businesses, enterprises, governments, and public sector organizations, if the right vendors provide the right tooling and the right services.” How enterprises are getting started with AI at scale In practice, the move from experimentation into real-world deployment is where the challenges become most visible. Despite the momentum, many are still working through how to scale AI beyond initial use cases. As they do, organizations quickly run into practical constraints. Many start in the cloud because of easy access to resources and services, but practical considerations like data, governance and control, and cost quickly come to the forefront. The cloud can be used to experiment, with the ultimate goal of bringing applications back on premises as they move toward production, using platforms that solve for security and cost. The use cases gaining the most traction include document search and knowledge retrieval, security and predictive threat detection, software development and coding workflows, and customer support and service operations. In the security realm, banking customers and others in Europe and the U.S. are deploying AI-driven tools including facial recognition and predictive threat detection. Meanwhile, there’s a growing focus on end-to-end, 360-degree customer engagement, from pre-sales through post-sales advocacy, in the customer support industry. Industry-specific AI transformation is already underway Across industries, the shift from experimentation to real deployment is already taking shape in distinct ways. In retail, AI is transforming store operations with cameras and robotics used for targeted in-aisle marketing at the moment of purchase decision, while cashier-less checkout is replacing traditional POS systems, and the human capital freed up is being redeployed to back-office and merchandising functions. In healthcare, Nutanix works with customers on applications spanning diagnosis, treatment, remote health, and hospital operations, with cloud partners including AWS and Azure. In manufacturing and logistics, the transformation is equally significant. The operational challenges of scaling enterprise AI As AI use cases scale, enterprises are running into a new class of operational challenges. Managing multiple AI workloads and agents, coordinating infrastructure access across teams, ensuring security and governance, and integrating AI systems with existing business processes are now top-of-mind concerns for IT and business leaders alike. The gap between AI developers pushing for speed and access, and infrastructure teams responsible for security, uptime, and governance, is one of the defining challenges of this moment. “Now I’m running agents, and they’re all going to fight to get access to resources to solve my problems,” Cornely said. “What you want now is infrastructure that allows you to set constraints, govern resources.” The AI factory: a shared platform for production AI These challenges are driving demand for what Maner and Cornely describe as the AI factory: a shared infrastructure environment that supports multiple users and workloads simultaneously, enabling both experimentation and production while balancing developer agility with enterprise governance. At GTC 2026, Nutanix announced the Nutanix Agentic AI Solution, a complete platform spanning core infrastructure, Kubernetes-based container services running on a topology-aware hypervisor, and advanced services for building and governing agents. “We’re launching a complete platform, from core infrastructure through PaaS and advanced PaaS services to the whole management framework for your AI factories,” Cornely said. “Really enabling self-service for the teams that will build these applications in the enterprise.” Hybrid environments are essential to enterprise AI strategy Operating this kind of environment requires flexibility across infrastructure. Hybrid infrastructure is not a compromise, but a requirement. Some workloads will always run in the public cloud, while others must remain on premises due to security requirements, regulatory compliance, data sovereignty, or competitive IP considerations. “Especially in the regulated industries, as sovereignty becomes a bigger issue, data gravity becomes a bigger issue, security, and also a lot of competitive differentiation in the industry, it’s going to depend on what the company wants for their own IP,” Maner said. This is the foundation of Nutanix’s platform position, he added. “We are the perfect harmony, bringing those applications, that data, and all the optimization for these use cases end to end, from on-prem to off-prem and in a hybrid mode,” he said. “Doing it not only in one cloud, but for multiple clouds.” That flexibility also extends to the broader ecosystem. Nutanix works across hyperscalers including AWS, Azure, and Google Cloud, as well as regional service providers and emerging neoclouds. Nutanix offers neoclouds a full software stack to run their own clouds and deliver advanced AI services, giving enterprise customers already running Nutanix a simple extension of compute, networking, and AI capabilities. Maner described the arrangement as a win for both sides. For enterprises, it means simplified access to hybrid AI services. For neoclouds, it means a proven platform to build on. It’s all automated and secure by default, Cornely added. “All of those governance problems that now come up with agentic AI are the same problems we’ve been solving for the last 16 years for every other application running in your cloud,” he said. From pilot to production: operationalizing AI across the enterprise Ultimately, the goal is not to run a successful AI pilot, but to operationalize AI across real-world use cases, manage infrastructure as a shared resource, support collaboration between infrastructure teams and AI developers, and scale from initial projects to enterprise-wide deployment. “There’s a massive gap right now between people building AI applications, those AI engineers, those agentic AI developers, and your classical infra teams,” Cornely said. “They need tooling to enable the infra teams, so they can support your AI engineers. That’s what we deliver with our agentic AI solution.” Sponsored articles are content produced by a company that is either paying for the post or has a business relationship with VentureBeat, and they’re always clearly marked. For more information, contact sales@venturebeat.com.

VentureBeat
~11 min readMay 5, 2026

Miami startup Subquadratic claims 1,000x AI efficiency gain with SubQ model; researchers demand independent proof.

A little-known Miami-based startup called Subquadratic emerged from stealth on Tuesday with a sweeping claim: that it has built the first large language model to fully escape the mathematical constraint that has defined — and limited — every major AI system since 2017. The company claims its first model, SubQ 1M-Preview, is the first LLM built on a fully subquadratic architecture — one where compute grows linearly with context length. If that claim holds, it would be a genuine inflection point in how AI systems scale. At 12 million tokens, the company says, its architecture reduces attention compute by almost 1,000 times compared to other frontier models — a figure that, if validated independently, would dwarf the efficiency gains of any existing approach. The company is also launching three products into private beta: an API exposing the full context window, a command-line coding agent called SubQ Code, and a search tool called SubQ Search. It has raised $29 million in seed funding from investors including Tinder co-founder Justin Mateen, former SoftBank Vision Fund partner Javier Villamizar, and early investors in Anthropic, OpenAI, Stripe, and Brex. The New Stack reported that the raise values the company at $500 million. The numbers Subquadratic is publishing are extraordinary. The reaction from the AI research community has been, to put it mildly, mixed — ranging from genuine curiosity to open accusations of vaporware. Understanding why requires understanding what the company claims to have solved, and why so many prior attempts to solve the same problem have fallen short. The quadratic scaling problem has shaped the economics of the entire AI industry Every transformer-based AI model — which includes virtually every frontier system from OpenAI, Anthropic, Google, and others — relies on an operation called "attention." Every token is compared against every other token, so as inputs grow, the number of interactions — and the compute required to process them — scales quadratically. In plain terms: double the input size, and the cost doesn't double. It quadruples. This relationship has shaped what gets built and what doesn't. The industry standard is 128,000 tokens for many AI models and up to 1 million tokens for frontier cloud models such as Claude Sonnet 4.7 and Gemini 3.1 Pro.  Even at those sizes, the cost of processing long inputs becomes punishing. The industry built an elaborate stack of workarounds to cope. RAG systems use a search engine to pull a small number of relevant results before sending them to the model, because sending the full corpus isn't feasible. Developers layer retrieval pipelines, chunking strategies, prompt engineering techniques, and multi-agent orchestration systems on top of models — all to route around the fundamental constraint that the model itself can't efficiently process everything at once. Subquadratic's argument is that these workarounds are expensive, brittle, and ultimately limiting. As CTO Alexander Whedon told SiliconANGLE in an interview, "I used to manually curate prompts and retrieval systems and evals and conditional logic to chain together the workflows. And I think that that is kind of a waste of human intelligence and also limiting to the product quality." Subquadratic's fix is deceptively simple: stop doing the math that doesn't matter The company's approach, called Subquadratic Sparse Attention or SSA, is built on a straightforward premise: most of the token-to-token comparisons in standard attention are wasted compute. Instead of comparing every token to every other token, SSA learns to identify which comparisons actually matter and computes attention only over those positions. Crucially, the selection is content-dependent — the model decides where to look based on meaning, not on fixed positional patterns. This allows it to retrieve specific information from arbitrary positions across a very long context without paying the quadratic tax. The practical payoff scales with context length — exactly the inverse of the problem it's trying to solve. According to the company's technical blog, SSA achieves a 7.2x prefill speedup over dense attention at 128,000 tokens, rising to 52.2x at 1 million tokens. As Whedon put it: "If you double the input size with quadratic scaling laws, you need four times the compute; with linear scaling laws, you need just twice." The company says it trained the model in three stages — pretraining, supervised fine-tuning, and a reinforcement learning stage specifically targeting long-context retrieval failures — teaching the model to aggressively use distant context rather than defaulting to nearby information, a subtle failure mode that quietly degrades performance in existing systems. Three benchmarks paint a strong picture, but what they leave out may matter more On the surface, SubQ's benchmark numbers are competitive with or superior to models built by organizations spending billions of dollars. On SWE-Bench Verified, it scored 81.8% compared to Opus 4.6's 80.8% and DeepSeek 4.0 Pro's 80.0%. On RULER at 128,000 tokens, a standard benchmark for reasoning over extended inputs, SubQ scored 95% — edging out Claude Opus 4.6 at 94.8%. On MRCR v2, a demanding test of multi-hop retrieval across long contexts, SubQ posted a third-party verified score of 65.9%, compared with Claude Opus 4.7 at 32.2%, GPT-5.5 at 74%, and Gemini 3.1 Pro at 26.3%. But several details warrant scrutiny. The benchmark selection is narrow — exactly three tests, all emphasizing long-context retrieval and coding, the precise tasks SubQ is designed for. Broader evaluations across general reasoning, math, multilingual performance, and safety have not been published. The company says a comprehensive model card is "coming soon." According to The New Stack, each benchmark model was run only once due to high inference cost, and the SWE-Bench margin is, as the company's own paper acknowledges, "harness as much as model." In benchmark methodology, single runs without confidence intervals leave room for variance. There is also a significant gap between SubQ's research results and its production model. On MRCR v2, the company reported a research score of 83 — but the third-party verified production model scored 65.9. That 17-point gap between the lab result and the shipping product is notable and largely unexplained. Subquadratic also told SiliconANGLE that on the RULER 128K benchmark, SubQ scored 95% accuracy at a cost of $8, compared with 94% accuracy and about $2,600 for Claude Opus — a remarkable cost claim. But the company has not publicly disclosed specific API pricing, making it impossible to independently verify the cost-per-task comparisons. The AI research community's verdict ranges from 'genuine breakthrough' to 'AI Theranos' Within hours of the announcement, the AI research community erupted into a debate that crystallized around a single question: Is this real? AI commentator Dan McAteer captured the binary mood in a widely shared post: "SubQ is either the biggest breakthrough since the Transformer... or it's AI Theranos." The comparison to the infamous blood-testing fraud company may be unfair, but it reflects the scale of the claims being made. Skeptics zeroed in on several pressure points. Prominent AI engineer Will Depue initially noted that SubQ is "almost surely a sparse attention finetune of Kimi or DeepSeek," referring to existing open-source models. Whedon confirmed this on X, writing that the company is "using weights from open-source models as a starting point, as a function of our funding and maturity as a company." Depue later escalated his criticism, writing that the company's O(n) scaling claims and the speedup numbers "don't seem to line up" and called the communication "either incredibly poorly communicated or just not real." Others raised structural questions. One developer noted that if SubQ truly reduces compute by 1,000x and costs less than 5% of Opus, the company should have no trouble serving it at scale — so why gate access through an early-access program? Developer Stepan Goncharov called the benchmarks "very interesting cherry-picked benchmarks," while another commenter described them as "suspiciously perfect." But not everyone was dismissive. AI researcher John Rysana pushed back on the Theranos framing, writing that the work is "just subquadratic attention done well which is very meaningful for long context workloads," and that "odds of it being BS are extremely low." Linus Ekenstam, a tech commentator, said he was "extremely intrigued to see the real-world implications" particularly for complex AI-powered software. Magic.dev made strikingly similar claims two years ago — and then went quiet Perhaps the most pointed critique of SubQ's launch comes not from its specific claims but from recent history. Magic.dev announced a 100-million-token context-window model in August 2024, with a claimed 1,000x efficiency advantage, and raised roughly $500 million on the strength of those claims. As of early 2026, there is no public evidence of LTM-2-mini being used outside Magic. The parallels are uncomfortable. Both companies claimed massive context windows. Both touted roughly 1,000x efficiency gains. Both targeted software engineering as their primary use case. And both launched with limited external access. The broader research landscape reinforces the caution. Kimi Linear, DeepSeek Sparse Attention, Mamba, and RWKV all promised subquadratic scaling, and all faced the same problem: architectures that achieve linear complexity in theory often underperform quadratic attention on downstream benchmarks at frontier scale, or they end up hybrid — mixing subquadratic layers with standard attention and losing the pure scaling benefits. A widely cited LessWrong analysis argued that these approaches "are all better thought of as 'incremental improvement number 93595 to the transformer architecture'" because practical implementations remain quadratic and "only improve attention by a constant factor." Subquadratic is directly aware of this history. Its own technical blog specifically addresses each prior approach — fixed-pattern sparse attention, state space models, hybrid architectures, and DeepSeek Sparse Attention — and argues that SSA avoids their tradeoffs. Whether it actually does remains an empirical question that only independent evaluation can settle. A five-time founder, a former Meta engineer, and $29 million to prove the doubters wrong The team behind the claims matters in evaluating them. CEO Justin Dangel is a five-time founder and CEO with a track record across health tech, insurancetech, and consumer goods, and his companies have scaled to hundreds of employees, attracted institutional backing, and reached liquidity. CTO Alexander Whedon previously worked as a software engineer at Meta and served as Head of Generative AI at TribeAI, where he led over 40 enterprise AI implementations. The team includes 11 PhD researchers with backgrounds from Meta, Google, Oxford, Cambridge, ByteDance, and Adobe. That is a credible collection of talent for an architecture-level research effort. But neither co-founder has published foundational AI research, and the company has not yet released a peer-reviewed paper. The technical report is listed as "coming soon." The funding profile is unusual for a company making frontier AI claims. Subquadratic raised $29 million at a reported $500 million valuation — a steep price for a seed-stage company with no publicly available model, no peer-reviewed research, and no disclosed revenue. The investor base, led by Tinder co-founder Mateen and former SoftBank partner Villamizar, skews toward consumer tech and growth investing rather than deep technical AI research. The company is not open-sourcing its weights but plans to offer training tools for enterprises to do their own post-training, and has set a 50-million-token context window target for Q4. The real test for SubQ isn't benchmarks — it's whether the math survives independent scrutiny Strip away the marketing language and the social media drama, and the underlying question Subquadratic is asking is genuinely important: Can AI systems break free of quadratic scaling without sacrificing the quality that makes them useful? The stakes are enormous. If attention can be made truly linear without degrading retrieval and reasoning, the economics of AI shift fundamentally. Enterprise applications that today require elaborate retrieval pipelines — processing entire codebases, contracts, regulatory filings, medical records — become single-pass operations. The billions of dollars currently spent on RAG infrastructure, context management, and agentic orchestration become partially redundant.  Whedon's willingness to engage publicly with technical criticism — posting a technical blog within hours of pushback — suggests a team that understands it needs to show its work, not just describe it. And to its credit, the company acknowledged openly that it builds on open-source foundations and that its model is smaller than those at the major labs. Every frontier model in 2026 advertises a context window of at least a million tokens, but almost none of them are actually great at making use of all that information. The gap between a nominal context window and a functional one — between what a model accepts and what it reliably reasons over — remains one of the most important unsolved problems in AI. Subquadratic says it has closed that gap. If independent evaluation confirms that claim, the implications would ripple far beyond a single startup's valuation. If it doesn't, the company joins a growing list of long-context promises that sounded revolutionary on launch day and unremarkable six months later. In computing, every fundamental constraint eventually falls. When it does, the breakthrough never comes from the direction the industry expected. The question hanging over Subquadratic is whether a team of 11 PhDs and a $29 million seed round actually found the answer that has eluded organizations spending thousands of times more — or whether they just found a better way to describe the problem.

VentureBeat
~7 min readMay 5, 2026

AI agents are missing all the discussions your team is having. SageOX has an answer: agentic context infrastructure

As AI model providers increasingly move downstream, launching products and agents for specific enterprise applications and sectors like finance, one big question still remains: how will said AI agents be equipped with the proper context surrounding a task — who assigned it, which other stakeholders are involved, what data or discussions have taken place about it and how it should be done? This practice of "context engineering" remains one of the great unsolved problems of the AI era. But SageOx, a Seattle-based startup founded by the veterans who built the original AWS EC2 and EBS infrastructure, believes it has the answer: a new systems layer it calls "agentic context infrastructure." Using a combination of small hardware recording devices and the existing applications enterprises already rely on — Slack, email, documents, files — and applying new, open-source frameworks and instructions atop it all, SageOX has developed a system by which enterprises can keep agents as "in-the-loop" and updated on the enterprise's tasks as their human employees are, and prevent them from "drifting" off their assigned tasks and the firm's larger goals. “We are capturing all of this context where it happens," said Ajit Banerjee, founder and CEO of SageOX and a former Hugging Face, Meta, Amazon and Apple engineer said in a recent video call interview with VentureBeat. "Product development is a team sport, and the context doesn’t just come from people typing on a keyboard. It happens in conversations.” By capturing the "why" behind the "what"—the intent that lives in Slack threads, whiteboarding sessions, and water-cooler conversations—SageOx aims to provide a "hivemind" that ensures agents don't drift and humans stay in flow. "The way people have to work is not old-school coordination, where I write down an issue and then it goes through a sequence. It has to be almost like playing jazz," Banerjee added. Today, the company emerged from stealth to announce its $15 million seed round led by Canaan and participation from A.Capital, Pioneer Square Labs, and Founders’ Co-op. The architecture of team memory Today’s AI agents operate in isolated sessions, lacking a shared memory of prior decisions or architectural intent. Every task effectively starts from scratch, forcing developers to manually recap context—a process that undermines the very speed agents are meant to provide. SageOx addresses this through a multi-surface product suite designed to capture context wherever it naturally occurs. At the center of this ecosystem is the Ox Dot. A customized hardware device designed for the shared office, the Dot captures meetings, standups, and design reviews with a single touch. Its most distinctive feature is "Auto Rewind"—a fail-safe for the spontaneous brilliance of a team. If a breakthrough happens during an unrecorded conversation, Auto Rewind allows the team to "go back" and capture the discussion after the fact. This audio is transcribed, speaker-identified, and distilled into team memory, where it becomes accessible to both humans and agents. For the developer, the open-source, MIT-licensed Ox CLI provides the bridge. Commands like ox agent prime allow coding assistants—including Claude Code and Codex—to consult the team's shared history before writing code. This ensures that if a team decided in a meeting to use a specific authentication pattern, the agent knows it without being explicitly told in a prompt. As Dr. Rupak Majumdar, Scientific Director, Max Planck Institute for Software Systems, noted after seeing the team’s development speed, they are effectively "treating code like assembler." Agentic engineering: moving Beyond "clean" code The shift to an agent-first workflow has forced the SageOx team to reconsider nearly every principle of modern software management. SageOX CTO Ryan Snodgrass, formerly of Amazon, notes in a blog post transcript that traditional branch management and "clean" commit histories are often "bad for the agents." In the old world, humans preferred large PRs that were easy to read during a single code review. In the agentic era, 10,000-line PRs spread across the codebase make it impossible for an agent to reason about intent. Instead, SageOx advocates for smaller, high-volume, and highly focused commits. This "agent-readable" history allows the machine to look back and understand exactly why a specific change was made. The team is even re-evaluating repo structures; while they currently utilize a monorepo for their 750,000 lines of code, they are exploring a future where agents manage a constellation of micro-repos, as agents can "get lost" when a codebase grows too large for their context window. This philosophy of "speed-over-stasis" allowed the team to build their own firmware for the Ox Dot in less than two weeks, despite having no recent hardware experience. By feeding technical PDFs and documentation into AI models, they bypassed months of traditional research. CEO Ajit Banerjee calls this the "unlearning" of old habits—realizing that the "undifferentiated heavy lifting" of knowledge work can now be offloaded to a system that remembers everything the team knows. Radical transparency: beyond open source to an "open work" model Perhaps as significant as the technology is SageOx’s commitment to "Open Work." Moving beyond traditional open-source software, the company is practicing a form of radical transparency in an effort to foster the acceleration of development across the entire open source community and any enterprises who wish to learn from the way they work. SageOx's team openly shares their internal prompts, their planning sessions, and even their unfiltered internal debates with the public. Users can sign in to the SageOx console and watch the team build SageOx in real-time. This "open kimono" approach was an intentional decision to lead by example. Banerjee argues that since they are asking teams to change how they work, they must be willing to show the "WTF" moments and the course corrections as they happen. "The revolution is not going to be televised," Banerjee says. "It's going to be SageOxed." This transparency is intended to prove that a small, lean team—"yoking up lean"—can outpace massive organizations by leveraging a shared context layer. As for how SageOx plans to monetize and become profitable, Banerjee said the revenue path is modeled on the AWS EC2 playbook: start with early adopters, especially small AI-native startups, then expand toward enterprises as the need becomes obvious. The pedigree of infrastructure The technical foundation of SageOx is rooted in the early days of cloud infrastructure. Banerjee was an original member of the AWS EC2 team, and Snodgrass was one of Amazon's first engineers, leading the transition from monolithic architectures to microservices. This background is reflected in the company’s name: the "Ox" represents the "Yeoman work" they aim to do—a dependable animal that handles the heavy lifting of data and context so the team can move forward. The SageOx vision is one where humans are no longer the manual assemblers of context. Instead, they act as the directors of a "parallel processing" engine. In a recent demonstration, a feature request moved from a verbal discussion to a completed implementation in under seven minutes. By priming coding agents with the recorded context of the original discussion, the team bypassed the need for formal specs or Jira tickets. The new way of work SageOx is currently focusing its efforts on "AI-native" startups—teams that operate primarily through prompts and rely heavily on agentic coworkers. Their suite of tools, from the open-source Ox CLI to the hardware-enabled Ox Dot, is designed to solve the immediate problem of alignment drift. As AI moves from being a tool to a teammate, the most valuable asset a company possesses is no longer its raw source code, but its shared context. SageOx suggests that the way forward is not to hoard information behind "private fences," but to create a communal ground where intent is visible to every teammate—human or machine. In this new epoch, the teams that win will be the ones that can remember as fast as they can execute.