
Enterprise Software for Data Analytics
Historically, one of the key ways of creating leverage in an organization was through enterprise software - transforming manual, error-prone processes into more systematic, robust processes. At scale, process improvements in an organization provide immense business benefit in the form of increasing revenue and decreasing cost. In the past decade, data analytics infrastructure and applications have transformed how organizations operate. Cheap, flexible cloud-based data warehouses enabled businesses of all sizes to analyze all of their business and product telemetry. Best practices around cloud-native ETL (rebranded as “ELT”) developed, allowing data ingestion and transformation pipelines to be more robust, scalable, and composable. And finally, easy-to-use business intelligence tools allowed data analysts of all skillsets to create data dashboards, primarily designed to summarize high-dimensional information in a compact, intuitive way. As of today, the data dashboard remains king as the primary interface of choice for last-mile analytics delivery, the interface for data insights.
LLMs disrupt data interfaces
In the past five years, the data analytics industry has seen innovation in the interface layer for last-mile analytics delivery. New form factors for data insights were developed and marketed at scale including notebooks, data apps, spreadsheets, and canvases. More than five years into this experiment, it has become more clear that none of these new form factors were fundamentally different enough and better enough to take hold as the new primary interface for data insights. Enter Large Language Models. The introduction of ChatGPT to the world at the end of 2022 caused a (perhaps unexpected, even to its creators) sensation, leading to mass adoption and change of user behavior that we have not seen since the mobile and cloud waves of the late 2000s and early 2010s. Unlike the false chatbot wave of the mid 2010s, this time, the technology works at scale, and users are sufficiently primed to take advantage of this new technology.

Introducing Oxy, an open-source framework for agentic analytics
To enable a future where data agents become the primary interface for data insights, we built Oxy, an open-source framework for agentic analytics. The core of Oxy is an open-source, declarative framework built in Rust that is purpose-built for agentic data analytics. Oxy is built with the following product principles: open-source, performant, code-native, declarative, composable, and secure. We were founded by second-time venture-backed (Khosla) entrepreneurs with research backgrounds from Harvard and MIT who led data and ML initiatives at companies like Airbnb, Wayfair, and QuantCo.



The Human-to-AI interface
Philosophically, the reason why a framework like Oxy is so powerful is that the AI interface it creates closely mimics existing human behavior, the human-to-human interface. As an analogy, seasoned human data analysts are valuable in an organization because of the degrees of freedom a human-to-human interface provides, as opposed to working with a pure software form factor. The human-to-human interface is superior at being robust against slightly erroneous input and output information. The human-to-human interface is superior at covering a very wide range of question-workflow pairs, starting with the most simple question about a particular metric in a point of time but also expanding into a very complex workflow that entails the synthesis of thousands of findings from hundreds of metrics and their correlations that result in a major business decision. This interface gets to the point directly, rather than using a rudimentary artifact like the dashboard as a repeatable yet poorly-powered intermediary. It is not difficult to envision that a human-to-AI interface that functions accurately and quickly that is engineered to recapture this natural paradigm will swiftly rise in primacy.
