About Olive Legal
Historically, quantitative models are domain specific. Brilliant people spend their best years testing features, tuning hyperparameters, and iterating architectures within a narrow domain. But scale is the panacea: large models will find patterns people, and specialized models, could not.
Forecasting generalizes. Zoa trains cross-domain event forecasting engines.
*Automating Iteration*
LLMs—embedded in multi-agent optimization loops and evaluated against fixed policies—can automate the build-test-improve modeling cycle. Think AlphaEvolve for forecasting problems.
*Sample-Efficient General Models*
Today’s forecasting models are narrowly crafted with deep human priors. But larger models will outperform state-of-the-art specialized models.
Unlike existing event models, our models leverage data from across contexts and rely less on human intuition. And compared to LLMs, our models are built with more inductive priors and rely more heavily on inference-time compute—improving sample efficiency.
*Why It Matters*
In the real economy, our models could be useful for forecasting supply chain volatility, energy supply and demand, even earthquake risk.
Science is, Ian Hacking writes, the taming of chance. It is the process of iteratively updating priors (something like: identify uncertainty, conceive experiment to reduce uncertainty, execute, update). If science is uncertainty-reduction, forecasting is a critical measure of progress.
Better forecasting improves our ability to select interesting experiments (roughly those with greatest expected uncertainty reduction) and update priors. Our models will be used by labs and academics in data-heavy domains.
Sam's ex-girlfriend introduced him to Greg back at Carnegie Mellon in 2017, and while that relationship didn't last, their friendship has. After college, Greg went to Harvard Law School, while Sam worked for three years at Jane Street on their Options desk, building & leading a satellite dev team.
Forecasting generalizes. Zoa trains cross-domain event forecasting engines.
*Automating Iteration*
LLMs—embedded in multi-agent optimization loops and evaluated against fixed policies—can automate the build-test-improve modeling cycle. Think AlphaEvolve for forecasting problems.
*Sample-Efficient General Models*
Today’s forecasting models are narrowly crafted with deep human priors. But larger models will outperform state-of-the-art specialized models.
Unlike existing event models, our models leverage data from across contexts and rely less on human intuition. And compared to LLMs, our models are built with more inductive priors and rely more heavily on inference-time compute—improving sample efficiency.
*Why It Matters*
In the real economy, our models could be useful for forecasting supply chain volatility, energy supply and demand, even earthquake risk.
Science is, Ian Hacking writes, the taming of chance. It is the process of iteratively updating priors (something like: identify uncertainty, conceive experiment to reduce uncertainty, execute, update). If science is uncertainty-reduction, forecasting is a critical measure of progress.
Better forecasting improves our ability to select interesting experiments (roughly those with greatest expected uncertainty reduction) and update priors. Our models will be used by labs and academics in data-heavy domains.
Sam's ex-girlfriend introduced him to Greg back at Carnegie Mellon in 2017, and while that relationship didn't last, their friendship has. After college, Greg went to Harvard Law School, while Sam worked for three years at Jane Street on their Options desk, building & leading a satellite dev team.
Latest Jobs
Founders & Key People
Financials Beta
Business Model: Not Specified
Revenues: Not Specified
Expenses: Not Specified
Debt: Request
Operating Status: Active
Funding Raised: $0
Investment Rounds: 0 Rounds
Funding Stage: Not Specified
Last Funding Date: Not Specified