New articles on Quantitative Finance


[1] 2508.02684

Strategic competition in informal risk sharing mechanism versus collective index insurance

The frequent occurrence of natural disasters has posed significant challenges to society, necessitating the urgent development of effective risk management strategies. From the early informal community-based risk sharing mechanisms to modern formal index insurance products, risk management tools have continuously evolved. Although index insurance provides an effective risk transfer mechanism in theory, it still faces the problems of basis risk and pricing in practice. At the same time, in the presence of informal community risk sharing mechanisms, the competitiveness of index insurance deserves further investigation. Here we propose a three-strategy evolutionary game model, which simultaneously examines the competitive relationship between formal index insurance purchasing (I), informal risk sharing strategies (S), and complete non-insurance (A). Furthermore, we introduce a method for calculating insurance company profits to aid in the optimal pricing of index insurance products. We find that basis risk and risk loss ratio have significant impacts on insurance adoption rate. Under scenarios with low basis risk and high loss ratios, index insurance is more popular; meanwhile, when the loss ratio is moderate, an informal risk sharing strategy is the preferred option. Conversely, when the loss ratio is low, individuals tend to forego any insurance. Furthermore, accurately assessing the degree of risk aversion and determining the appropriate ratio of risk sharing are crucial for predicting the future market sales of index insurance.


[2] 2508.02685

Benchmarking Classical and Quantum Models for DeFi Yield Prediction on Curve Finance

The rise of decentralized finance (DeFi) has created a growing demand for accurate yield and performance forecasting to guide liquidity allocation strategies. In this study, we benchmark six models, XGBoost, Random Forest, LSTM, Transformer, quantum neural networks (QNN), and quantum support vector machines with quantum feature maps (QSVM-QNN), on one year of historical data from 28 Curve Finance pools. We evaluate model performance on test MAE, RMSE, and directional accuracy. Our results show that classical ensemble models, particularly XGBoost and Random Forest, consistently outperform both deep learning and quantum models. XGBoost achieves the highest directional accuracy (71.57%) with a test MAE of 1.80, while Random Forest attains the lowest test MAE of 1.77 and 71.36% accuracy. In contrast, quantum models underperform with directional accuracy below 50% and higher errors, highlighting current limitations in applying quantum machine learning to real-world DeFi time series data. This work offers a reproducible benchmark and practical insights into model suitability for DeFi applications, emphasizing the robustness of classical methods over emerging quantum approaches in this domain.


[3] 2508.02686

Adaptive Market Intelligence: A Mixture of Experts Framework for Volatility-Sensitive Stock Forecasting

This study develops and empirically validates a Mixture of Experts (MoE) framework for stock price prediction across heterogeneous volatility regimes using real market data. The proposed model combines a Recurrent Neural Network (RNN) optimized for high-volatility stocks with a linear regression model tailored to stable equities. A volatility-aware gating mechanism dynamically weights the contributions of each expert based on asset classification. Using a dataset of 30 publicly traded U.S. stocks spanning diverse sectors, the MoE approach consistently outperforms both standalone models. Specifically, it achieves up to 33% improvement in MSE for volatile assets and 28% for stable assets relative to their respective baselines. Stratified evaluation across volatility classes demonstrates the model's ability to adapt complexity to underlying market dynamics. These results confirm that no single model suffices across market regimes and highlight the advantage of adaptive architectures in financial prediction. Future work should explore real-time gate learning, dynamic volatility segmentation, and applications to portfolio optimization.


[4] 2508.02691

Statistical modeling of SOFR term structure

SOFR derivatives market remains illiquid and incomplete so it is not amenable to classical risk-neutral term structure models which are based on the assumption of perfect liquidity and completeness. This paper develops a statistical SOFR term structure model that is well-suited for risk management and derivatives pricing within the incomplete markets paradigm. The model incorporates relevant macroeconomic factors that drive central bank policy rates which, in turn, cause jumps often observed in the SOFR rates. The model is easy to calibrate to historical data, current market quotes, and the user's views concerning the future development of the relevant macroeconomic factors. The model is well suited for large-scale simulations often required in risk management, portfolio optimization and indifference pricing of interest rate derivatives.


[5] 2508.02702

Evaluating Transfer Learning Methods on Real-World Data Streams: A Case Study in Financial Fraud Detection

When the available data for a target domain is limited, transfer learning (TL) methods can be used to develop models on related data-rich domains, before deploying them on the target domain. However, these TL methods are typically designed with specific, static assumptions on the amount of available labeled and unlabeled target data. This is in contrast with many real world applications, where the availability of data and corresponding labels varies over time. Since the evaluation of the TL methods is typically also performed under the same static data availability assumptions, this would lead to unrealistic expectations concerning their performance in real world settings. To support a more realistic evaluation and comparison of TL algorithms and models, we propose a data manipulation framework that (1) simulates varying data availability scenarios over time, (2) creates multiple domains through resampling of a given dataset and (3) introduces inter-domain variability by applying realistic domain transformations, e.g., creating a variety of potentially time-dependent covariate and concept shifts. These capabilities enable simulation of a large number of realistic variants of the experiments, in turn providing more information about the potential behavior of algorithms when deployed in dynamic settings. We demonstrate the usefulness of the proposed framework by performing a case study on a proprietary real-world suite of card payment datasets. Given the confidential nature of the case study, we also illustrate the use of the framework on the publicly available Bank Account Fraud (BAF) dataset. By providing a methodology for evaluating TL methods over time and in realistic data availability scenarios, our framework facilitates understanding of the behavior of models and algorithms. This leads to better decision making when deploying models for new domains in real-world environments.


[6] 2508.02738

CreditARF: A Framework for Corporate Credit Rating with Annual Report and Financial Feature Integration

Corporate credit rating serves as a crucial intermediary service in the market economy, playing a key role in maintaining economic order. Existing credit rating models rely on financial metrics and deep learning. However, they often overlook insights from non-financial data, such as corporate annual reports. To address this, this paper introduces a corporate credit rating framework that integrates financial data with features extracted from annual reports using FinBERT, aiming to fully leverage the potential value of unstructured text data. In addition, we have developed a large-scale dataset, the Comprehensive Corporate Rating Dataset (CCRD), which combines both traditional financial data and textual data from annual reports. The experimental results show that the proposed method improves the accuracy of the rating predictions by 8-12%, significantly improving the effectiveness and reliability of corporate credit ratings.


[7] 2508.02739

Kronos: A Foundation Model for the Language of Financial Markets

The success of large-scale pre-training paradigm, exemplified by Large Language Models (LLMs), has inspired the development of Time Series Foundation Models (TSFMs). However, their application to financial candlestick (K-line) data remains limited, often underperforming non-pre-trained architectures. Moreover, existing TSFMs often overlook crucial downstream tasks such as volatility prediction and synthetic data generation. To address these limitations, we propose Kronos, a unified, scalable pre-training framework tailored to financial K-line modeling. Kronos introduces a specialized tokenizer that discretizes continuous market information into token sequences, preserving both price dynamics and trade activity patterns. We pre-train Kronos using an autoregressive objective on a massive, multi-market corpus of over 12 billion K-line records from 45 global exchanges, enabling it to learn nuanced temporal and cross-asset representations. Kronos excels in a zero-shot setting across a diverse set of financial tasks. On benchmark datasets, Kronos boosts price series forecasting RankIC by 93% over the leading TSFM and 87% over the best non-pre-trained baseline. It also achieves a 9% lower MAE in volatility forecasting and a 22% improvement in generative fidelity for synthetic K-line sequences. These results establish Kronos as a robust, versatile foundation model for end-to-end financial time series analysis. Our pre-trained model is publicly available at this https URL.


[8] 2508.02758

CTBench: Cryptocurrency Time Series Generation Benchmark

Synthetic time series are essential tools for data augmentation, stress testing, and algorithmic prototyping in quantitative finance. However, in cryptocurrency markets, characterized by 24/7 trading, extreme volatility, and rapid regime shifts, existing Time Series Generation (TSG) methods and benchmarks often fall short, jeopardizing practical utility. Most prior work (1) targets non-financial or traditional financial domains, (2) focuses narrowly on classification and forecasting while neglecting crypto-specific complexities, and (3) lacks critical financial evaluations, particularly for trading applications. To address these gaps, we introduce \textsf{CTBench}, the first comprehensive TSG benchmark tailored for the cryptocurrency domain. \textsf{CTBench} curates an open-source dataset from 452 tokens and evaluates TSG models across 13 metrics spanning 5 key dimensions: forecasting accuracy, rank fidelity, trading performance, risk assessment, and computational efficiency. A key innovation is a dual-task evaluation framework: (1) the \emph{Predictive Utility} task measures how well synthetic data preserves temporal and cross-sectional patterns for forecasting, while (2) the \emph{Statistical Arbitrage} task assesses whether reconstructed series support mean-reverting signals for trading. We benchmark eight representative models from five methodological families over four distinct market regimes, uncovering trade-offs between statistical fidelity and real-world profitability. Notably, \textsf{CTBench} offers model ranking analysis and actionable guidance for selecting and deploying TSG models in crypto analytics and strategy development.


[9] 2508.02949

Ukrainian-style oligarchic economies: how concentrated power undermines value added in production chains

Oligarchic control exerts significant distortions on economic efficiency. Ukraine exemplifies this phenomenon, where oligarchs dominate key sectors and achieve economies of scale through vertical integration of coal mines, steel mills, and power plants while controlling critical infrastructure (e.g. access to transportation networks) to stifle competition. Their Soviet-era production chain monopolization strategies, coupled with political patronage networks (including both local and national governments), reinforce systemic inefficiencies and barriers to market entry. Although existing studies highlight the developmental benefits of de-oligarchization, this work advances the literature through computational modeling. We develop an agent-based model of a partially oligarch-controlled economy, where firms with heterogeneous production functions interact within a value-added network. Through numerical simulations, we quantify how different de-oligarchization policies affect aggregate GDP growth. The results indicate that the optimal de-oligarchization strategies are determined by the position of the oligarch in the production chain. Depending on the oligarch's position, dismantling oligarchic structures should either focus on removing oligarchs' access to raw materials or on breaking oligarchs' influence on other transactions in the production chain.


[10] 2508.02966

Measuring Human Leadership Skills with Artificially Intelligent Agents

We show that the ability to lead groups of humans is predicted by leadership skill with Artificially Intelligent agents. In a large pre-registered lab experiment, human leaders worked with AI agents to solve problems. Their performance on this 'AI leadership test' was strongly correlated with their causal impact on human teams, which we estimate by repeatedly randomly assigning leaders to groups of human followers and measuring team performance. Successful leaders of both humans and AI agents ask more questions and engage in more conversational turn-taking; they score higher on measures of social intelligence, fluid intelligence, and decision-making skill, but do not differ in gender, age, ethnicity or education. Our findings indicate that AI agents can be effective proxies for human participants in social experiments, which greatly simplifies the measurement of leadership and teamwork skills.


[11] 2508.02971

Modeling Loss-Versus-Rebalancing in Automated Market Makers via Continuous-Installment Options

This paper mathematically models a constant-function automated market maker (CFAMM) position as a portfolio of exotic options, known as perpetual American continuous-installment (CI) options. This model replicates an AMM position's delta at each point in time over an infinite time horizon, thus taking into account the perpetual nature and optionality to withdraw of liquidity provision. This framework yields two key theoretical results: (a) It proves that the AMM's adverse-selection cost, loss-versus-rebalancing (LVR), is analytically identical to the continuous funding fees (the time value decay or theta) earned by the at-the-money CI option embedded in the replicating portfolio. (b) A special case of this model derives an AMM liquidity position's delta profile and boundaries that suffer approximately constant LVR, up to a bounded residual error, over an arbitrarily long forward window. Finally, the paper describes how the constant volatility parameter required by the perpetual option can be calibrated from the term structure of implied volatilities and estimates the errors for both implied volatility calibration and LVR residual error. Thus, this work provides a practical framework enabling liquidity providers to choose an AMM liquidity profile and price boundaries for an arbitrarily long, forward-looking time window where they can expect an approximately constant, price-independent LVR. The results establish a rigorous option-theoretic interpretation of AMMs and their LVR, and provide actionable guidance for liquidity providers in estimating future adverse-selection costs and optimizing position parameters.


[12] 2508.03217

Measuring DEX Efficiency and The Effect of an Enhanced Routing Method on Both DEX Efficiency and Stakeholders' Benefits

The efficiency of decentralized exchanges (DEXs) and the influence of token routing algorithms on market performance and stakeholder outcomes remain underexplored. This paper introduces the concept of Standardized Total Arbitrage Profit (STAP), computed via convex optimization, as a systematic measure of DEX efficiency. We prove that executing the trade order maximizing STAP and reintegrating the resulting transaction fees eliminates all arbitrage opportunities-both cyclic arbitrage within DEXs and between DEXs and centralized exchanges (CEXs). In a fully efficient DEX (i.e., STAP = 0), the monetary value of target tokens received must not exceed that of the source tokens, regardless of the routing algorithm. Any violation indicates arbitrage potential, making STAP a reliable metric for arbitrage detection. Using a token graph comprising 11 tokens and 18 liquidity pools based on Uniswap V2 data, we observe a decline in DEX efficiency between June 21 and November 8, 2024. Simulations comparing two routing algorithms-Yu Zhang et al.'s line-graph-based method and the depth-first search (DFS) algorithm-show that employing more profitable routing improves DEX efficiency and trader returns over time. Moreover, while total value locked (TVL) remains stable with the line-graph method, it increases under the DFS algorithm, indicating greater aggregate benefits for liquidity providers.


[13] 2508.03230

To Bubble or Not to Bubble: Asset Price Dynamics and Optimality in OLG Economies

We study an overlapping generations (OLG) exchange economy with an asset that yields dividends. First, we derive general conditions, based on exogenous parameters, that give rise to three distinct scenarios: (1) only bubbleless equilibria exist, (2) a bubbleless equilibrium coexists with a continuum of bubbly equilibria, and (3) all equilibria are bubbly. Under stationary endowments and standard assumptions, we provide a complete characterization of the equilibrium set and the associated asset price dynamics. In this setting, a bubbly equilibrium exists if and only if the interest rate in the economy without the asset is strictly lower than the population growth rate and the sum of per capita dividends is finite. Second, we establish necessary and sufficient conditions for Pareto optimality. Finally, we investigate the relationship between asset price behaviors and the optimality of equilibria.


[14] 2508.02773

Web3 x AI Agents: Landscape, Integrations, and Foundational Challenges

The convergence of Web3 technologies and AI agents represents a rapidly evolving frontier poised to reshape decentralized ecosystems. This paper presents the first and most comprehensive analysis of the intersection between Web3 and AI agents, examining five critical dimensions: landscape, economics, governance, security, and trust mechanisms. Through an analysis of 133 existing projects, we first develop a taxonomy and systematically map the current market landscape (RQ1), identifying distinct patterns in project distribution and capitalization. Building upon these findings, we further investigate four key integrations: (1) the role of AI agents in participating in and optimizing decentralized finance (RQ2); (2) their contribution to enhancing Web3 governance mechanisms (RQ3); (3) their capacity to strengthen Web3 security via intelligent vulnerability detection and automated smart contract auditing (RQ4); and (4) the establishment of robust reliability frameworks for AI agent operations leveraging Web3's inherent trust infrastructure (RQ5). By synthesizing these dimensions, we identify key integration patterns, highlight foundational challenges related to scalability, security, and ethics, and outline critical considerations for future research toward building robust, intelligent, and trustworthy decentralized systems with effective AI agent interactions.


[15] 2508.03474

Unravelling the Probabilistic Forest: Arbitrage in Prediction Markets

Polymarket is a prediction market platform where users can speculate on future events by trading shares tied to specific outcomes, known as conditions. Each market is associated with a set of one or more such conditions. To ensure proper market resolution, the condition set must be exhaustive -- collectively accounting for all possible outcomes -- and mutually exclusive -- only one condition may resolve as true. Thus, the collective prices of all related outcomes should be \$1, representing a combined probability of 1 of any outcome. Despite this design, Polymarket exhibits cases where dependent assets are mispriced, allowing for purchasing (or selling) a certain outcome for less than (or more than) \$1, guaranteeing profit. This phenomenon, known as arbitrage, could enable sophisticated participants to exploit such inconsistencies. In this paper, we conduct an empirical arbitrage analysis on Polymarket data to answer three key questions: (Q1) What conditions give rise to arbitrage (Q2) Does arbitrage actually occur on Polymarket and (Q3) Has anyone exploited these opportunities. A major challenge in analyzing arbitrage between related markets lies in the scalability of comparisons across a large number of markets and conditions, with a naive analysis requiring $O(2^{n+m})$ comparisons. To overcome this, we employ a heuristic-driven reduction strategy based on timeliness, topical similarity, and combinatorial relationships, further validated by expert input. Our study reveals two distinct forms of arbitrage on Polymarket: Market Rebalancing Arbitrage, which occurs within a single market or condition, and Combinatorial Arbitrage, which spans across multiple markets. We use on-chain historical order book data to analyze when these types of arbitrage opportunities have existed, and when they have been executed by users. We find a realized estimate of 40 million USD of profit extracted.


[16] 2506.18829

The Theory of Economic Complexity

Economic complexity methods aim to estimate the combined presence of economic factors without having to explicitly define them. A key method in this literature is the Economic Complexity Index or ECI, an eigenvector derived from specialization matrices that explains variation in economic growth, inequality, and sustainability. Yet, despite the widespread use of ECI in economic development, economic geography, and innovation studies, we still lack a principled theory that can deduce it from a mechanistic model. Here, we calculate $ECI$ analytically for a model where the output of an economy in an activity increases if the economy is more likely to be endowed with the factors required by the activity. We derive ECI analytically and numerically and show that it is a monotonic function of the probability that an economy is endowed with many factors, validating the idea that ECI is an agnostic estimate of the presence of multiple factors in an economy. We then generalize this result to other production functions and to a short-run equilibrium framework with prices, wages, and consumption, finding that the derived wage function is consistent with economies converging to an income that is compatible with their complexity. Finally, we show this model explains differences in the shapes of networks of related activities, such as the product space and the research space. These findings solve long standing puzzles in the literature and validate metrics of economic complexity as estimates of the combined presence of multiple factors.


[17] 2507.11480

Pricing energy spread options with variance gamma-driven Ornstein-Uhlenbeck dynamics

We consider the pricing of energy spread options for spot prices following an exponential Ornstein-Uhlenbeck process driven by a sum of independent multivariate variance gamma processes, which gives rise to mean-reverting, infinite activity price dynamics. Within this class of driving processes, the Esscher transform is used to obtain an equivalent martingale measure with a focus on the weak variance alpha-gamma process. By deriving an analytic expression for the cumulant generating function of the innovation term, we obtain a pricing formula for forwards and apply the FFT method of Hurd and Zhou to price spread options. Lastly, we demonstrate how the model should be both estimated on energy prices under the real world measure and calibrated on forward or call prices, and provide numerical results for the pricing of spread options.