We propose a highly efficient and accurate methodology for generating synthetic financial market data using a diffusion model approach. The synthetic data produced by our methodology align closely with observed market data in several key aspects: (i) they pass the two-sample Cramer - von Mises test for portfolios of assets, and (ii) Q - Q plots demonstrate consistency across quantiles, including in the tails, between observed and generated market data. Moreover, the covariance matrices derived from a large set of synthetic market data exhibit significantly lower condition numbers compared to the estimated covariance matrices of the observed data. This property makes them suitable for use as regularized versions of the latter. For model training, we develop an efficient and fast algorithm based on numerical integration rather than Monte Carlo simulations. The methodology is tested on a large set of equity data.
We examine the empirical performance of some parametric and nonparametric estimators of prices of options with a fixed time to maturity, focusing on variance-gamma and Heston models on one side, and on expansions in Hermite functions on the other side. The latter class of estimators can be seen as perturbations of the classical Black-Scholes model. The comparison between parametric and Hermite-based models having the same "degrees of freedom" is emphasized. The main criterion is the out-of-sample relative pricing error on a dataset of historical option prices on the S&P500 index. Prior to the main empirical study, the approximation of variance-gamma and Heston densities by series of Hermite functions is studied, providing explicit expressions for the coefficients of the expansion in the former case, and integral expressions involving the explicit characteristic function in the latter case. Moreover, these approximations are investigated numerically on a few test cases, indicating that expansions in Hermite functions with few terms achieve competitive accuracy in the estimation of Heston densities and the pricing of (European) options, but they perform less effectively with variance-gamma densities. On the other hand, the main large-scale empirical study show that parsimonious Hermite estimators can even outperform the Heston model in terms of pricing errors. These results underscore the trade-offs inherent in model selection and calibration, and their empirical fit in practical applications.
This paper studies the time-varying structure of the equity market with respect to market capitalization. First, we analyze the distribution of the 100 largest companies' market capitalizations over time, in terms of inequality, concentration at the top, and overall discrepancies in the distribution between different times. In the next section, we introduce a mathematical framework of linear and nonlinear functionals of time-varying portfolios. We apply this to study the market capitalization exposure and spread of optimal portfolios chosen by a Sharpe optimization procedure. These methods could be more widely used to study various measures of optimal portfolios and measure different aspects of market exposure while holding portfolios selected by an optimization routine that changes over time.
This paper examines the impact of racial discrimination in hiring on employment, wage, and wealth disparities between black and white workers. Using a labor search-and-matching model with racially prejudiced and non-prejudiced firms, we show that labor market frictions sustain discriminatory practices as an equilibrium outcome. These practices account for 44% to 52% of the average wage gap and 16% of the median wealth gap. Discriminatory hiring also amplifies unemployment and wage volatility for black workers, increasing their labor market risks over business cycles. Eliminating prejudiced firms reduces these disparities and improves black workers' welfare, though it slightly decreases overall economic welfare.
We study risk sharing among agents with preferences modeled by heterogeneous distortion risk measures, who are not necessarily risk averse. Pareto optimality for agents using risk measures is often studied through the lens of inf-convolutions, because allocations that attain the inf-convolution are Pareto optimal, and the converse holds true under translation invariance. Our main focus is on groups of agents who exhibit varying levels of risk seeking. Under mild assumptions, we derive explicit solutions for the unconstrained inf-convolution and the counter-monotonic inf-convolution, which can be represented by a generalization of distortion risk measures. Furthermore, for a group of agents with different levels of risk aversion or risk seeking, we consider a portfolio manager's problem and explicitly determine the optimal investment strategies. Interestingly, we observe a counterintuitive phenomenon of comparative statics: even if all agents in the group become more risk seeking, the portfolio manager acting on behalf of the group may not necessarily allocate a larger proportion of investments to risky assets, which is in sharp contrast to the case of risk-averse agents.
A new modular approximate Bayesian inferential framework is proposed that enables fast calculation of probabilistic predictions of future option prices. We exploit multiple information sources, including daily spot returns, high-frequency spot data and option prices. A benefit of this modular Bayesian approach is that it allows us to work with the theoretical option pricing model, without needing to specify an arbitrary statistical model that links the theoretical prices to their observed counterparts. We show that our approach produces accurate probabilistic predictions of option prices in realistic scenarios and, despite not explicitly modelling pricing errors, the method is shown to be robust to their presence. Predictive accuracy based on the Heston stochastic volatility model, with predictions produced via rapid real-time updates, is illustrated empirically for short-maturity options.
Decentralized Finance (DeFi) and smart contracts are the next generation and fast-growing market for quick and safe interaction between lenders and borrowers. However for maintaining a streamline ecosystem it is necessary to understand the risk associated with the particular user under consideration. In this paper we have developed 'On Chain Credit Risk Score' of a wallet is an answer to quantifying the risk of the particular wallet, namely the probability that the particular wallet may have a loan liquidated. 'On Chain Credit Risk Score (OCCR Score)' of wallets, will help lending borrowing protocols and other DeFi institutes to understand the risk involved in giving out loans to a wallet and may change the Loan-to-Value (LTV) Ratio and subsequently the Liquidation Threshold (LT) if required.
The bullwhip study has received a lot of attention in the literature, but with conflicting results, especially in the context of data aggregation. In this paper, we investigate three widely studied factors in bullwhip measurement: time aggregation, product aggregation, and seasonality. In time aggregation, we decompose the variance into two components: the expectation of the subset variances and the variance of subset expectations, thus decomposing the bullwhip ratio into four components to explore the underlying mechanism of time aggregation. In product aggregation, the bullwhip ratio is analyzed in the context of products with either uncorrelated or correlated demands and orders. Seasonality is also examined to study its effect on the bullwhip ratio. Our key findings are: (a) Time aggregation can increase, decrease, or maintain the bullwhip ratio in different scenarios. (b) Aggregated bullwhip ratio of uncorrelated products is a weighted average of bullwhip ratios from individual products, with corresponding demand variance as the weights. However, aggregated bullwhip ratio of correlated products could break the boundaries. (c) Seasonality can be considered as a standalone product with a bullwhip ratio of one, which can drive the overall bullwhip ratio closer to one.
An axiomatic approach to macroeconomics based on the mathematical structure of thermodynamics is presented. It deduces relations between aggregate properties of an economy, concerning quantities and flows of goods and money, prices and the value of money, without any recourse to microeconomic foundations about the preferences and actions of individual economic agents. The approach has three important payoffs. 1) it provides a new and solid foundation for aspects of standard macroeconomic theory such as the existence of market prices, the value of money, the meaning of inflation, the symmetry and negative-definiteness of the macro-Slutsky matrix, and the Le Chatelier-Samuelson principle, without relying on implausibly strong rationality assumptions over individual microeconomic agents. 2) the approach generates new results, including implications for money flow and trade when two or more economies are put in contact, in terms of new concepts such as economic entropy, economic temperature, goods' values and money capacity. Some of these are related to standard economic concepts (eg marginal utility of money, market prices). Yet our approach derives them at a purely macroeconomic level and gives them a meaning independent of usual restrictions. Others of the concepts, such as economic entropy and temperature, have no direct counterparts in standard economics, but they have important economic interpretations and implications, as aggregate utility and the inverse marginal aggregate utility of money, respectively. 3) this analysis promises to open up new frontiers in macroeconomics by building a bridge to ideas from non-equilibrium thermodynamics. More broadly, we hope that the economic analogue of entropy (governing the possible transitions between states of economic systems) may prove to be as fruitful for the social sciences as entropy has been in the natural sciences.
Traditional genetic programming (GP) often struggles in stock alpha factor discovery due to its vast search space, overwhelming computational burden, and sporadic effective alphas. We find that GP performs better when focusing on promising regions rather than random searching. This paper proposes a new GP framework with carefully chosen initialization and structural constraints to enhance search performance and improve the interpretability of the alpha factors. This approach is motivated by and mimics the alpha searching practice and aims to boost the efficiency of such a process. Analysis of 2020-2024 Chinese stock market data shows that our method yields superior out-of-sample prediction results and higher portfolio returns than the benchmark.
This paper contributes to the literature on parametric demand estimation by using deep learning to model consumer preferences. Traditional econometric methods often struggle with limited within-product price variation, a challenge addressed by the proposed neural network approach. The proposed method estimates the functional form of the demand and demonstrates higher performance in both simulations and empirical applications. Notably, under low price variation, the machine learning model outperforms econometric approaches, reducing the mean squared error of initial price parameter estimates by nearly threefold. In empirical setting, the ML model consistently predicts a negative relationship between demand and price in 100% of cases, whereas the econometric approach fails to do so in 20% of cases. The suggested model incorporates a wide range of product characteristics, as well as prices of other products and competitors.
We study a problem of optimal irreversible investment and emission reduction formulated as a nonzero-sum dynamic game between an investor with environmental preferences and a firm. The game is set in continuous time on an infinite-time horizon. The firm generates profits with a stochastic dynamics and may spend part of its revenues towards emission reduction (e.g., renovating the infrastructure). The firm's objective is to maximize the discounted expectation of a function of its profits. The investor participates in the profits and may decide to invest to support the firm's production capacity. The investor uses a profit function which accounts for both financial and environmental factors. Nash equilibria of the game are obtained via a system of variational inequalities. We formulate a general verification theorem for this system in a diffusive setup and construct an explicit solution in the zero-noise limit. Our explicit results and numerical approximations show that both the investor's and the firm's optimal actions are triggered by moving boundaries that increase with the total amount of emission abatement.
We investigate how advanced large language models (LLMs), specifically GPT-4, process corporate disclosures to forecast earnings. Using earnings press releases issued around GPT-4's knowledge cutoff date, we address two questions: (1) Do GPT-generated earnings forecasts outperform analysts in accuracy? (2) How is GPT's performance related to its processing of textual and quantitative information? Our findings suggest that GPT forecasts are significantly less accurate than those of analysts. This underperformance can be traced to GPT's distinct textual and quantitative approaches: its textual processing follows a consistent, generalized pattern across firms, highlighting its strengths in language tasks. In contrast, its quantitative processing capabilities vary significantly across firms, revealing limitations tied to the uneven availability of domain-specific training data. Additionally, there is some evidence that GPT's forecast accuracy diminishes beyond its knowledge cutoff, underscoring the need to evaluate LLMs under hindsight-free conditions. Overall, this study provides a novel exploration of the "black box" of GPT-4's information processing, offering insights into LLMs' potential and challenges in financial applications.
Economic growth in Sweden during the early 20th Century was largely driven by industry. A significant contributor to this growth was the installation of different kinds of engines used to power factories. We use newly digitized data on engines and their energy source by industry sector, and combine this with municipality-level data of workers per industry sector to construct a new variable reflecting economic output using dirty engines. In turn, we assess the average externality of dirty output on mortality in the short-run, as defined by deaths over the population in the baseline year. Our results show substantial increases of up to 17% higher mortality in cities where large increases to dirty engine installations occurred, which is largely driven by the elderly. We also run a placebo test using clean powered industry and find no effect on mortality.
In the literature, insurance and reinsurance pricing is typically determined by a premium principle, characterized by a risk measure that reflects the policy seller's risk attitude. Building on the work of Meyers (1980) and Chen et al. (2016), we propose a new performance-based variable premium scheme for reinsurance policies, where the premium depends on both the distribution of the ceded loss and the actual realized loss. Under this scheme, the insurer and the reinsurer face a random premium at the beginning of the policy period. Based on the realized loss, the premium is adjusted into either a ''reward'' or ''penalty'' scenario, resulting in a discount or surcharge at the end of the policy period. We characterize the optimal reinsurance policy from the insurer's perspective under this new variable premium scheme. In addition, we formulate a Bowley optimization problem between the insurer and the monopoly reinsurer. Numerical examples demonstrate that, compared to the expected-value premium principle, the reinsurer prefers the variable premium scheme as it reduces the reinsurer's total risk exposure.
Virtual bidding plays an important role in two-settlement electric power markets, as it can reduce discrepancies between day-ahead and real-time markets. Renewable energy penetration increases volatility in electricity prices, making accurate forecasting critical for virtual bidders, reducing uncertainty and maximizing profits. This study presents a Transformer-based deep learning model to forecast the price spread between real-time and day-ahead electricity prices in the ERCOT (Electric Reliability Council of Texas) market. The proposed model leverages various time-series features, including load forecasts, solar and wind generation forecasts, and temporal attributes. The model is trained under realistic constraints and validated using a walk-forward approach by updating the model every week. Based on the price spread prediction results, several trading strategies are proposed and the most effective strategy for maximizing cumulative profit under realistic market conditions is identified through backtesting. The results show that the strategy of trading only at the peak hour with a precision score of over 50% produces nearly consistent profit over the test period. The proposed method underscores the importance of an accurate electricity price forecasting model and introduces a new method of evaluating the price forecast model from a virtual bidder's perspective, providing valuable insights for future research.
This paper presents the system description of our entry for the COLING 2025 FMD challenge, focusing on misinformation detection in financial domains. We experimented with a combination of large language models, including Qwen, Mistral, and Gemma-2, and leveraged pre-processing and sequential learning for not only identifying fraudulent financial content but also generating coherent, and concise explanations that clarify the rationale behind the classifications. Our approach achieved competitive results with an F1-score of 0.8283 for classification, and ROUGE-1 of 0.7253 for explanations. This work highlights the transformative potential of LLMs in financial applications, offering insights into their capabilities for combating misinformation and enhancing transparency while identifying areas for future improvement in robustness and domain adaptation.
We propose risk models for a portfolio of risks, each following a compound Poisson distribution, with dependencies introduced through a family of tree-based Markov random fields with Poisson marginal distributions inspired in C\^ot\'e et al. (2024b, arXiv:2408.13649). The diversity of tree topologies allows for the construction of risk models under several dependence schemes. We study the distribution of the random vector of risks and of the aggregate claim amount of the portfolio. We perform two risk management tasks: the assessment of the global risk of the portfolio and its allocation to each component. Numerical examples illustrate the findings and the efficiency of the computation methods developed throughout. We also show that the discussed family of Markov random fields is a subfamily of the multivariate Poisson distribution constructed through common shocks.
High-frequency trading (HFT) represents a pivotal and intensely competitive domain within the financial markets. The velocity and accuracy of data processing exert a direct influence on profitability, underscoring the significance of this field. The objective of this work is to optimise the real-time processing of data in high-frequency trading algorithms. The dynamic feature selection mechanism is responsible for monitoring and analysing market data in real time through clustering and feature weight analysis, with the objective of automatically selecting the most relevant features. This process employs an adaptive feature extraction method, which enables the system to respond and adjust its feature set in a timely manner when the data input changes, thus ensuring the efficient utilisation of data. The lightweight neural networks are designed in a modular fashion, comprising fast convolutional layers and pruning techniques that facilitate the expeditious completion of data processing and output prediction. In contrast to conventional deep learning models, the neural network architecture has been specifically designed to minimise the number of parameters and computational complexity, thereby markedly reducing the inference time. The experimental results demonstrate that the model is capable of maintaining consistent performance in the context of varying market conditions, thereby illustrating its advantages in terms of processing speed and revenue enhancement.