Localisation and circularity in perishable food supply chains are essential for sustainability. Poor allocation of time-sensitive food leads to waste, higher transport emissions, and unnecessary long-distance sourcing. Algorithms used in digital trading platforms and allocation systems can help address these problems by improving how local supply is matched with demand under real operational constraints. This paper examines localisation and circularity in the UK apple supply chain. Apples are an informative case because they are perishable, consumed fresh as dessert fruit, used as inputs across multiple food industries, and generate valuable by-products. We present a weighted-sum mixed-integer linear programming formulation for supply-demand allocation. The model encodes a single global objective with explicit weights on four operational criteria: price matching, quantity alignment, freshness requirements, and geographic distance. These weights make priorities explicit and adjustable, enabling transparent balancing between economic and sustainability considerations. The framework also supports the circulation of unallocated supply across allocation cycles. Using a realistic apple supply-demand dataset, we evaluate allocation outcomes under different priority settings. Results indicate that allocation outcomes are strongly shaped by both priority settings and the structure of the underlying supply network characteristics.
Purpose: This paper examines the prevalence of long COVID across different demographic groups in the U.S. and the extent to which workers with impairments associated with long COVID have engaged in pandemic-related remote work. Methods: We use the U.S. Household Pulse Survey to evaluate the proportion of all adults who self-reported to (1) have had long COVID, and (2) have activity limitations due to long COVID. We also use data from the U.S. Current Population Survey to estimate linear probability regressions for the likelihood of pandemic-related remote work among workers with and without disabilities. Results: Findings indicate that women, Hispanic people, sexual and gender minorities, individuals without four-year college degrees, and people with preexisting disabilities are more likely to have long COVID and to have activity limitations from long COVID. Remote work is a reasonable arrangement for people with such activity limitations and may be an unintentional accommodation for some people who have undisclosed disabilities. However, this study shows that people with disabilities were less likely than people without disabilities to perform pandemic-related remote work. Conclusion: The data suggest this disparity persists because people with disabilities are clustered in jobs that are not amenable to remote work. Employers need to consider other accommodations, especially shorter workdays and flexible scheduling, to hire and retain employees who are struggling with the impacts of long COVID.
Background. Long COVID symptoms (which include brain fog, depression, and fatigue) are mild at best and debilitating at worst. Some U.S. health surveys have found that women, lower income individuals, and those with less education are overrepresented among adults with long COVID, but these studies do not address intersectionality. Methods. We use 10 rounds of Household Pulse Survey (HPS) data from 2022 to 2023 to perform an intersectional analysis using descriptive statistics that evaluate the prevalence of long COVID and the interference of long COVID symptoms with day-to-day activities. We also estimate multivariate logistic regressions that relate the odds of having long COVID and activity limitations due to long COVID to a set of individual characteristics and intersections by sex, race/ethnicity, education, and sexual orientation and gender identity. Results. Women, some people of color, sexual and gender minorities, and people without college degrees are more likely to have long COVID and to have activity limitations from long COVID. Intersectional analysis reveals a striking step-like pattern: college-educated men have the lowest prevalence of long COVID while women without college educations have the highest prevalence. Daily activity limitations are more evenly distributed across demographics, but a different step-like pattern is present: fewer women with degrees have activity limitations while limitations are more widespread among men without degrees. Regression results confirm the negative association of long COVID with being a woman, less educated, Hispanic, and a sexual and gender minority, while results for the intersectional effects are more nuanced. Conclusions. Results point to systematic disparities in health, highlighting the need for policies that increase access to quality healthcare, strengthen the social safety net, and reduce economic precarity.
This article extends the analysis of Atkinson, Foley, and Ganz in "Beyond the Spoiler Effect: Can Ranked-Choice Voting Solve the Problem of Political Polarization?". Their work uses a one-dimensional spatial model based on survey data from the Cooperative Election Survey (CES) to examine how instant-runoff voting (IRV) and Condorcet methods promote candidate moderation. Their model assumes an idealized electoral environment in which all voters possess complete information regarding candidates' ideological positions, all voters provide complete preference rankings, etc. Under these assumptions, their results indicate that Condorcet methods tend to yield winners who are substantially more moderate than those produced by IRV. We construct new models based on CES data which take into account more realistic voter behavior, such as the presence of partial ballots. Our general finding is that under more realistic models the differences between Condorcet methods and IRV largely disappear, implying that in real-world settings the moderating effect of Condorcet methods may not be nearly as strong as what is suggested by more theoretical models.
Spatial autocorrelation in regression models can lead to downward biased standard errors and thus incorrect inference. The most common correction in applied economics is the spatial heteroskedasticity and autocorrelation consistent (HAC) standard error estimator introduced by Conley (1999). A critical input is the kernel bandwidth: the distance within which residuals are allowed to be correlated. However, this is still an unresolved problem and there is no formal guidance in the literature. In this paper, I first document that the relationship between the bandwidth and the magnitude of spatial HAC standard errors is inverse-U shaped. This implies that both too narrow and too wide bandwidths lead to underestimated standard errors, contradicting the conventional wisdom that wider bandwidths yield more conservative inference. I then propose a simple, non-parametric, data-driven bandwidth selector based on the empirical covariogram of regression residuals. In extensive Monte Carlo experiments calibrated to empirically relevant spatial correlation structures across the contiguous United States, I show that the proposed method controls the false positive rate at or near the nominal 5% level across a wide range of spatial correlation intensities and sample configurations. I compare six kernel functions and find that the Bartlett and Epanechnikov kernels deliver the best size control. An empirical application using U.S. county-level data illustrates the practical relevance of the method. The R package SpatialInference implements the proposed bandwidth selection method.
We propose a Random Rule Model (RRM) in which behavior is generated by switching among a small library of transparent, parameter-free decision rules. A differentiable gate learns environment-dependent rule propensities, producing an interpretable mixture over named procedures. We develop a global identification theory based on two verifiable conditions on the observed support. Applied to 10,000 binary lottery problems, rule-gating substantially outperforms structured neural benchmarks based on expected utility and prospect theory, approaching the most flexible benchmark while remaining highly restrictive under permutation-fit tests, and retains predictive content on an independent dataset. Mechanism diagnostics reveal that extreme-outcome screening, salience, and attention rules carry the largest responsibility weights, with systematic shifts along tradeoff complexity and dispersion asymmetry. Robustness checks confirm that the findings are not driven by the ex-ante library choice, marginal dominance relationships, or the availability of additional regressors.
In causal analysis, understanding the causal mechanisms through which an intervention or treatment affects an outcome is often of central interest. We propose a test to evaluate (i) whether the causal effect of a treatment that is randomly assigned conditional on covariates is fully mediated by, or operates exclusively through, observed intermediate outcomes (referred to as mediators or surrogate outcomes), and (ii) whether the various causal mechanisms operating through different mediators are identifiable conditional on covariates. We demonstrate that if both full mediation and identification of causal mechanisms hold, then the conditionally random treatment is conditionally independent of the outcome given the mediators and covariates. Furthermore, we extend our framework to settings with non-randomly assigned treatments. We show that, in this case, full mediation remains testable, while identification of causal mechanisms is no longer guaranteed. We propose a double machine learning framework for implementing the test that can incorporate high-dimensional covariates and is root-n consistent and asymptotically normal under specific regularity conditions. We also present a simulation study demonstrating good finite-sample performance of our method, along with two empirical applications revisiting randomized experiments on maternal mental health and social norms.
We study how to allocate resources to participants who can strategically misrepresent their deservingness at a cost. A principal assigns item(s) (or money) among multiple agents on the basis of their costly signals. Each agent's signal reflects their private type in the absence of misrepresentation but can be inflated above their true type at a cost. The principal is a social planner who aims to maximize the weighted average of matching efficiency and a utilitarian objective. Strategic misrepresentation introduces novel incentive-compatibility constraints, under which we characterize the optimal mechanism. We apply our characterization to two kinds of markets, distinguished by resource scarcity, and show that the principal strictly benefits from randomizing the allocations based on costly signals when the population of participants is large enough. Interestingly, in large markets with scarce resources, the format of the optimal mechanism converges to a winner-takes-all contest; however, there is a non-diminishing value in randomizing allocations to middle types as the population of participants grows.
We introduce inference methods for score decompositions, which partition scoring functions for predictive assessment into three interpretable components: miscalibration, discrimination, and uncertainty. Our estimation and inference relies on a linear recalibration of the forecasts, which is applicable to general multi-step ahead point forecasts such as means and quantiles due to its validity for both smooth and non-smooth scoring functions. This approach ensures desirable finite-sample properties, enables asymptotic inference, and establishes a direct connection to the classical Mincer-Zarnowitz regression. The resulting inference framework facilitates tests for equal forecast calibration or discrimination, which yield three key advantages. They enhance the information content of predictive ability tests by decomposing scores, deliver higher statistical power in certain scenarios, and formally connect scoring-function-based evaluation to traditional calibration tests, such as financial backtests. Applications demonstrate the method's utility. We find that for survey inflation forecasts, discrimination abilities can differ significantly even when overall predictive ability does not. In an application to financial risk models, our tests provide deeper insights into the calibration and information content of volatility and Value-at-Risk forecasts. By disentangling forecast accuracy from backtest performance, the method exposes critical shortcomings in current banking regulation.
We present a new method for computation of the index of completely mixed equilibria in finite games, based on the work of Eisenbud et al.(1977). We apply this method to solving two questions about the relation of the index of equilibria and the index of fixed points, and the index of equilibria and payoff-robustness: any integer can be the index of an isolated completely mixed equilibrium of a finite game. In a particular class of isolated completely mixed equilibria, called monogenic, the index can be $0$, $+1$ or $-1$ only. In this class non-zero index is equivalent to payoff-robustness. We also discuss extensions of the method of computation to extensive-form games, and cases where the equilibria might be located on the boundary of the strategy set.
We study the design of fair allocation rules for the abatement of riparian pollution. To do so, we consider the so-called river pollution claims model, recently introduced by Yang et al. (2025) to distribute a budget of emissions permits among agents (cities, provinces, or countries) located along a river. In such a model, each agent has a claim reflecting population, emission history, and business-as-usual emissions, and the issue is to allocate among them a budget that is lower (or equal) than the aggregate claim. For environmental reasons, the specific location along the river where pollutants are emitted is an important concern (the more upstream the location is the higher the damage of polluting the river). We characterize a class of geometric rules that adjust proportional allocations to compromise between fairness and environmental concerns. Our class is an alternative to the one proposed by Yang et al. (2025). We compare both alternatives through an axiomatic study, as well as an illustration for the case study of the Tuojiang Basin in China.
Western governments have adopted an assortment of counter-hybrid threat measures to defend against hostile actions below the conventional military threshold. The impact of these measures is unclear because of the ambiguity of hybrid threats, their cross-domain nature, and uncertainty about how countermeasures shape adversarial behavior. This paper offers a novel approach to clarifying this impact by unifying previously bifurcating hybrid threat modeling methods through a (multi-agent) influence diagram framework. The model balances the costs of countermeasures, their ability to dissuade the adversary from executing hybrid threats, and their potential to mitigate the impact of hybrid threats. We run 1000 semi-synthetic variants of a real-world-inspired scenario simulating the strategic interaction between attacking agent A and defending agent B over a cyber attack on critical infrastructure to explore the effectiveness of a set of five different counter-hybrid threat measures. Counter-hybrid measures range from strengthening resilience and denial of the adversary's ability to execute a hybrid threat to dissuasion through the threat of punishment. Our analysis primarily evaluates the overarching characteristics of counter-hybrid threat measures. This approach allows us to generalize the effectiveness of these measures and examine parameter impact sensitivity. In addition, we discuss policy relevance and outline future research avenues.
The growing use of unstructured text in business research makes topic modeling a central tool for constructing explanatory variables from reviews, social media, and open-ended survey responses, yet existing approaches function poorly as measurement instruments. Prior work shows that textual content predicts outcomes such as sales, satisfaction, and firm performance, but probabilistic models often generate conceptually diffuse topics, neural topic models are difficult to interpret in theory-driven settings, and large language model approaches lack standardization, stability, and alignment with document-level representations. We introduce LX Topic, a neural topic method that conceptualizes topics as latent linguistic constructs and produces calibrated document-level topic proportions for empirical analysis. LX Topic builds on FASTopic to ensure strong document representativeness and integrates large language model refinement at the topic-word level using alignment and confidence-weighting mechanisms that enhance semantic coherence without distorting document-topic distributions. Evaluations on large-scale Amazon and Yelp review datasets demonstrate that LX Topic achieves the highest overall topic quality relative to leading models while preserving clustering and classification performance. By unifying topic discovery, refinement, and standardized output in a web-based system, LX Topic establishes topic modeling as a reproducible, interpretable, and measurement-oriented instrument for marketing research and practice.
The difference-in-differences (DiD) design is a quasi-experimental method for estimating treatment effects. In staggered DiD with multiple treatment groups and periods, estimation based on the two-way fixed effects model yields negative weights when averaging heterogeneous group-period treatment effects into an overall effect. To address this issue, we first define group-period average treatment effects on the treated (ATT), and then define groupwise, periodwise, dynamic, and overall ATTs nonparametrically, so that the estimands are model-free. We propose doubly robust estimators for these types of ATTs in the form of augmented inverse variance weighting (AIVW). The proposed framework allows time-varying covariates that partially explain the time trends in outcomes. Even if part of the working models is misspecified, the proposed estimators still consistently estimate the parameter of interest. The asymptotic variance can be explicitly computed from influence functions. Under a homoskedastic working model, the AIVW estimator is simplified to an augmented inverse probability weighting (AIPW) estimator. We demonstrate the desirable properties of the proposed estimators through simulation and an application that compares the effects of a parallel admission mechanism with immediate admission on the China National College Entrance Examination.
We study infinite-horizon Markov decision processes (MDPs) where the decision maker evaluates each of her strategies by aggregating the infinite stream of expected stage-rewards. The crucial feature of our approach is that the aggregation is performed by means of a given diffuse charge (a diffuse finitely additive probability measure) on the set of stages. The results of Neyman [2023] imply that in this setting, in every MDP with finite state and action spaces, the decision maker has a pure optimal strategy as long as the diffuse charge satisfies the time value of money principle. His result raises the question of existence of an optimal strategy without additional assumptions on the aggregation charge. We answer this question in the negative with a counterexample. With a delicately constructed aggregation charge, the MDP has no optimal strategy at all, neither pure nor randomized.
Large language models (LLMs) are trained on enormous amounts of data and encode knowledge in their parameters. We propose a pipeline to elicit causal relationships from LLMs. Specifically, (i) we sample many documents from LLMs on a given topic, (ii) we extract an event list from from each document, (iii) we group events that appear across documents into canonical events, (iv) we construct a binary indicator vector for each document over canonical events, and (v) we estimate candidate causal graphs using causal discovery methods. Our approach does not guarantee real-world causality. Rather, it provides a framework for presenting the set of causal hypotheses that LLMs can plausibly assume, as an inspectable set of variables and candidate graphs.
We study the deployment performance of machine learning based enforcement systems used in cryptocurrency anti money laundering (AML). Using forward looking and rolling evaluations on Bitcoin transaction data, we show that strong static classification metrics substantially overstate real world regulatory effectiveness. Temporal nonstationarity induces pronounced instability in cost sensitive enforcement thresholds, generating large and persistent excess regulatory losses relative to dynamically optimal benchmarks. The core failure arises from miscalibration of decision rules rather than from declining predictive accuracy per se. These findings underscore the fragility of fixed AML enforcement policies in evolving digital asset markets and motivate loss-based evaluation frameworks for regulatory oversight.
Counterfactuals in quantitative trade and spatial models are functions of the current state of the world and the model parameters. Common practice treats the current state of the world as perfectly observed, but there is good reason to believe that it is measured with error. This paper provides tools for quantifying uncertainty about counterfactuals when the current state of the world is measured with error. I recommend an empirical Bayes approach to uncertainty quantification, and show that it is both practical and theoretically justified. I apply the proposed method to the settings in Adao, Costinot, and Donaldson (2017) and Allen and Arkolakis (2022) and find non-trivial uncertainty about counterfactuals.
When treatments are non-randomly assigned, continuous, and yield heterogeneous effects at the same intensity, causal identification becomes particularly challenging. In such contexts, existing approaches often fail to provide policy-relevant estimates of the relationship between treatment intensity and outcomes, especially in the presence of limited common support. To fill this gap, we introduce the Clustered Dose-Response Function (Cl-DRF), a novel estimator designed to uncover the continuous causal relationship between treatment intensity and the dependent variable across distinct subgroups. Our approach leverages both theoretical and data-driven sources of heterogeneity, relying on relaxed versions of the conditional independence and positivity assumptions that are plausible across various observational settings. We apply the Cl-DRF estimator to estimate subgroup-specific dose-response relationships between European Cohesion Funds and economic growth. In contrast to much of the literature, higher funding increases growth in more developed regions without diminishing returns, while limited absorptive capacity prevents other regions from fully benefiting.
Contest success function (CSF) maps contestants' efforts to their winning probability. This paper provides axiomatizations of CSFs with headstarts. The results extend the classic axiomatization of the Tullock CSF and connect to CSFs that allow for draws. The central axiom is relative homogeneity of counterfactual deviation, which requires the pairwise influence of one contestant's effort on opponent's probabilistic allocation to be scale-invariant. Two fairness axioms and no advantageous reallocation further restrict the admissible functional forms with headstarts. We also introduce dummy consistency, requiring allocations to be consistent with and without inactive contestants, to clarify the relationship with earlier axiomatic work that rules out headstarts. Finally, we discuss an extension that drops the assumption of full allocation.
We study price regulation for a monopolist operating in networked markets with demand spillovers. Achieving efficiency requires price reductions proportional to consumers' Katz-Bonacich centralities, which generally cannot be implemented by commonly used price regulations. Moreover, these regulations become asymptotically welfare neutral as spillovers grow. Nevertheless, some price regulations may still benefit consumers. In particular, average-price regulation robustly increases consumer surplus. By contrast, banning price discrimination increases consumer surplus only when more central consumers have higher intrinsic willingness to pay.
Advanced space technology systems often face high fixed costs, can serve limited non-government demand, and are significantly driven by non-market motivations. While increased entrepreneurial activity and national ambitions in space have encouraged planners at public space agencies to develop markets around such systems, the very factors that make the recent growth of the space economy so remarkable also challenge planners' efforts to develop and sustain markets for space-related goods and services. I propose a graphical framework to visualize the number of competitors a market can sustain as a function of the industry's cost structure; the distribution of government support across direct purchases, direct investments, and shared infrastructure; and the magnitude of non-government demand. Building on public goods theory, the framework shows how marginal dollars invested in shared infrastructure can create non-rival benefits supporting more competitors per dollar than direct purchases or subsidies. I demonstrate the framework with a stylized application inspired by NASA's Commercial LEO Destinations program. Under cost and demand conditions consistent with public data, independent stations generate industry-wide losses of \$355 million annually, while shared core infrastructure enables industry-wide profits of \$154 million annually. I also outline key directions for future research on public investment and market development strategies for advanced technologies.
Nonparametric regression and regression-discontinuity designs suffer from smoothing bias that distorts conventional confidence intervals. Solutions based on robust bias correction (RBC) are now central to the economist's toolbox. In this paper, we establish a novel connection between RBC methods and bootstrap prepivoting. Revisiting RBC through the lens of bootstrapping allows us to develop a novel bias correction procedure which delivers improved nonparametric inference. The resulting confidence intervals are 17% shorter than the usual intervals employed in curve estimation and regression discontinuity designs, without compromising asymptotic coverage. This holds regardless of evaluation point location, bandwidth choice, or regressor and error distribution.
The growing adoption of artificial intelligence (AI) technologies has heightened interest in the labor market value of AI related skills, yet causal evidence on their role in hiring decisions remains scarce. This study examines whether AI skills serve as a positive hiring signal and whether they can offset conventional disadvantages such as older age or lower formal education. We conducted an experimental survey with 1,725 recruiters from the United Kingdom, the United States and Germany. Using a paired conjoint design, recruiters evaluated hypothetical candidates represented by synthetically designed resumes. Across three occupations of graphic design, office assistance, and software engineering, AI skills significantly increase interview invitation probabilities by approximately 8 to 15 percentage points, compared with candidates without such skills. AI credentials, such as university or company backed skill certificates, only lead to a moderate increase in invitation probabilities compared with self declaration of AI skills. AI skills also partially or fully offset disadvantages related to age and lower education, with effects strongest for office assistants, for whom formal AI certificates play a significant additional compensatory role. Effects are weaker for graphic designers, consistent with more skeptical recruiter attitudes toward AI in creative work. Finally, recruiters own background and AI usage significantly moderate these effects. Overall, the findings demonstrate that AI skills function as a powerful hiring signal and can mitigate traditional labor market disadvantages, with implications for workers skill acquisition strategies and firms recruitment practices.
The classical theory of efficient allocations of an aggregate endowment in a pure-exchange economy has hitherto primarily focused on the Pareto-efficiency of allocations, under the implicit assumption that transfers between agents are frictionless, and hence costless to the economy. In this paper, we argue that certain transfers cause frictions that result in costs to the economy. We show that these frictional costs are tantamount to a form of subadditivity of the cost of transferring endowments between agents. We suggest an axiomatic study of allocation mechanisms, that is, the mechanisms that transform feasible allocations into other feasible allocations, in the presence of such transfer costs. Among other results, we provide an axiomatic characterization of those allocation mechanisms that admit representations as robust (worst-case) linear allocation mechanisms, as well as those mechanisms that admit representations as worst-case conditional expectations. We call the latter Robust Conditional Mean Allocation mechanisms, and we relate our results to the literature on (decentralized) risk sharing within a pool of agents.
Score-driven (SD) models are a standard tool in statistics and econometrics, with applications in hundreds of published articles in the past decade. We provide an information-theoretic characterization of SD updates based on reductions in the expected Kullback-Leibler (EKL) divergence relative to the true -- but unknown -- data-generating density. EKL reductions occur if and only if the expected update direction aligns with the expected score; i.e., their inner product should be positive. This equivalence condition uniquely identifies SD updates (including scaled or clipped variants) as being EKL reducing, even in non-concave, multivariate, and misspecified settings. We further derive explicit bounds on admissible learning rates in terms of score moments, linking SD methods to adaptive optimization techniques. By contrast, alternative performance measures in the literature impose stronger conditions (e.g., concave logarithmic densities) and do not characterize SD updates: other updating rules may improve these measures, while SD updates need not. Our results provide a rigorous justification for SD models and establish EKL as their natural information-theoretic foundation.