Research

Partial Identification Moment Inequalities Randomization Tests Regression Discontinuity Covariate-Adaptive Randomization Clustered Data Treatment Effects Stochastic Dominance Nonparametric Identification

Working Papers

Current work in progress. Both papers develop new inferential tools using induced order statistics and study their convergence properties.

Working Paper

On the Rate of Convergence of Induced Ordered Statistics and Their Applications

Federico A. Bugni, Ivan A. Canay, Deborah Kim

Induced order statistics (IOS) arise when sample units are reordered according to the value of an auxiliary variable, and the associated responses are analyzed in that induced order. We derive sharp convergence rates for IOS under weak assumptions that accommodate both interior and boundary conditioning points — addressing limitations in existing methods used for regression discontinuity designs and k-nearest-neighbor approaches. The paper establishes sharp marginal rates for approximating target conditional distributions in Hellinger and total variation distances, uncovering a clear trade-off between smoothness and speed of convergence, with explicit conditions governing how the number of nearest neighbors can grow with sample size.
Working Paper

Testing Conditional Stochastic Dominance at Target Points

Federico A. Bugni, Ivan A. Canay, Deborah Kim

This paper introduces a test for conditional stochastic dominance (CSD) at specific values of the conditioning covariates, referred to as target points. The test is relevant for analyzing income inequality, evaluating treatment effects, and studying discrimination. We propose a Kolmogorov–Smirnov-type test statistic using induced order statistics from independent samples. Notably, the test features a data-independent critical value, eliminating the need for resampling techniques such as the bootstrap. Our approach avoids kernel smoothing and parametric assumptions, relying instead on a tuning parameter to select relevant observations. We establish asymptotic validity under weak regularity conditions and show that the test reduces, in the limit experiment, to a problem analogous to testing unconditional stochastic dominance in finite samples. Monte Carlo simulations confirm strong finite-sample performance.

Partial Identification & Moment Inequalities

This strand studies inference in models where the data restrict parameters to a set rather than a point. Work covers confidence regions, tests, subvector inference, specification testing, and computational guidance for moment inequality models.

Forthcoming

A User's Guide for Inference in Models Defined by Moment Inequalities

Ivan A. Canay, Gastón Illanes, Amílcar Velez

Journal of Econometrics, accepted 2026

Models defined by moment inequalities have become a standard modeling framework for empirical economists across a wide range of fields. We present a guide to empirical practice intended to help applied researchers navigate all the decisions required to frame a model as a moment inequality model and construct confidence intervals for the parameters of interest. The template has four steps: (a) a behavioral decision model, (b) moving from the decision model to a moment inequality model, (c) choosing a test statistic and critical value, and (d) accounting for computational challenges. Each step is illustrated in an empirical application studying identification of expected sunk costs of offering a product in a market.

Practical and Theoretical Advances for Inference in Partially Identified Models

Ivan A. Canay, Azeem M. Shaikh

Advances in Economics and Econometrics (Eleventh World Congress), Vol. 2, pp. 271–306, Cambridge University Press, 2017

This paper surveys recent literature on inference in partially identified models. After reviewing basic concepts — including the definition of a partially identified model and the identified set — we turn to the construction of confidence regions in partially identified settings, emphasizing the importance of requiring uniform consistency in level over relevant classes of distributions. The survey is mainly limited to models in which the identified set is characterized by a finite number of moment inequalities, or the closely related class in which the identified set is a function of such a set. We conclude with some thoughts on fruitful directions for future research.

Inference for Subvectors and Other Functions of Partially Identified Parameters in Moment Inequality Models

Federico A. Bugni, Ivan A. Canay, Xiaoxia Shi

Quantitative Economics, 8(1), pp. 1–38, 2017

This paper introduces a bootstrap-based inference method for functions of the parameter vector in a moment (in)equality model. The procedure controls asymptotic size uniformly over a large class of data distributions and improves upon the two existing methods with uniform size control — projection-based and subsampling inference. Relative to projection-based methods, ours weakly dominates in terms of finite-sample power, strictly dominates in terms of asymptotic power, and is typically less computationally demanding. Relative to subsampling, it strictly dominates in asymptotic power and appears less sensitive to the choice of tuning parameter.

Specification Test for Partially Identified Models Defined by Moment Inequalities

Federico A. Bugni, Ivan A. Canay, Xiaoxia Shi

Journal of Econometrics, 185(1), pp. 259–282, 2015

This paper studies specification testing in partially identified models defined by moment (in)equalities. The literature has suggested a test based on checking whether confidence sets are empty, referred to as Test BP. We propose two new specification tests — Test RS and Test RC — that achieve uniform asymptotic size control and dominate Test BP in terms of power in any finite sample and in the asymptotic limit.

Distortions of Asymptotic Confidence Size in Locally Misspecified Moment Inequality Models

Federico A. Bugni, Ivan A. Canay, Patrik Guggenberger

Econometrica, 80(4), pp. 1741–1768, 2012

This paper studies the behavior, under local misspecification, of several confidence sets commonly used for inference in moment (in)equality models. We propose the amount of asymptotic confidence size distortion as a criterion to choose among competing inference methods. Under weak assumptions we find two main results: (i) confidence sets based on subsampling and generalized moment selection suffer from the same degree of asymptotic confidence size distortion, despite the latter yielding smaller expected volume under correct specification; and (ii) the asymptotic confidence size of confidence sets based on the quasi-likelihood ratio statistic can be an arbitrarily small fraction of that based on the modified method of moments statistic.

EL Inference for Partially Identified Models: Large Deviations Optimality and Bootstrap Validity

Ivan A. Canay

Journal of Econometrics, 156(2), pp. 408–425, 2010

This paper addresses optimal inference for parameters that are partially identified in moment inequality models. Using a large-deviations criterion for optimality, I show that inference based on the empirical likelihood ratio statistic is optimal. I also introduce a new empirical likelihood bootstrap that provides a valid resampling method for moment inequality models, overcoming implementation challenges that arise from non-pivotal limit distributions. Monte Carlo simulations confirm favorable finite-sample performance of the proposed framework.

Randomization & Permutation Inference

This strand develops randomization-based and permutation tests valid under weak distributional assumptions. Applications include regression discontinuity designs, small-cluster settings, and tests for stochastic dominance at target points.

On the Implementation of Approximate Randomization Tests in Linear Models with a Small Number of Clusters

Yong Cai, Ivan A. Canay, Deborah Kim, Azeem M. Shaikh

Journal of Econometric Methods, 12(1), pp. 85–103, 2023

This paper provides a user's guide to the general theory of approximate randomization tests (ARTs) developed in Canay, Romano, and Shaikh (2017) when specialized to linear regressions with clustered data. An important feature of the methodology is that it applies to settings in which the number of clusters is small — even as small as five. We provide a step-by-step algorithmic description of how to implement the test and construct confidence intervals for the parameter of interest. We also present three novel results: the method admits an equivalent implementation based on weighted scores; the test and confidence intervals are invariant to studentization; and the confidence intervals are convex for scalar parameters. Companion R and Stata packages facilitate implementation and replication of the empirical exercises.

Testing Continuity of a Density via g-Order Statistics in the Regression Discontinuity Design

Federico A. Bugni, Ivan A. Canay

Journal of Econometrics, 221(1), pp. 138–159, 2021

In the regression discontinuity design (RDD), it is common practice to assess design credibility by testing continuity of the density of the running variable at the cutoff. We propose an approximate sign test for density continuity based on g-order statistics, studied under two asymptotic frameworks: one where the number of local observations is fixed as sample size grows, and one where it grows slowly. Under both, the test has limiting rejection probability under the null not exceeding the nominal level. The test is easy to implement, asymptotically valid under weaker conditions than competing methods, and exhibits finite-sample validity under stronger conditions. Simulations confirm good size control. We apply the test to Lee (2008)'s well-known RDD application on incumbency advantage.

Approximate Permutation Tests and Induced Order Statistics in the Regression Discontinuity Design

Ivan A. Canay, Vishal Kamat

The Review of Economic Studies, 85(3), pp. 1577–1608, 2018

It is common in the RDD to assess design credibility by testing whether means of baseline covariates do not change at the cutoff. We propose a permutation test based on induced ordered statistics for the null hypothesis of continuity of the distribution of baseline covariates, and introduce a novel asymptotic framework that keeps the number of local observations fixed as the overall sample size grows. The new test is easy to implement, asymptotically valid under weak conditions, exhibits finite-sample validity under stronger conditions, and has favorable power relative to tests based on means. Simulations show excellent size control across designs. We apply the test to Lee (2008)'s RDD application on incumbency advantage.

Randomization Tests Under an Approximate Symmetry Assumption

Ivan A. Canay, Joseph P. Romano, Azeem M. Shaikh

Econometrica, 85(3), pp. 1013–1030, 2017

This paper develops a theory of randomization tests under an approximate symmetry assumption. Randomization tests provide a general means of constructing tests that control size in finite samples when the data distribution exhibits symmetry under the null. We provide conditions under which the same construction yields tests that asymptotically control the probability of false rejection when the limiting distribution of a function of the data exhibits symmetry under the null. An important application is to settings where data are grouped into a fixed number of clusters with many observations each. We show the approximate symmetry requirement holds under weak assumptions, including cluster heterogeneity and cross-cluster dependence.

Hodges–Lehmann Optimality for Testing Moment Conditions

Ivan A. Canay, Taisuke Otsu

Journal of Econometrics, 171(1), pp. 45–53, 2012

This paper studies Hodges–Lehmann optimality of tests in a general setup, comparing tests by the exponential rates of growth to one of their power functions at a fixed alternative while keeping asymptotic sizes bounded. We present two sets of sufficient conditions for Hodges–Lehmann optimality that extend the scope of the analysis to setups not covered by existing conditions. Applications to testing moment conditions and overidentifying restrictions show that the empirical likelihood test satisfies our new conditions, and that the GMM and generalized empirical likelihood tests are Hodges–Lehmann optimal under mild primitive conditions, supporting the view that Hodges–Lehmann optimality is a weak asymptotic requirement.

Experiments & Clustered Data

This strand covers inference in randomized experiments with covariate-adaptive assignment and settings where treatment is assigned at the cluster level. Topics include bootstrap validity with few clusters, non-ignorable cluster sizes, and decomposition of treatment effects when outcomes are delayed.

Forthcoming

Decomposition and Interpretation of Treatment Effects in Settings with Delayed Outcomes

Federico A. Bugni, Ivan A. Canay, Steve McBride

Journal of Econometrics, accepted 2026

This paper studies settings where the outcome is not immediately realized after treatment assignment — a feature ubiquitous in empirical settings. The period between treatment and outcome allows other observed actions to occur and affect the outcome. We study several regression-based estimands routinely used to capture the average treatment effect and shed light on their interpretation in terms of ceteris paribus effects, indirect causal effects, and selection terms. The three most popular estimands do not generally satisfy strong sign preservation: they may be negative even when treatment positively affects the outcome conditional on any combination of other actions. The most popular regression that includes other actions as controls satisfies strong sign preservation if and only if those actions are mutually exclusive binary variables.

Inference for Cluster Randomized Experiments with Non-ignorable Cluster Sizes

Federico A. Bugni, Ivan A. Canay, Azeem M. Shaikh, Max Tabord-Meehan

Journal of Political Economy: Microeconomics, 3(2), pp. 255–288, 2025

This paper considers inference in cluster randomized experiments when cluster sizes are non-ignorable — that is, when treatment effects may depend non-trivially on cluster sizes. We frame the analysis in a super-population framework in which cluster sizes are random, departing from earlier work that treats them as fixed. We distinguish between the equally-weighted and size-weighted cluster-level average treatment effects, and provide inference methods for each in an asymptotic framework where the number of clusters grows and treatment is assigned via covariate-adaptive stratified randomization. We also permit sampling only a subset of units within each cluster and demonstrate implications for commonly used estimators.

The Wild Bootstrap with a "Small" Number of "Large" Clusters

Ivan A. Canay, Andrés Santos, Azeem M. Shaikh

The Review of Economics and Statistics, 103(2), pp. 346–363, 2021

This paper studies the wild bootstrap–based test of Cameron, Gelbach, and Miller (2008). Existing analyses require a large number of clusters. In an asymptotic framework with a small number of clusters, we provide conditions under which an unstudentized version of the test is valid, including homogeneity-like restrictions on the distribution of covariates. We establish that a studentized version may only overreject by an amount that decreases exponentially with the number of clusters. A qualitatively similar result holds for score bootstrap-based tests, which permit testing in nonlinear models.

Inference under Covariate Adaptive Randomization with Multiple Treatments

Federico A. Bugni, Ivan A. Canay, Azeem M. Shaikh

Quantitative Economics, 10(4), pp. 1747–1785, 2019

This paper studies inference in randomized controlled trials with covariate-adaptive randomization and multiple treatments. We study inference about the average effect of one or more treatments relative to other treatments or a control. Tests based on a fully saturated linear regression using the usual heteroskedasticity-consistent standard errors are invalid — they may have limiting rejection probability strictly greater than the nominal level. Tests based on suitable variance estimators that we provide are instead exact. For the case where treatment proportions are constant across strata, we additionally consider tests based on a linear regression with strata fixed effects.

Inference under Covariate Adaptive Randomization

Federico A. Bugni, Ivan A. Canay, Azeem M. Shaikh

Journal of the American Statistical Association, 113(524), pp. 1784–1796, 2018

This paper studies inference for the average treatment effect in randomized controlled trials with covariate-adaptive randomization — schemes that stratify on baseline covariates and assign treatment to achieve balance within strata. The usual two-sample t-test is conservative, with limiting rejection probability strictly less than the nominal level. A simple adjustment to the standard error yields a test that is exact. Analogous results hold for the t-test in a regression of outcomes on treatment and strata indicators. We also propose a covariate-adaptive permutation test that only permutes treatment status within strata and show it is exact for relevant special cases.

Other Methods

Work on detecting discrimination through outcome tests, testability of nonparametric identification with endogeneity, quantile regression with panel data, and moment selection in GMM.

On the Use of Outcome Tests for Detecting Bias in Decision Making

Ivan A. Canay, Magne Mogstad, Jack Mountjoy

The Review of Economic Studies, 91(4), pp. 2135–2167, 2024

The decisions of judges, lenders, journal editors, and other gatekeepers often generate significant disparities across affected groups. An important question is whether these disparities reflect relevant differences in individual characteristics or biased decision making. Becker (1957, 1993) proposed an outcome test of bias based on differences in post-decision outcomes across groups, inspiring a large empirical literature. We offer a methodological blueprint for this approach, showing that models underpinning outcome tests can be recast as Roy models, since heterogeneous potential outcomes enter directly into the decision maker's choice equation. Different members of the Roy model family are distinguished by the tightness of the link between potential outcomes and decisions, with important implications for defining bias and deriving logically valid outcome tests.

On the Testability of Identification in Some Nonparametric Models with Endogeneity

Ivan A. Canay, Andrés Santos, Azeem M. Shaikh

Econometrica, 81(6), pp. 2535–2559, 2013

This paper examines three hypothesis testing problems arising in the identification of some nonparametric models with endogeneity. The first concerns testing necessary conditions for identification involving mean independence restrictions (completeness conditions). The second and third concern testing identification directly in models involving quantile independence restrictions. For each problem, we provide conditions under which any test will have power no greater than size against any alternative — concluding that no nontrivial tests for these hypothesis testing problems exist.

A Simple Approach to Quantile Regression for Panel Data

Ivan A. Canay

The Econometrics Journal, 14(3), pp. 368–386, 2011

This paper provides sufficient conditions that point identify a quantile regression model with fixed effects, and proposes a simple transformation of the data that eliminates the fixed effects when they are location shifters. The resulting two-step estimator is consistent and asymptotically normal as both n and T grow. The estimator is straightforward to compute and can be implemented in standard econometrics packages.

Simultaneous Selection and Weighting of Moments in GMM Using a Trapezoidal Kernel

Ivan A. Canay

Journal of Econometrics, 156(2), pp. 284–303, 2010

This paper proposes a procedure to estimate linear models when the number of instruments is large, addressing the tradeoff between asymptotic efficiency (favoring more instruments) and bias (adversely affected by adding instruments). I propose a kernel weighted GMM estimator using a trapezoidal kernel. I derive the higher-order mean squared error and show that the trapezoidal kernel generates lower asymptotic variance than regular kernels. Monte Carlo simulations show the estimator performs on par with optimal-instrument estimators and improves upon a GMM estimator that uses all available instruments.