Featured Researches

Econometrics

On the Nuisance of Control Variables in Regression Analysis

Control variables are included in regression analyses to estimate the causal effect of a treatment on an outcome. In this note we argue that the estimated effect sizes of control variables are unlikely to have a causal interpretation themselves though. We therefore recommend to refrain from reporting marginal effects of controls in regression tables and instead to focus exclusively on the variables of interest in the results sections of empirical research papers.

Read more
Econometrics

On the Size Control of the Hybrid Test for Predictive Ability

We show that the hybrid test for superior predictability is not pointwise asymptotically of level under standard conditions, and may lead to rejection rates over 11% when the significance level α is 5% in a simple case. We propose a modified hybrid test which is uniformly asymptotically of level α by properly adapting the generalized moment selection method.

Read more
Econometrics

On the Time Trend of COVID-19: A Panel Data Study

In this paper, we study the trending behaviour of COVID-19 data at country level, and draw attention to some existing econometric tools which are potentially helpful to understand the trend better in future studies. In our empirical study, we find that European countries overall flatten the curves more effectively compared to the other regions, while Asia & Oceania also achieve some success, but the situations are not as optimistic elsewhere. Africa and America are still facing serious challenges in terms of managing the spread of the virus, and reducing the death rate, although in Africa the virus spreads slower and has a lower death rate than the other regions. By comparing the performances of different countries, our results incidentally agree with Gu et al. (2020), though different approaches and models are considered. For example, both works agree that countries such as USA, UK and Italy perform relatively poorly; on the other hand, Australia, China, Japan, Korea, and Singapore perform relatively better.

Read more
Econometrics

On the plausibility of the latent ignorability assumption

The estimation of the causal effect of an endogenous treatment based on an instrumental variable (IV) is often complicated by attrition, sample selection, or non-response in the outcome of interest. To tackle the latter problem, the latent ignorability (LI) assumption imposes that attrition/sample selection is independent of the outcome conditional on the treatment compliance type (i.e. how the treatment behaves as a function of the instrument), the instrument, and possibly further observed covariates. As a word of caution, this note formally discusses the strong behavioral implications of LI in rather standard IV models. We also provide an empirical illustration based on the Job Corps experimental study, in which the sensitivity of the estimated program effect to LI and alternative assumptions about outcome attrition is investigated.

Read more
Econometrics

On-Demand Transit User Preference Analysis using Hybrid Choice Models

In light of the increasing interest to transform the fixed-route public transit (FRT) services into on-demand transit (ODT) services, there exists a strong need for a comprehensive evaluation of the effects of this shift on the users. Such an analysis can help the municipalities and service providers to design and operate more convenient, attractive, and sustainable transit solutions. To understand the user preferences, we developed three hybrid choice models: integrated choice and latent variable (ICLV), latent class (LC), and latent class integrated choice and latent variable (LC-ICLV) models. We used these models to analyze the public transit user's preferences in Belleville, Ontario, Canada. Hybrid choice models were estimated using a rich dataset that combined the actual level of service attributes obtained from Belleville's ODT service and self-reported usage behaviour obtained from a revealed preference survey of the ODT users. The latent class models divided the users into two groups with different travel behaviour and preferences. The results showed that the captive user's preference for ODT service was significantly affected by the number of unassigned trips, in-vehicle time, and main travel mode before the ODT service started. On the other hand, the non-captive user's service preference was significantly affected by the Time Sensitivity and the Online Service Satisfaction latent variables, as well as the performance of the ODT service and trip purpose. This study attaches importance to improving the reliability and performance of the ODT service and outlines directions for reducing operational costs by updating the required fleet size and assigning more vehicles for work-related trips.

Read more
Econometrics

Optimal Decision Rules for Weak GMM

This paper studies optimal decision rules, including estimators and tests, for weakly identified GMM models. We derive the limit experiment for weakly identified GMM, and propose a theoretically-motivated class of priors which give rise to quasi-Bayes decision rules as a limiting case. Together with results in the previous literature, this establishes desirable properties for the quasi-Bayes approach regardless of model identification status. We further propose weighted average power-optimal identification-robust frequentist tests and confidence sets, and prove a Bernstein-von Mises-type result for the quasi-Bayes posterior under weak identification.

Read more
Econometrics

Optimal Policy Learning: From Theory to Practice

Following in the footsteps of the literature on empirical welfare maximization, this paper wants to contribute by stressing the policymaker perspective via a practical illustration of an optimal policy assignment problem. More specifically, by focusing on the class of threshold-based policies, we first set up the theoretical underpinnings of the policymaker selection problem, to then offer a practical solution to this problem via an empirical illustration using the popular LaLonde (1986) training program dataset. The paper proposes an implementation protocol for the optimal solution that is straightforward to apply and easy to program with standard statistical software.

Read more
Econometrics

Optimal Portfolio Using Factor Graphical Lasso

Graphical models are a powerful tool to estimate a high-dimensional inverse covariance (precision) matrix, which has been applied for a portfolio allocation problem. The assumption made by these models is a sparsity of the precision matrix. However, when stock returns are driven by common factors, such assumption does not hold. We address this limitation and develop a framework, Factor Graphical Lasso (FGL), which integrates graphical models with the factor structure in the context of portfolio allocation by decomposing a precision matrix into low-rank and sparse components. Our theoretical results and simulations show that FGL consistently estimates the portfolio weights and risk exposure and also that FGL is robust to heavy-tailed distributions which makes our method suitable for financial applications. FGL-based portfolios are shown to exhibit superior performance over several prominent competitors including equal-weighted and Index portfolios in the empirical application for the S&P500 constituents.

Read more
Econometrics

Optimal probabilistic forecasts: When do they work?

Proper scoring rules are used to assess the out-of-sample accuracy of probabilistic forecasts, with different scoring rules rewarding distinct aspects of forecast performance. Herein, we re-investigate the practice of using proper scoring rules to produce probabilistic forecasts that are `optimal' according to a given score, and assess when their out-of-sample accuracy is superior to alternative forecasts, according to that score. Particular attention is paid to relative predictive performance under misspecification of the predictive model. Using numerical illustrations, we document several novel findings within this paradigm that highlight the important interplay between the true data generating process, the assumed predictive model and the scoring rule. Notably, we show that only when a predictive model is sufficiently compatible with the true process to allow a particular score criterion to reward what it is designed to reward, will this approach to forecasting reap benefits. Subject to this compatibility however, the superiority of the optimal forecast will be greater, the greater is the degree of misspecification. We explore these issues under a range of different scenarios, and using both artificially simulated and empirical data.

Read more
Econometrics

Optimal selection of the number of control units in kNN algorithm to estimate average treatment effects

We propose a simple approach to optimally select the number of control units in k nearest neighbors (kNN) algorithm focusing in minimizing the mean squared error for the average treatment effects. Our approach is non-parametric where confidence intervals for the treatment effects were calculated using asymptotic results with bias correction. Simulation exercises show that our approach gets relative small mean squared errors, and a balance between confidence intervals length and type I error. We analyzed the average treatment effects on treated (ATET) of participation in 401(k) plans on accumulated net financial assets confirming significant effects on amount and positive probability of net asset. Our optimal k selection produces significant narrower ATET confidence intervals compared with common practice of using k=1.

Read more

Ready to get started?

Join us today