Next Article in Journal
Bifurcations and Slow-Fast Analysis in a Cardiac Cell Model for Investigation of Early Afterdepolarizations
Next Article in Special Issue
The Odd Exponentiated Half-Logistic Exponential Distribution: Estimation Methods and Application to Engineering Data
Previous Article in Journal
A New Formulation for the Capacitated Lot Sizing Problem with Batch Ordering Allowing Shortages
Previous Article in Special Issue
Determination of the Factors Affecting King Abdul Aziz University Published Articles in ISI by Multilayer Perceptron Artificial Neural Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Inference from Non-Probability Surveys with Statistical Matching and Propensity Score Adjustment Using Modern Prediction Techniques

by
Luis Castro-Martín
,
Maria del Mar Rueda
* and
Ramón Ferri-García
Department of Statistics and Operational Research, Faculty of Sciencies, University of Granada, 18071 Granada, Spain
*
Author to whom correspondence should be addressed.
Mathematics 2020, 8(6), 879; https://doi.org/10.3390/math8060879
Submission received: 7 May 2020 / Revised: 25 May 2020 / Accepted: 27 May 2020 / Published: 1 June 2020
(This article belongs to the Special Issue Statistics 2020)

Abstract

:
Online surveys are increasingly common in social and health studies, as they provide fast and inexpensive results in comparison to traditional ones. However, these surveys often work with biased samples, as the data collection is often non-probabilistic because of the lack of internet coverage in certain population groups and the self-selection procedure that many online surveys rely on. Some procedures have been proposed to mitigate the bias, such as propensity score adjustment (PSA) and statistical matching. In PSA, propensity to participate in a nonprobability survey is estimated using a probability reference survey, and then used to obtain weighted estimates. In statistical matching, the nonprobability sample is used to train models to predict the values of the target variable, and the predictions of the models for the probability sample can be used to estimate population values. In this study, both methods are compared using three datasets to simulate pseudopopulations from which nonprobability and probability samples are drawn and used to estimate population parameters. In addition, the study compares the use of linear models and Machine Learning prediction algorithms in propensity estimation in PSA and predictive modeling in Statistical Matching. The results show that statistical matching outperforms PSA in terms of bias reduction and Root Mean Square Error (RMSE), and that simpler prediction models, such as linear and k-Nearest Neighbors, provide better outcomes than bagging algorithms.

1. Introduction

Surveys are a fundamental tool for data collection in areas like social studies and health sciences. Probability sampling methods have been widely adopted by researchers in those areas, as well as by official statistics. The main reason is that it provides valid statistical inferences about large finite populations by using relatively small samples, based on a solid mathematical theory, with the right combination of a random sample design and an approximately design-unbiased estimator.
Over the last decade, new alternatives to survey sample data have became popular as data sources. Examples are big data and web surveys that have the potential of providing estimates in nearly real time, an easier data access, and lower data collection costs than traditional probability sampling [1]. Very often, the data-generating process of such sources is nonprobabilistic, given that the probability of being part of the sample is not known and/or is null for some groups of the target population, and, as a result, these methods produce nonprobability samples. There are serious issues on the use of nonprobability survey samples; the most relevant is that the data-generating process is unknown and may have serious coverage, nonresponse, and selection biases, which may not be ignorable and could deeply affect estimates [2]. These biases tend to be more disruptive as the population size gets larger, regardless of the sample size [3].
In order to correct this selection bias produced by non-random selection mechanisms, some inference procedures are proposed in the literature. A first class of methods includes statistical models aiming to predict the non-sampled units of the population [4,5,6]. Specifying an appropriate super-population model capable of learning the variation of the target variables is important for this model-based method. For this approach, auxiliary features X must be available for each unit of the observed and the unobserved parts of the population. This situation is complicated in practice.
Some studies combine a nonprobability sample with a reference probability sample for constructing models for units in the latter or to adjust selection probabilities. The most important methods for this case are statistical matching and propensity score adjustment (PSA). There are many works studying the properties and performance of PSA [7,8,9,10,11,12], but there is not much of a bibliography that develops statistical matching in this context.
In this article, we apply machine learning prediction techniques to build statistical matching estimators and compare their performance with PSA estimators. Since there is not a sampling design that allows us to determine the main statistical properties (sampling distribution, expected value, variance, etc.) of random quantities calculated from the non-probability sample, we cannot include theoretical properties of the estimators obtained, but their behavior is studied through simulation studies that also include several propensity score techniques. Although PSA performance was compared with linear calibration in [10] and the combination of PSA with machine learning was already studied in [12], to the best of our knowledge, this is the first time that these methodologies are compared in practice and the first time that machine learning techniques are used for estimation with statistical matching from nonprobability samples.
The description of the conducted study is organized as follows: In Section 2, we introduce the notation and explain the estimation problem that can be solved with the aforementioned methods. In Section 3 and Section 4, we describe the mathematical foundations of PSA and statistical matching, respectively, and their properties according to previous research. In Section 5, we briefly explain the ideas behind each of the algorithms tested in the study. In Section 6, we describe the data and the simulation study used to compare the performance of PSA and statistical matching, as well as the metrics used to measure it. Finally, in Section 7 and Section 8, we show the results of the study and discuss some of their implications in the comparison between methods.

2. Background

Suppose that the finite population U consists of i = 1 , N subjects. Let y be a survey variable and y i be the y-value attached to the i-th unit, i = 1 , , N .
Let s v be a volunteer nonprobability sample of size n v , obtained from U v U observing the study variable y.
Without any auxiliary information, the population mean Y ¯ is usually estimated with the unweighted sample mean
Y ¯ ^ = k s v y k n v
that produces biased estimates of the population mean. The size and direction of the bias depend on the proportion of the population with no chance of inclusion in the sample (coverage) and differences in the inclusion probabilities among the different members of the sample with a non-zero probability of taking part in the survey (selection) [2,13]. The selection bias cannot be estimated in practice for most survey variables of interest.
We consider the situation where there is a probability sample available and compare two inference methods to treat selection biases in a general framework. Let s r be the reference probability sample selected under the sampling design ( s d , p d ) with π i the first-order inclusion probability for the i-th individual. Let us assume that in s r , we observe some other study variables that are common to both samples, denoted by x. The available data are denoted by { ( i , y i , x i ) , i s v } , and { ( i , x i ) , i s r } . We are interested in estimating a linear parameter θ N = U a i y i , where a i are known constants. Examples include the population total T y = U y i , the population mean Y ¯ , and the population proportion p A = U y i / N , where y i = 1 if the unit i belongs to the interest group A, and 0 otherwise.

3. Propensity Score Adjustment

The most popular adjustment method in nonprobability settings is propensity score adjustment (PSA) or propensity weighting. This method, firstly developed by [14], was originally intended to correct the confounding bias in the experimental design context, and it is the most widely used method in practice [2,7,8,9,10,12,15,16,17]. In this approach, the propensity for an individual to participate in the volunteer survey is estimated by binning the data from both samples, s r and s v , together and training a machine learning model (usually logistic regression) on the variable δ , with δ k = 1 if k s v and δ k = 0 if k s r . We assume that the selection mechanism of s v is ignorable; this is:
P ( δ k = i | y k , x k ) = P ( δ k = i | x k ) , i = 0 , 1 ; k s v .
We also assume that the mechanism follows a parametric model:
P ( δ k = 1 | y k , x k ) = p k ( x ) = 1 e ( γ T x k ) + 1
for some vector γ . We obtain the pseudo maximum likelihood of parameter γ and use the inverse of the estimated response propensity as weight for constructing the estimator [11]:
θ ^ P S A 1 = k s v a k y k / p ^ k ( x k ) ,
where p ^ k ( x k ) denotes the estimated response propensity for the individual k s v . Alternative estimators can be constructed by slightly modifying the formula in (4) [18]:
θ ^ P S A 2 = k s v 1 p ^ k ( x k ) a k y k / p ^ k ( x k ) .
Other alternatives involve the stratification of propensities in a fixed number of groups, with the idea of grouping individuals with similar volunteering propensities. For instance, in [7,8], adjustment factors f c are obtained for the cth strata of individuals:
f c = k s r c d k r / k s r d k r j s v c d j v / j s v d j v ,
where s r c and s v c are individuals from the s r and s v sample respectively who belong to the cth group, while d k r = 1 / π k and d j v = 1 / p ^ j are the design weights for the kth individual of the reference sample and the jth individual of the volunteer sample, respectively. The final weights are:
w j = f c d j v = k s r c d k r / k s r d k r j s v c d j v / j s v d j v d j v .
The weights are then used in the Horvitz–Thompson estimator:
θ ^ P S A 3 = k s v w k a k y k .
The approach used in [9] does not require the calculation of f c . It only uses the average propensity within each group
π ¯ c = k s r c s v c p k ( x ) / ( n r c + n v c ) ,
where n r c and n v c are the number of individuals from the reference and the volunteer sample, respectively, that belong to the cth group. The mean propensity for each member of the volunteer sample is used in the Horvitz–Thompson estimator:
θ ^ P S A 4 = c k s v c a k y k / π ¯ c .
PSA in nonprobability online surveys has been proven to be efficient if the selection mechanism is ignorable and the right covariates are used for modeling [7]. If some of these conditions do not apply, the use of PSA can induce biased estimates that would need further adjustments [9]. The combination of PSA and calibration has shown successful results in terms of bias removal [8,10].
Several machine learning models have been suggested as alternatives to logistic regression for the estimation of propensity scores in the experimental design context, with promising results. Ref. [19,20] examined the performance of various classification and regression trees (CART) for PSA in sample balancing. Other applications of machine learning algorithms in PSA involve their use in nonresponse adjustments; more precisely, they have been studied using Random Forests as propensity predictors [21]. Regarding the nonprobability sampling context covered in this study, [12] presented a simulation study using decision trees, k-Nearest Neighbors, Naive Bayes, Random Forests, and a Gradient Boosting Machine that support the view given in [6] about machine learning methods being used for removing selection bias in nonprobability samples. All of those algorithms, along with Discriminant Analysis and Model Averaged Neural Networks, will be used for propensity estimation in this study. Further details can be consulted in Section 5.

4. Statistical Matching

Statistical matching (also known as data fusion, data merging, or synthetic matching) is a model-based approach introduced by [22] and further developed by [23] for nonresponse in probability samples. The idea in this context is to model the relationship between y k and x k using the volunteer sample s V in order to predict y k for the reference sample.
Suppose that the finite population { ( i , y i , x i ) , i U } can be viewed as a random sample from the superpopulation model:
y i = m ( x i ) + e i , i = 1 , 2 , . . . , N ,
where m ( x i ) = E m ( y i | x i ) and the random vector e = ( e 1 , . . . , e N ) is assumed to have zero mean.
Under the design-based approach, the usual estimator of a population’s linear parameter is the Horvitz–Thompson estimator given by:
θ ^ H T = k s r a k y k d k
where d k = 1 / π k is the sampling weight of the unit k that is design-unbiased, consistent for θ , and asymptotically normally distributed under mild conditions [24]. This estimator cannot be calculated because y k is not observed for the units k s r ; thus, we substitute y i by the predicted values from the above model. Thus, the matching estimator is given by:
θ ^ S M = s r a k y ^ k d k ,
where y ^ k is the predicted value of y k .
The key is how to predict the values of y k . Usually, the linear regression model is considered for estimation; E m ( y i | x i ) = x i T β is easy to implement in most of the existent statistical packages, but several drawbacks have to be considered. Parametric models require assumptions regarding variable selection, the functional form and distributions of variables, and specification of interactions. If any of these assumptions are incorrect, the bias reduction could be incomplete or nonexistent. Contrary to statistical modeling approaches that assume a data model with parameters estimated from the data, more advanced machine learning algorithms aim to extract the relationship between an outcome and predictor without an a priori data model. These methods have not been widely applied in the statistical matching literature. Now, we propose the use of machine learning methods as an alternative to linear regression modeling. The ML prediction methods considered in this article are described in the following section.

5. Prediction Modeling

5.1. Generalized Linear Models (GLM)

The most basic regression model consists of calculating coefficients, β , of linear regression based on input data. The coefficients that satisfy the optimality criteria based on minimizing the Ordinary Least Squares are estimated with the formula β = ( X X ) 1 X Y . This method is only stable as long as X X is relatively close to a unit matrix [25]. Quite often, covariates suffer from multicollinearity. For those cases, ridge regression proposes an identity term to control instability: β = ( X X + k I ) 1 X Y , where k 0 can be chosen arbitrarily or via parameter tuning. β can also be considered as the posterior mean of a prior normal distribution with zero mean and a variance of I σ 2 / k [26]. From that point of view, Bayesian estimates can be obtained via Gibbs sampling.
Instead, the Least Absolute Shrinkage and Selection Operator (LASSO) regression [27] proposes using a penalty parameter, α , according to the following optimization problem:
a r g m i n i = 1 N ( y i α j β j x i j ) 2 s u b j e c t t o j | β j | t .
t is a hyperparameter that forces the shrinkage of the coefficients. In this case, coefficients are allowed to be equal to zero. Therefore, the main difference is that LASSO allows the optimization procedure to select variables, while ridge regression may produce very small coefficients for some cases without reaching zero. Alternatively, LASSO coefficients can be estimated considering the posterior mode of prior Laplace distributions. Bayesian estimates can then be calculated as described in [28]. Ridge and LASSO are both considered standard penalized regression models [29].
For PSA, these methods can be used for estimating the propensities. First, the target variable for the model is defined as y i = 1 if k s v and y i = 0 if k s r . The pseudo maximum likelihood can then be optimized via logistic regression or any of its variants described above.
For statistical matching, the target variable for the model is the survey variable itself. Therefore, the model is trained with the volunteer sample and then used to obtain the estimated responses for the reference sample.

5.2. Discriminant Analysis

When the predicted variable is discrete, Discriminant Analysis can be used for classification of individuals. Let y be the dependent variable with K classes, π k the probability of an individual of belonging to the kth class, X the matrix of covariates n × p , and f k ( x ) the joint distribution of x conditioned to y taking the kth class. As described in [30], Linear Discriminant Analysis (LDA) assigns an individual the class that maximizes the probability:
P ( y i = k | x = x i ) = π k f k ( x i ) j = 1 K π j f j ( x i ) , k = 1 , , K .
Assuming that X | y = k follows a multivariate Gaussian distribution N p ( μ k , Σ ) , LDA works by assigning an instance the class for which the coefficient δ k ( x i ) defined as
δ k ( x i ) = x i T Σ 1 μ k 1 2 μ k T Σ 1 μ k + l o g ( π k )
is largest. Note that the decision depends on a linear combination of the multivariate Gaussian distribution parameters; hence, the classifier gets the name of Linear Discriminant Analysis. When used for PSA, K = 2 and, as a result, the outcome of LDA is the posterior probability obtained in (15) for the class δ = 1 .
LDA can provide good results; however, its simplicity can be a handicap in some cases where relationships between covariates and target are complex, and if the covariates are correlated, its performance gets worse [31]. For these reasons, alternatives considering smoothing, such as Penalized Discriminant Analysis (FDA) or Shrinkage Discriminant Analysis (SDA), can be used. The former expands the covariate matrix and applies penalization coefficients in the calculation of thresholds [32], while the latter performs a shrinkage of covariates similar to that performed in the ridge or LASSO models.
LDA is only suitable for classification and, therefore, it cannot be used for statistical matching when the survey variable is continuous. However, its probabilistic nature makes it appropriate for estimating propensities in PSA, as described above.

5.3. Decision Trees, Bagged Trees, and Random Forests

Decision trees sequentially split the input data via conditional clauses until they reach a terminal node, which assigns a specific class or value. This process results in the following estimation for the expectance E m ( y i | x i ) :
E m ( y i | x i ) = y ( s J 1 ) ¯ { i s / x i J 1 } y ( s J k ) ¯ { i s / x i J k } ,
where y ( s J i ) ¯ denotes the mean of y among the members of the sampled population, s, meeting the criteria of the i-th terminal node.
Bagged trees combine this approach with bagging [33]. Bagging averages the predictions of multiple weak classifiers (in this case, m unpruned trees). In order for them to complement each other, they are trained with a bootstrapped subsample of the complete dataset. Therefore:
E m ( y i | x i ) = j = 1 m ϕ j ( x i ) m , ϕ j ( x i ) = y ( s J 1 j ) ¯ { i s / x i J 1 j } y ( s J k j ) ¯ { i s / x i J k j } ,
where y ( s J i j ) ¯ denotes the mean of y among the members of the sampled population, s, meeting the criteria of the i-th terminal node of the j-th tree. This technique is known to improve the accuracy of the predictions [34]. Alternatively, Random Forests can also be used for both regression and classification using weak classifiers [35]. In this algorithm, the input variables for each weak classifier are a random subset of all of the covariates, instead of taking the whole x i vector as in bagged trees.
This approach is easy to apply for statistical matching. As usual, a model is trained using the volunteer sample in order to predict a response based on the covariates. Said model is then applied to the reference sample covariates. However, tree-based models are not good for estimating probabilities [36]. They can still be used for PSA, taking the proportion of weak classifiers that agree as the estimated propensity.

5.4. Gradient Boosting Machine

Gradient Boosting Machine (GBM) also works as an ensemble of weak classifiers. Boosting is an iterative process that trains subsequent models, giving more importance to the data for which previous models failed. This idea can be interpreted as an optimization problem [37], and, therefore, it is suitable for the gradient descent algorithm [38]. Then, the estimates for y are:
E m ( y i | x i ) = v T J ( x i ) ,
where J ( x i ) stands for a matrix of terminal nodes of m decision trees and v is a vector representing the weight of each tree. GBM has improved previous state-of-the-art models for some cases [39].
GBM can be used for PSA and statistical matching in the same way as the previous ensemble models considered.

5.5. k-Nearest Neighbors

k-Nearest Neighbors is “one of the most fundamental and simple classification methods” [40]. It does not need training. The algorithm simply averages the value of the target variable for the k individuals closest to the estimated individual (its k nearest neighbors), given a certain distance dependent on the covariates. This is:
E m ( y i | x i ) = j s / d ( x i , x j ) d ( x i , x ( k ) ) y j k
where x ( 1 ) ,..., x ( n 1 ) are, respectively, the individuals closest to and furthest from x i . Choosing the right k is important for the proper performance of the algorithm.
For classification, k-Nearest Neighbors would usually simply predict the most repeated label among its k-nearest neighbors. However, it can instead take into account the proportion in order to estimate probabilities. This idea can be applied for PSA, taking y i = 1 if k s v and y i = 0 if k s r , as always. For statistical matching, k-Nearest Neighbors can also be used normally to predict the responses.

5.6. Naive Bayes

The Naive Bayes algorithm is a classifier (that is, it can only be used to predict discrete variables) based on the Bayes theorem. In this study, Naive Bayes has been used only for propensity estimation in PSA. In this case, the Bayes theorem can be used with the probabilities that the participants have of being part of the volunteer sample and the occurrence of a given vector for X , that is, the values of the covariates for a given individual i.
p ^ i ( x i ) = P ( δ i = 1 ) P ( X = x i | δ i = 1 ) P ( X = x i ) .
The Naive Bayes classifier is simple in its reasoning, but can provide precise results in PSA under certain conditions [12]. On the other hand, predictions from Naive Bayes can turn unstable when covariates with high cardinality (e.g., numerical variables) are present, as discrete domains are required for computation of probabilities [41].
As was the case with discriminant analysis, Naive Bayes works naturally with probabilities; therefore, it is suitable for estimating propensities in PSA, but not for statistical matching if the survey variable is continuous.

5.7. Neural Networks with Bayesian Regularization

Neural networks calculate the expectance of y i as:
E m ( y i | x i ) = g k = 1 L v k f k ( · ) + b ,
where g and f k stand for the activation functions, v k are the weights of the k-th neuron, and b is the activation threshold [42]. The inputs follow an iterative process through one or more hidden layers until reaching the last layer, which produces the final output. The weights are initialized randomly and then optimized via gradient descent with the backpropagation algorithm [43].
Overfitting is an important problem for neural networks so prior distributions can be imposed to v k weights as a regularization method. They are then optimized to maximize the posterior density or the likelihood, as described in [42]. Another option is bagging of neural networks, as explained in [44]. The same neural network model is fitted using different seeds, and the results are averaged to obtain the predictions. This approach is known as Model Averaged Neural Networks.
Neural networks have already been considered for superpopulation modeling [45]. They are the state of the art for many domains [46]; Bayesian neural networks in particular are “fairly robust with respect to the problems of overfitting and hyperparameter choice” [47].
Since they work as universal approximators [48], neural networks can be used for PSA and statistical matching in the same way as generalized linear models.

6. The Simulation Study

6.1. Data

All of the experiments were performed using three different populations. In addition, two different sampling strategies were selected for each one in order to recreate the behavior of the estimates under the lack of representativeness of the potentially sampled subpopulation and under selection mechanisms tied to individual features (e.g., voluntariness).
The first population, which will be referred to as P1, corresponds to the microdata of the Spanish Life Conditions Survey (2012 edition) [49]. It collects data about economic and life conditions variables for 28,610 adults living in Spain. We took the mean health, as reported by the individuals themselves on a scale from 1 to 5, as the objective variable to estimate. The algorithms were trained using the 56 most related variables, excluding “health issues in the last six months”, “chronic conditions”, “household income difficulties”, and “civil status” (as they are too correlated with the target variable). The first sampling strategy for this population, which will be referred to as P1S1, was a simple random sampling excluding the individuals without internet access. In the second sampling strategy, P2S2, we also included a propensity to participate in the sample using the formula P r ( y r ) = y r 2 1900 2 1996 2 1900 2 , where y r is the year in which the individual was born. This way, linear models should have more problems learning the relations.
BigLucy [50], P2, was chosen as the second population. It consists of various financial variables of 85,396 industrial companies of a city for a particular fiscal year. The target variable chosen was the annual income in the previous fiscal year. The algorithms were trained using the size of the company (small, medium, or big), the number of employees, the company’s income tax, and whether it is ISO certified. The first sampling strategy for this population, P2S1, was simple random sampling among the companies with SPAM options that are not small companies. This approach tested whether the models were able to correctly estimate the annual income for companies that were not in the training data. The second sampling strategy, P2S2, was simple random sampling among the companies with SPAM options, including a propensity to participate calculated as P r ( t a x e s ) = m i n ( t a x e s 2 / 30 , 1 ) , where t a x e s is the company’s income tax. This scenario is similar but it implies a quadratic dependence.
The Bank Marketing Data Set [51], P3, is the third population. It includes information about 41,188 phone calls related to direct marketing campaigns of a Portuguese banking institution. Our goal is to predict the mean contact duration. A total of 18 variables were used for training. We excluded two of the dataset variables, the month, and whether the client has subscribed for a term deposit in order to make the inference more difficult. For the first sampling strategy, P3S1, we applied simple random sampling among the clients contacted more than three times. For the second sampling strategy, P3S2, we applied simple random sampling among the clients contacted more than twice. Surprisingly, filtering less led to worse estimations for some cases.

6.2. Simulation

Each population and sampling strategy was simulated using various sample sizes: 1000, 2000, and 5000. The same size is taken for the convenience sample and for the reference sample. For each sample size, 500 simulations were executed. In each simulation, PSA (using weights defined in Formula (4) in Section 3), PSA with stratification (using weights defined in Formula (10) in Section 3), and statistical matching estimates were obtained using several predictive algorithms.
For PSA (with and without stratification), the following classification algorithms were used: Logistic regression (glm), generalized linear model via penalized maximum likelihood (glmnet), Naive Bayes (naivebayes), k-Nearest Neighbors (knn), C4.5 decision tree (J48), Bagged Trees (treebag), Random Forests (rf), Gradient Boosting Machine (gbm), Model Averaged Neural Network (avNNet), Linear Discriminant Analysis (lda), Penalized Discriminant Analysis (pda), and Shrinkage Discriminant Analysis (sda).
For statistical matching, the following regression algorithms were used: linear regression (glm), Ridge regression with and without Bayesian priors (bridge and ridge respectively), LASSO regression via penalized maximum likelihood (glmnet), LARS-EN algorithm (lasso) and using Bayesian priors on the estimates (blasso), k-Nearest Neighbors (knn), Bagged Trees (treebag), Gradient Boosting Machine (gbm) and Bayesian-regularized Neural Networks (brnn).
These represent standard variants from different model types: Linear regression, penalized regression, Bayesian models, prototype models, trees, gradient boosting, neural networks, and discriminant analysis. All of the methods were trained using default hyperparameters, except for k-Nearest Neighbors, Naive Bayes, and C4.5, because their performance improved greatly after hypeparameter tuning. Said tuning was performed with bootstrap. The framework used for training, optimization, and prediction was caret [52], an R [53] package.
Different metrics are considered for evaluating each scenario: Relative mean bias, relative standard deviation, and relative Root Mean Square Error (RMSE).
R B i a s ( % ) = i = 1 500 p ^ y i 500 p y × 100 p y
R S t a n d a r d d e v i a t i o n ( % ) = i = 1 500 ( p ^ y i p ¯ ^ y ) 2 499 × 100 p y
R M S E ( % ) = R B i a s 2 + R S D 2 ,
where p y is the value of the target variable, p ¯ ^ y is the mean of the 500 estimations of p y , and p ^ y i is the estimation of p y in the i-th simulation.
In order to rank each estimator, the mean efficiency, the median efficiency, and the number of times it has been among the best are measured. An estimator is considered to be among the best when its RMSE differs from the minimum RMSE by less than 1%. The efficiency is defined as follows:
E f f i c i e n c y ( % ) = B a s e l i n e R M S E B a s e l i n e × 100 ,
where the baseline is the RMSE of using the sample average as the estimation.
To complete the comparison analysis, the results of relative bias, standard deviation, and RMSE were analyzed using linear mixed-effects regression. This approach provides estimates of the effect size of each adjustment method and algorithm. Datasets were considered random effects, while adjustment (matching, PSA, or PSA with propensity stratification) and algorithm (glm, gbm, glmnet, knn, and treebag, as they were the only algorithms used in all adjustments) were the fixed effects variables. All three response variables take non-negative values, and the interpretation is the same: The lower their value is, the better (less biased and/or less variable) the estimations are. Following this rule, negative Beta coefficients indicate that a given factor is contributing to better estimations, and vice versa for positive coefficients.

7. Results

Table A1, Table A2 and Table A3 in the Appendix A show, respectively, the resulting biases, deviations, and RMSEs. In general, it can be observed that statistical matching outperforms PSA, which outperforms PSA with stratification. Nevertheless, all three methods consistently reduce the sample bias.
In terms of machine learning algorithms, basic linear models seem the most robust. Others, like Naive Bayes, can achieve outstanding results, but only for some cases. It is also interesting noting that C4.5 trees can even lead to worse estimations than simply using the sample average.
The final ranking confirming this impression can be seen in Table 1. Statistical matching with linear or ridge regression has the best mean efficiency and has been among the best more than the rest of approaches. This is not a surprise, since their simplicity avoids overfitting, presumably one of the main problems of matching. Ridge regression should be preferred if the data suffer from multicollinearity. Otherwise, linear regression alone is very effective (and faster).
The results of linear mixed-effects modeling can be consulted in Table 2, Table 3 and Table 4. It is noticeable how linear models, LASSO with LARS-EN algorithm, and k-Nearest Neighbors outperform Bagged Trees in all metrics (modulus of relative bias, relative standard deviation, and RMSE), while there is no evidence that GBM is different from any of them. Regarding adjustment methods, PSA (both with and without propensity stratification) showed significantly more bias and RMSE than statistical matching. In the case of standard deviation, there is also evidence that PSA without propensity stratification provides higher values (deviation) than matching. Altogether, these results would indicate that matching has a larger effect on bias reduction than PSA.

8. Discussion

Nonprobability samples are increasingly common due to the growing internet penetration and the subsequent rise of online questionnaires. These questionnaries are a faster, less expensive, and more comfortable method of information collection in comparison to traditional ones. However, samples obtained with this technique deal with several sources of bias: Despite the increasing internet penetration, large population groups (less educated or elderly people) are still not properly represented. In addition, questionnaires are often administered with non-probabilistic sampling methods (e.g., snowballing), which imply that the selection is controlled by the interviewees themselves, causing a selection bias.
In this study, we focus on two of the proposed methods to reduce biases produced by nonprobability sampling: PSA and matching. We also compare the outcomes when the predictive modeling, required in both methods, is done through linear regression and through machine learning algorithms. PSA and matching require a probability sample on which the target variable has not been measured. The unit sampling performed in the simulations captures different self-selection scenarios in nonprobability sampling, while probability samples are drawn by simple random sampling with no sources of bias. This canonical representation is not usual, as reference samples are mostly drawn with complex sampling methods and the amount of bias is non-null. Further research could take into account these imperfect situations.
Results show that statistical matching provides better results than PSA on bias reduction and RMSE, regardless of the dataset and selection mechanism. In addition, linear models and k-nearest neighbors provided, on average, better results in terms of bias reduction than more complex models, such as GBM and Bagged Trees. These results are relevant since, even though there are comparative studies between adjustment techniques in nonprobability surveys [11,54], to the best of our knowledge, no comparison has been done before between these two methods.
Before closing, several limitations of our analysis should be mentioned. Given that the datasets used for simulation are real-life examples, we cannot ensure whether a selection bias mechanism is Missing At Random (MAR) or Missing Not At Random (MNAR), as the causality relationships are not known. It is known that the selection mechanism makes a difference in terms of how challenging bias reduction can be, but, in this study, it was not possible to assess.
In the near future, it is planned to explore how to combine PSA and statistical matching techniques. Shrinkage is a natural way to improve the available estimates in terms of the mean squared error that has been used by many authors in other contexts (e.g., [55,56,57]). The idea is to shrink the estimator θ ^ S M towards the estimator θ ^ P S A and obtain θ ^ s r k = K θ ^ S M + ( 1 K ) θ ^ P S A , where K is a constant satisfying 0 < K < 1 .
Another way can be considered by taking into account that most machine learning models allow weighting of the data used for training. Therefore, the weights obtained via PSA for the volunteer sample can be used when training the model used for statistical matching, since it is trained with said sample.

Author Contributions

Conceptualization, M.d.M.R.; formal analysis, M.d.M.R.; funding acquisition, M.d.M.R.; methodology, R.F.-G.; software, L.C.-M.; supervision, R.F.-G.; validation, L.C.-M.; writing, L.C.-M. All authors have read and agreed to the published version of the manuscript.

Funding

The work was supported by the Ministerio de Economia, Industria y Competitividad, Spain, under Grant MTM2015-63609-R and, in terms of the third author, an FPU grant from the Ministerio de Ciencia, Innovación y Universidades, Spain (FPU17/02177).

Acknowledgments

We would like to thank the anonymous referees for their remarks and comments that have improved the presentation and the contents of the paper.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Table A1. Relative mean bias (%) of each estimator for each population, sampling method, and sample size. The best values among the methods are shown in bold.
Table A1. Relative mean bias (%) of each estimator for each population, sampling method, and sample size. The best values among the methods are shown in bold.
P1S1 P1S2 P2S1 P2S2 P3S1 P3S2
Estimator100020005000 100020005000 100020005000 100020005000 100020005000 100020005000
baseline8.48.58.5 12.912.812.8 70.670.470.4 32.732.632.9 13.413.213.3 5.95.85.9
matching blasso3.332.7 6.65.85.2 24.624.624.6 12.612.612.6 8.95.12.7 30.42
matching bridge3.43.12.8 6.96.15.3 24.524.624.5 12.312.612.7 9.14.82.3 40.51.5
matching brnn2.42.32.2 4.84.54.3 25.324.724.7 13.313.313.3 4.32.50 0.42.14
matching gbm5.35.35.4 8.88.98.9 46.246.747.1 1717.417.5 0.42.35 5.85.33.2
matching glm2.62.42.5 4.74.74.6 24.424.624.7 12.612.712.7 0.60.50.8 3.43.33
matching glmnet2.62.62.5 4.84.74.8 25.525.625.6 12.712.912.8 0.70.80.9 3.23.23.2
matching knn54.43.6 7.56.96.1 34.234.134 75.84.2 6.15.85.5 0.90.50.3
matching lasso7.17.27.2 10.81111.1 65.565.565.5 29.929.829.7 6.86.46.6 1.81.31.2
matching ridge2.42.52.5 4.74.74.7 24.524.624.7 12.612.812.7 0.50.90.8 3.23.13.2
matching treebag3.53.74.3 6.56.16.4 45.545.746 16.416.616.6 6.60.18.8 10.861.7
psa avNNet3.94.23.9 6.36.76.6 66.864.262.6 8.66.910 10.58.910 5.14.54.6
psa gbm5.95.75.4 9.49.29 66.166.667.2 15.115.315.6 10.911.111.5 1.31.31.4
psa glm3.43.53.5 5.65.85.8 67.76868.1 15.115.115.1 4.85.25.1 1.31.31.3
psa glmnet3.73.63.6 5.965.9 66.666.766.9 15.215.115.1 5.45.65.4 1.21.11.1
psa J484.655.2 9.510.510.9 70.670.469.6 23.522.115.1 19.322.323.4 4.22.22
psa knn4.34.24.1 7.37.17 68.468.566.1 18.818.213.2 77.57.9 0.60.60.6
psa lda3.73.63.6 6.16.26.3 67.367.267.1 14.814.914.8 6.56.26.4 0.20.30.2
psa naive2.210.1 4.73.62.5 22.424.829.8 4.56.47.1 3.44.14.3 4.34.45.3
psa pda3.63.63.6 6.26.36.2 67.767.267 14.814.914.8 6.86.46.3 0.50.20.3
psa rf77.17.4 1111.111.4 117.2125.3134.9 26.430.730.9 11.611.912.4 4.14.85.1
psa sda3.53.73.6 6.26.26.2 67.16767 14.814.714.8 6.36.26.3 0.50.30.3
psa treebag6.877.3 10.91111.4 66.26659.6 13.323.425.6 12.112.412.7 4.44.75.6
psa strat avNNet6.46.56.5 9.49.59.4 68.168.567.6 17.916.418.3 12.81312.8 3.53.23.5
psa strat gbm7.27.17 1110.710.6 66.365.264.1 22.822.923 12.212.212 3.94.24.2
psa strat glm6.26.16.1 98.88.8 71.369.666.9 2322.923 10.810.710.8 2.933
psa strat glmnet6.36.26.1 9.18.98.9 77.378.781 2322.923 10.910.910.9 3.133
psa strat J486.26.96.9 10.511.511.8 70.670.470 27.426.922.5 14.817.718.8 0.50.21.4
psa strat knn6.76.66.6 9.19.19.1 69.569.767.6 16.915.511.2 11.411.111.1 3.43.63.5
psa strat lda6.26.26.2 9.19.19 71.469.967.2 22.722.722.8 10.510.510.5 3.43.53.6
psa strat naive5.65.55.1 6.45.85 79.879.277.4 3.43.42.9 10.410.410.4 2.51.90.3
psa strat pda6.26.16.1 9.29.19.1 727067.5 22.822.722.8 12.612.812.9 3.63.53.6
psa strat rf7.47.58.1 12.512.412.1 8283.385 21.426.830.2 10.510.610.5 5.55.55.6
psa strat sda6.36.26.1 9.39.19.1 70.87066.7 22.822.722.8 12.913.213.1 3.43.63.6
psa strat treebag7.97.98.1 12.412.412.4 68.468.767.1 6.920.524.3 9.88.911.2 5.95.65.7
Table A2. Relative deviation (%) of each estimator for each population, sampling method, and sample size. The best values among the methods are in bold.
Table A2. Relative deviation (%) of each estimator for each population, sampling method, and sample size. The best values among the methods are in bold.
P1S1 P1S2 P2S1 P2S2 P3S1 P3S2
Estimator100020005000 100020005000 100020005000 100020005000 100020005000 100020005000
baseline1.10.80.5 1.30.90.5 1.71.20.7 2.41.71 3.11.90.8 32.11.2
matching blasso1.51.10.6 1.81.30.8 1.51.20.7 1.71.20.7 3.93.11.5 3.82.71.6
matching bridge1.61.10.7 21.41.1 1.72.30.7 1.81.20.7 5.13.51.6 4.131.6
matching brnn1.610.6 1.91.40.8 2.11.10.7 1.91.20.8 6.25.12.3 5.341.6
matching gbm1.410.6 1.71.20.7 1.30.90.6 1.71.20.7 7.64.31.6 6.65.13.6
matching glm1.410.6 21.40.9 1.61.10.6 1.71.20.7 4.12.71.2 42.61.4
matching glmnet1.510.6 21.30.8 1.61.10.6 1.61.20.7 3.92.81.1 42.51.4
matching knn1.610.7 21.40.9 1.30.80.6 1.71.30.9 4.32.81.3 3.92.81.5
matching lasso1.20.80.5 1.410.6 1.71.20.7 21.40.9 3.92.51.1 3.52.41.4
matching ridge1.610.6 1.91.30.8 1.51.10.7 1.81.20.7 42.71.2 3.92.51.4
matching treebag1.41.10.7 1.91.40.9 1.410.7 1.61.20.8 116.41.5 95.62
psa avNNet1.61.10.9 21.41.1 8.814.616 6.83.25.3 563.9 3.432
psa gbm1.30.90.5 1.610.7 2.71.81.1 1.81.30.7 3.92.61.1 4.12.71.4
psa glm1.510.6 21.40.8 321.3 1.71.20.8 4.12.71.2 42.61.3
psa glmnet1.510.6 21.30.8 2.71.91.3 1.71.20.8 42.61.1 3.72.61.4
psa J482.61.81.2 2.92.11.2 1.71.21.6 8.78.56.5 21.919.918.6 25.217.911.5
psa knn1.91.20.7 2.11.60.9 3.42.61.4 3.42.51.6 4.93.31.5 4.53.21.9
psa lda1.510.6 1.91.20.8 3.62.61.6 1.91.20.7 3.52.41 3.62.41.3
psa naive4.32.31.2 3.42.41.4 8.48.45.6 17.24.80.8 13.183.8 10.76.45.9
psa pda1.510.5 1.81.20.7 3.92.81.6 1.71.20.7 3.62.31 3.62.61.3
psa rf1.310.6 1.51.10.7 11.79.78.7 5.23.63.2 5.23.62.4 4.53.32.1
psa sda1.410.6 1.81.20.7 3.92.71.6 1.71.20.7 3.52.41 3.62.51.3
psa treebag1.61.21.3 1.81.20.8 9.713.98.8 21.18.74.2 5.74.32.8 5.342.4
psa strat avNNet1.20.80.5 1.61.10.7 5.45.57.2 3.61.93.7 3.22.31.1 42.71.6
psa strat gbm1.20.80.5 1.410.6 4.94.43.6 1.51.10.8 3.52.20.9 3.52.21.2
psa strat glm1.20.80.5 1.510.6 7.86.74.5 1.61.10.8 3.22.10.9 3.52.11.2
psa strat glmnet1.20.80.5 1.51.10.6 6.44.92.5 1.71.10.7 3.21.90.9 3.32.21.2
psa strat J482.21.10.7 2.41.50.8 2.11.44.2 5.44.94 15.412.310.5 20.114.610
psa strat knn1.20.90.5 1.71.20.7 2.71.91.4 53.72 3.22.31.2 3.42.51.3
psa strat lda1.20.80.5 1.510.6 7.16.64.8 1.71.20.8 3.220.9 3.32.31.3
psa strat naive1.61.10.6 2.71.70.9 4.13.62.7 2.821.3 3.120.9 64.84
psa strat pda1.20.80.5 1.410.6 7.66.84.9 1.71.30.8 2.91.90.9 3.22.21.2
psa strat rf1.20.80.4 1.40.90.6 3.22.51.7 4.23.32.9 3.11.90.9 3.32.11.2
psa strat sda1.20.80.5 1.51.10.6 7.57.14.4 1.81.30.8 321 3.32.21.2
psa strat treebag1.20.80.5 1.40.90.6 5.67.74.5 20.88.84 10.510.43.7 3.12.21.3
Table A3. Relative Root Mean Square Error (RMSE) (%) of each estimator for each population, sampling method, and sample size. The best values among the methods are shown in bold.
Table A3. Relative Root Mean Square Error (RMSE) (%) of each estimator for each population, sampling method, and sample size. The best values among the methods are shown in bold.
P1S1 P1S2 P2S1 P2S2 P3S1 P3S2
Estimator100020005000 100020005000 100020005000 100020005000 100020005000 100020005000
baseline8.58.58.5 12.912.912.8 70.670.470.5 32.832.732.9 13.813.413.3 6.66.26
matching blasso3.63.22.8 6.965.3 24.724.624.6 12.712.712.6 9.763 4.82.72.6
matching bridge3.83.32.8 7.16.35.4 24.624.724.5 12.512.612.7 10.45.92.8 5.732.2
matching brnn2.92.52.5 5.24.74.4 25.424.724.7 13.413.413.3 7.65.62.3 5.34.54.3
matching gbm5.55.45.4 998.9 46.246.747.1 17.117.417.5 7.64.95.2 8.87.34.8
matching glm32.62.5 5.14.94.7 24.424.624.7 12.712.712.7 4.22.81.4 5.24.23.3
matching glmnet32.82.6 5.24.94.9 25.525.625.6 12.812.912.8 42.91.4 5.243.4
matching knn5.34.63.6 7.876.2 34.234.134 7.264.3 7.56.55.7 42.91.5
matching lasso7.27.37.3 10.911.111.1 65.665.565.5 3029.829.7 7.86.86.7 3.92.71.8
matching ridge2.92.72.6 5.14.94.8 24.524.624.7 12.712.912.7 42.91.5 5.143.5
matching treebag3.83.84.3 6.76.26.4 45.545.746 16.516.616.7 12.86.49 148.22.7
psa avNNet4.24.34 6.66.86.7 67.465.864.6 117.611.3 11.710.810.7 6.15.45
psa gbm6.15.85.5 9.59.39 66.166.667.2 15.215.315.6 11.611.411.5 4.332
psa glm3.73.63.5 665.9 67.868.168.1 15.215.115.2 6.45.85.2 4.231.9
psa glmnet43.83.7 6.26.16 66.766.866.9 15.315.115.1 6.76.25.6 3.92.81.8
psa J485.35.35.4 9.910.710.9 70.670.469.6 2523.716.4 29.229.929.9 25.518.111.7
psa knn4.74.44.2 7.67.37.1 68.568.566.1 19.118.413.3 8.58.28 4.53.32
psa lda43.73.6 6.46.36.3 67.467.267.2 14.914.914.8 7.46.76.4 3.62.41.3
psa naive4.82.61.2 5.84.32.9 23.926.230.3 17.887.1 13.595.7 11.67.88
psa pda3.93.73.6 6.56.56.3 67.867.267 14.91514.9 7.76.86.3 3.72.61.3
psa rf7.17.17.4 11.111.211.4 117.8125.7135.2 26.930.931.1 12.712.412.7 6.15.85.5
psa sda3.83.83.7 6.56.36.3 67.267.167 14.914.814.8 7.36.66.4 3.62.51.3
psa treebag77.17.4 11.111.111.4 66.967.460.2 2524.926 13.313.113 6.96.16.1
psa strat avNNet6.56.66.5 9.59.69.4 68.368.768 18.316.518.6 13.213.212.8 5.34.23.8
psa strat gbm7.37.27 11.110.810.7 66.565.364.2 22.922.923 12.712.412 5.24.84.4
psa strat glm6.36.26.2 9.18.88.8 71.769.967.1 23.122.923 11.210.910.8 4.53.73.2
psa strat glmnet6.46.26.2 9.299 77.578.881 23.122.923 11.311.110.9 4.53.83.3
psa strat J486.677 10.711.611.9 70.670.570.2 27.927.322.9 21.421.521.6 20.114.610.1
psa strat knn6.86.76.6 9.39.29.1 69.569.767.6 17.615.911.3 11.811.311.2 4.84.33.7
psa strat lda6.36.26.2 9.29.19 71.770.267.4 22.722.722.9 1110.710.5 4.84.23.8
psa strat naive5.95.65.2 6.965.1 79.979.377.4 4.43.93.2 10.810.610.5 6.55.24
psa strat pda6.36.26.2 9.39.29.1 72.470.467.7 22.822.722.9 12.912.913 4.94.23.8
psa strat rf7.57.68.1 12.512.512.1 8283.385 21.82730.3 10.910.810.5 6.45.95.8
psa strat sda6.46.26.2 9.49.19.1 71.270.366.8 22.822.822.8 13.213.313.1 4.74.33.8
psa strat treebag888.1 12.512.412.4 68.669.167.2 21.922.324.6 14.413.711.8 6.665.9

References

  1. Rada, D. Ventajas e inconvenientes de la encuesta por Internet. Pap. Rev. Sociol. 2012, 97, 193–223. [Google Scholar]
  2. Elliott, M.R.; Valliant, R. Inference for nonprobability samples. Stat. Sci. 2017, 32, 249–264. [Google Scholar] [CrossRef]
  3. Meng, X.L. Statistical paradises and paradoxes in big data (I), Law of large populations, big data paradox, and the 2016 US presidential election. Ann. Appl. Stat. 2018, 12, 685–726. [Google Scholar] [CrossRef] [Green Version]
  4. Royall, R.M.; Herson, J. Robust estimation in finite populations I. J. Am. Stat. Assoc. 1973, 68, 880–889. [Google Scholar] [CrossRef]
  5. Valliant, R.; Dorfman, A.H.; Royall, R.M. Finite Population Sampling and Inference: A Prediction Approach; John Wiley: New York, NY, USA, 2000; No. 04, QA276. 6, V3. [Google Scholar]
  6. Buelens, B.; Burger, J.; van den Brakel, J.A. Comparing inference methods for non-probability samples. Int. Stat. Rev. 2018, 86, 322–343. [Google Scholar] [CrossRef]
  7. Lee, S. Propensity score adjustment as a weighting scheme for volunteer panel web surveys. J. Off. Stat. 2006, 22, 329–349. [Google Scholar]
  8. Lee, S.; Valliant, R. Estimation for volunteer panel web surveys using propensity score adjustment and calibration adjustment. Sociol. Methods Res. 2009, 37, 319–343. [Google Scholar] [CrossRef]
  9. Valliant, R.; Dever, J.A. Estimating propensity adjustments for volunteer web surveys. Sociol. Methods Res. 2011, 40, 105–137. [Google Scholar] [CrossRef]
  10. Ferri-García, R.; Rueda, M.D.M. Efficiency of propensity score adjustment and calibration on the estimation from non-probabilistic online surveys. Stat. Oper. Res. Trans. 2018, 1, 159–182. [Google Scholar]
  11. Valliant, R. Comparing Alternatives for Estimation from Nonprobability Samples. J. Surv. Stat. Methodol. 2020, 8, 231–263. [Google Scholar] [CrossRef]
  12. Ferri-García, R.; Rueda, M.D.M. Propensity score adjustment using machine learning classification algorithms to control selection bias in online surveys. PLoS ONE 2020, 15, e0231500. [Google Scholar]
  13. Couper, M.P. The future of modes of data collection. Public Opin. Q. 2011, 75, 889–908. [Google Scholar] [CrossRef] [Green Version]
  14. Rosenbaum, P.R.; Rubin, D.B. The central role of the propensity score in observational studies for causal effects. Biometrika 1983, 70, 41–55. [Google Scholar] [CrossRef]
  15. Taylor, H. Does internet research work? Int. J. Mark. 2000, 42, 1–11. [Google Scholar] [CrossRef]
  16. Taylor, H.; Bremer, J.; Overmeyer, C.; Siegel, J.W.; Terhanian, G. The record of internet-based opinion polls in predicting the results of 72 races in the November 2000 US elections. Int. J. Mark. Res. 2001, 43, 127–135. [Google Scholar]
  17. Schonlau, M.; Van Soest, A.; Kapteyn, A. Are ‘Webographic’ or attitudinal questions useful for adjusting estimates from Web surveys using propensity scoring? Surv. Res. Methods 2007, 1, 155–163. [Google Scholar] [CrossRef] [Green Version]
  18. Schonlau, M.; Couper, M.P. Options for conducting web surveys. Stat. Sci. 2017, 32, 279–292. [Google Scholar] [CrossRef]
  19. Lee, B.K.; Lessler, J.; Stuart, E.A. Improving propensity score weighting using machine learning. Stat. Med. 2010, 29, 337–346. [Google Scholar] [CrossRef] [Green Version]
  20. Phipps, P.; Toth, D. Analyzing establishment nonresponse using an interpretable regression tree model with linked administrative data. Ann. Appl. Stat. 2012, 6, 772–794. [Google Scholar] [CrossRef] [Green Version]
  21. Buskirk, T.D.; Kolenikov, S. Finding respondents in the forest: A comparison of logistic regression and random forest models for response propensity weighting and stratification. Surv. Methods Insights Field 2015, 1–17. [Google Scholar] [CrossRef]
  22. Rivers, D. Sampling for Web Surveys. In Proceeding of the Joint Statistical Meetings, Salt Lake City, UT, USA, 1 August 2007. [Google Scholar]
  23. Beaumont, J.F.; Bissonnette, J. Variance Estimation under Composite Imputation. The methodology behind SEVANI. Surv. Methodol. 2011, 37, 171–179. [Google Scholar]
  24. Fuller, W.A. Regression estimation for survey samples. Surv. Methodol. 2002, 28, 5–24. [Google Scholar]
  25. Hoerl, A.E.; Kennard, R.W. Ridge regression: Biased estimation for nonorthogonal problems. Technometrics 1970, 12, 55–67. [Google Scholar] [CrossRef]
  26. Hsiang, T. A Bayesian view on ridge regression. J. R. Soc. Ser. D 1975, 24, 267–268. [Google Scholar] [CrossRef]
  27. Tibshirani, R. Regression shrinkage and selection via the lasso. J. R. Stat. Soc. Ser. B 1996, 58, 267–288. [Google Scholar] [CrossRef]
  28. Park, T.; Casella, G. The bayesian lasso. J. Am. Stat. Assoc. 2008, 103, 681–686. [Google Scholar] [CrossRef]
  29. Van Houwelingen, J. Shrinkage and penalized likelihood as methods to improve predictive accuracy. Stat. Neerl. 2001, 55, 17–34. [Google Scholar] [CrossRef]
  30. James, G.; Witten, D.; Hastie, T.; Tibshirani, R. An Introduction to Statistical Learning; Springer: Berlin, Germany, 2013. [Google Scholar]
  31. Hastie, T.; Buja, A.; Tibshirani, R. Penalized discriminant analysis. Ann. Stat. 1995, 23, 73–102. [Google Scholar] [CrossRef]
  32. Hastie, T.; Tibshirani, R.; Friedman, J. The Elements of Statistical Learning: Data Mining, Inference, and Prediction; Springer Science & Business Media: Berlin, Germany, 2009. [Google Scholar]
  33. Breiman, L. Bagging predictors. Mach. Learn. 1996, 24, 123–140. [Google Scholar] [CrossRef] [Green Version]
  34. Sutton, C.D. Classification and regression trees, bagging, and boosting. Handb. Stat. 2005, 24, 303–329. [Google Scholar]
  35. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef] [Green Version]
  36. Niculescu-Mizil, A.; Caruana, R. Predicting good probabilities with supervised learning. In Proceedings of the 22nd International Conference on Machine Learning, Bergamo, Italy, 5–7 September 2015; pp. 625–632. [Google Scholar]
  37. Breiman, L. Arcing the Edge. Tech. Rep., Technical Report 486; Statistics Department, University of California: Berkeley, CA, USA, 1997. [Google Scholar]
  38. Friedman, J.H. Greedy function approximation: A gradient boosting machine. Ann. Stat. 2001, 1189–1232. [Google Scholar] [CrossRef]
  39. Touzani, S.; Granderson, J.; Fernandes, S. Gradient boosting machine for modeling the energy consumption of commercial buildings. Energy Build. 2018, 158, 1533–1543. [Google Scholar] [CrossRef] [Green Version]
  40. Peterson, L.E. K-nearest neighbor. Scholarpedia 2009, 4, 1883. [Google Scholar] [CrossRef]
  41. García, S.; Luengo, J.; Herrera, F. Data Preprocessing in Data Mining; Springer International Publishing: Cham, Switzerland, 2015. [Google Scholar]
  42. Okut, H. Bayesian regularized neural networks for small n big p data. In Artificial Neural Networks-Models and Applications; IN-TECH: Munich, Germany, 2016. [Google Scholar]
  43. Rumelhart, D.E.; Hinton, G.E.; Williams, R.J. Learning representations by back-propagating errors. Nature 1986, 323, 533–536. [Google Scholar] [CrossRef]
  44. Ripley, B.D. Pattern Recognition and Neural Networks; Cambridge University Press: Cambridge, UK, 1996. [Google Scholar]
  45. Breidt, F.J.; Opsomer, J.D. Model-assisted survey estimation with modern prediction techniques. Stat. Sci. 2017, 32, 190–205. [Google Scholar] [CrossRef]
  46. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  47. Baesens, B.; Viaene, S.; Poel, V.d.D.; Vanthienen, J.; Dedene, G. Bayesian neural network learning for repeat purchase modelling in direct marketing. Eur. J. Oper. Res. 2002, 138, 191–211. [Google Scholar] [CrossRef] [Green Version]
  48. Csáji, B.C. Approximation with artificial neural networks. Fac. Sci. Etvs Lornd Univ. Hung. 2001, 24, 7. [Google Scholar]
  49. National Institute of Statistics (INE). Life Conditions Survey. Microdata. Available online: https://www.ine.es/dyngs/INEbase/en/operacion.htm?c=Estadistica_C&cid=1254736176807menu=resultados&&idp=1254735976608#!tabs-1254736195153 (accessed on 30 May 2020).
  50. Gutiérrez, H.A. Estrategias de Muestreo Diseño de Encuestas y Estimacion de Parametros; Universidad Santo Tomas: Bogota, Colombia, 2009. [Google Scholar]
  51. Moro, S.; Cortez, P.; Rita, P. A data-driven approach to predict the success of bank telemarketing. Decis. Support Syst. 2014, 62, 22–31. [Google Scholar] [CrossRef] [Green Version]
  52. Kuhn, M. Caret: Classification and Regression Training; R Package Version 6.0-81; R Foundation for Statistical Computing: Vienna, Austria, 2018. [Google Scholar]
  53. R Core Team. R: A Language and Environment for Statistical Computing; R Foundation for Statistical Computing: Vienna, Austria, 2018. [Google Scholar]
  54. Chen, J.K.T.; Valliant, R.L.; Elliott, M.R. Calibrating non-probability surveys to estimated control totals using LASSO, with an application to political polling. J. R. Stat. Soc. Ser. C (Appl. Stat.) 2019, 68, 657–681. [Google Scholar] [CrossRef]
  55. James, W.; Stein, C. Estimation with quadratic loss. In Proceedings of the Fourth Berkeley Symposium on Mathematical Statistics and Probability; University of California Press: Berkeley, CA, USA, 1961; Volume 1, pp. 311–319. [Google Scholar]
  56. Copas, J.B. The shrinkage of point scoring methods. J. R. Stat. Soc. Ser. C (Appl. Stat.) 1993, 42, 315–331. [Google Scholar] [CrossRef]
  57. Arcos, A.; Contreras, J.M.; Rueda, M. A Novel Calibration Estimator in Social Surveys. Sociol. Methods Res. 2014, 43, 465–489. [Google Scholar] [CrossRef]
Table 1. Mean and median efficiency (%) of each estimator and the number of times it has been among the best.
Table 1. Mean and median efficiency (%) of each estimator and the number of times it has been among the best.
MeanMedianBest MeanMedianBest
matching ridge61.563.810psa gbm30.728.83
matching glm61.564.210psa strat naive30.532.13
matching glmnet6162.87psa strat knn25.624.70
matching brnn57.361.77matching lasso24.614.43
matching blasso57.159.96psa strat glm24.528.61
matching bridge55.861.27psa strat lda23.427.50
matching knn55.851.73psa strat sda23.227.20
psa glm46.453.85psa strat pda23.227.20
psa sda46.251.74psa strat avNNet21.823.40
psa glmnet46.153.43psa strat glmnet21.728.11
psa lda4651.23psa strat gbm16.816.20
psa pda45.751.74psa treebag10.111.60
psa naive41.256.96psa strat treebag7.63.50
psa knn38.542.43psa strat rf3.64.40
psa avNNet34.233.20psa rf−4.57.80
matching gbm32.234.90psa strat J48−23.93.80
matching treebag31.449.11psa J48−36.77.90
Table 2. Linear mixed-effects model on the modulus of relative bias ( | R B i a s | ) considering adjustment methods and algorithms as fixed effects and datasets as random effects. Reference levels: Bagged Trees (treebag) algorithm and matching adjustment.
Table 2. Linear mixed-effects model on the modulus of relative bias ( | R B i a s | ) considering adjustment methods and algorithms as fixed effects and datasets as random effects. Reference levels: Bagged Trees (treebag) algorithm and matching adjustment.
CoefficientEstimateStd. ErrorD. f.t ValueIC 95%p-Value
(Intercept)12.7558.475.12761.506[−8.857; 34.367]0.19104
glm−3.3591.224258.00−2.745[−5.769; −0.950]0.00647
glmnet−2.7531.224258.00−2.250[−5.163; −0.343]0.02530
knn−2.9601.224258.00−2.419[−5.370; −0.551]0.01624
gbm−0.6241.224258.00−0.510[−3.034; 1.785]0.61042
psa6.8440.948258.007.221[4.978; 8.710]5.79 ×   10 12
psa strat9.6010.948258.0010.129[7.734; 11.467]<2 ×   10 16
GroupVarianceStd. Dev.
Dataset424.2120.596
Residual40.426.358
DatasetSamplingIntercept
P1P1S11.590
P1P1S24.522
P2P2S153.693
P2P2S213.029
P3P3S14.110
P3P3S2−0.414
Table 3. Linear mixed-effects model on relative standard deviation considering adjustment methods and algorithms as fixed effects and datasets as random effects. Reference levels: treebag algorithm and Matching adjustment.
Table 3. Linear mixed-effects model on relative standard deviation considering adjustment methods and algorithms as fixed effects and datasets as random effects. Reference levels: treebag algorithm and Matching adjustment.
CoefficientEstimateStd. ErrorD. f.t ValueIC 95%p-Value
(Intercept)3.6860.51814.417.112[2.578; 4.795]4.5 ×   10 6
glm−2.0950.430258.00−4.866[−2.942; −1.247]2.0 ×   10 6
glmnet−2.2240.430258.00−5.167[−3.071; −1.376]4.8 ×   10 7
knn−1.9860.430258.00−4.614[−2.833; −1.138]6.2 ×   10 6
gbm−2.0200.430258.00−4.694[−2.868; −1.173]4.4 ×   10 6
psa0.6230.333258.001.869[−0.033; 1.280]0.0628
psa strat0.4040.333258.001.213[−0.252; 1.061]0.2262
GroupVarianceStd. Dev.
Dataset0.8340.913
Residual5.0022.236
DatasetSamplingIntercept
P1P1S12.479
P1P1S22.703
P2P2S14.391
P2P2S24.064
P3P3S14.226
P3P3S24.254
Table 4. Linear mixed-effects model on RMSE considering adjustment methods and algorithms as fixed effects and datasets as random effects. Reference levels: treebag algorithm and matching adjustment.
Table 4. Linear mixed-effects model on RMSE considering adjustment methods and algorithms as fixed effects and datasets as random effects. Reference levels: treebag algorithm and matching adjustment.
CoefficientEstimateStd. ErrorD. f.t ValueIC 95%p-Value
(Intercept)14.0928.3745.131.683[−7.271; 35.455]0.15174
glm−4.0941.222258.00−3.351[−6.500; −1.688]0.00093
glmnet−3.5151.222258.00−2.877[−5.920; −1.109]0.00435
knn−3.6391.222258.00−2.978[−6.045; −1.233]0.00317
gbm−1.2881.222258.00−1.054[−3.694; 1.118]0.29272
psa6.7440.946258.007.126[4.880; 8.607]1.0 ×   10 11
psa strat9.2460.946258.009.771[7.383; 11.110]<2 ×   10 16
GroupVarianceStd. Dev.
Dataset414.520.359
Residual40.36.348
DatasetSamplingIntercept
P1P1S12.436
P1P1S25.367
P2P2S154.549
P2P2S214.515
P3P3S15.949
P3P3S21.737

Share and Cite

MDPI and ACS Style

Castro-Martín, L.; Rueda, M.d.M.; Ferri-García, R. Inference from Non-Probability Surveys with Statistical Matching and Propensity Score Adjustment Using Modern Prediction Techniques. Mathematics 2020, 8, 879. https://doi.org/10.3390/math8060879

AMA Style

Castro-Martín L, Rueda MdM, Ferri-García R. Inference from Non-Probability Surveys with Statistical Matching and Propensity Score Adjustment Using Modern Prediction Techniques. Mathematics. 2020; 8(6):879. https://doi.org/10.3390/math8060879

Chicago/Turabian Style

Castro-Martín, Luis, Maria del Mar Rueda, and Ramón Ferri-García. 2020. "Inference from Non-Probability Surveys with Statistical Matching and Propensity Score Adjustment Using Modern Prediction Techniques" Mathematics 8, no. 6: 879. https://doi.org/10.3390/math8060879

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop