A. Survey MethodologyThe sampling frame for the survey was respondents to the Baseline Survey that was administered to applicants to YSC projects. Data of this type were available on 3,656 youth. Distinguishing successful applicants (“participants”) from unsuccessful ones (“non-participants”) required identifying participants using SICs from administrative data provided by the Program. Difficulties of selection owing to missing SIC data in the Baseline data and other factors led to the decision to survey what was essentially a census of all applicants and then post-code respondents as participants or non-participants on the basis of their self-identification. However, 120 names were eliminated at the outset because these applicants had previously been surveyed by the YSC program. The survey proceeded as follows:
The record of call, shown below, indicates that despite efforts to improve contact and tracing of applicants a great many respondents could not be contacted because no active phone number could be obtained. We started with 3100 unduplicated names and addresses, but 38.2% of these yielded no reply or did not find an applicant, were not in service, or could not generate a number. Outright refusals by potential respondents were relatively few (4.8%), but respondents who were never available or situations where terminations occurred for a variety of reasons added to that proportion. A curious category of result is found in the 158 persons who said that they did not apply. This raises questions about how the Baseline surveys were administered within projects, but the outcome may also be explained by some implicit refusals, misperceptions of the process or poor memory of events.
B. Non-Response Bias AnalysisA logistic regression model indicated several variables from the baseline survey associated with the likelihood of responding to the follow-up survey. The following conditions were associated to a statistically significant degree with greater likelihood of responding:
The following conditions were associated to a statistically significant degree with lesser likelihood of responding:
We tabulated the relative frequencies of response to the follow-up survey for the above-listed variables. For each variable that had several levels of response, the levels were combined, on the basis of similarity of response rates, into just two categories, and the response rates recalculated for the resulting categories. The inverses of these response rates were then applied to data for each of the respondents to the follow-up survey, with an appropriate adjusting factor to ensure that the total weighted sample size was equal to the unweighted sample size (1,242). The weights ranged from 0.68 to 2.05, with a mean (by design) of 1.00. These weights were then used in all the multivariate analyzes performed. The purpose of doing so was simply to ensure that the people who responded to the follow-up survey represented the baseline population as closely as possible on the basis of characteristics associated with differing response rates to the follow-up survey. For example, the response rate was significantly higher in the Atlantic region than elsewhere in the country. Therefore, responses from that region would receive relatively less weight in the analyzes because they represent fewer people than do responses from other parts of the country.
|