Teens, kindness and cruelty on social network sites
Prepared by Princeton Survey Research Associates International for the Pew Research Center’s Internet and American Life Project
The 2011 Teens and Digital Citizenship Survey sponsored by the Pew Research Center’s Internet and American Life Project obtained telephone interviews with a nationally representative sample of 799 teens ages 12 to 17 years old and their parents living in the continental United States. The survey was conducted by Princeton Survey Research Associates International. The interviews were conducted in English and Spanish by Princeton Data Source, LLC from April 19 to July 14, 2011. Statistical results are weighted to correct known demographic discrepancies. The margin of sampling error for the complete set of weighted data is ±4.8 percentage points.
Unless otherwise noted, the adult data presented in this report are from the Pew Research Center’s Internet and American Life Project’s August Tracking survey of adults 18 and older. The survey was conducted from July 25 through August 26, 2011 with a sample size of 2260 adults, including 916 from cell phone interviews. Interviews were conducted in English and Spanish. The margin of error for the entire sample is ±2 percentage points and ±3 percentage points for results based on social media users. Social media users are defined as those who report that they use a social network website or Twitter.
In addition to the two surveys, this study conducted 7 focus groups with teens between the ages of 12 and 19 in the greater Washington, DC metro area in January and February 2011. Participants were recruited via word of mouth, email, schools, and non-profit organizations. A total of 57 youth participated in the focus groups, though each group averaged 8 to 14 people. Groups were co-ed, but were broken into middle school and high school aged youth. The groups were balanced for gender and crossed the socio-economic and family structure spectrum. Black youth were over-represented. All participants were required to have access to either a computer or a cell phone to participate. Participants were paid a $40 cash incentive for their participation. Parental consent was obtained for all minor participants, as was the assent of the minor participants themselves. Eighteen and 19 year-old participants consented to their own participation.
Further details on the design, execution, and analysis of the teen and parent telephone survey are discussed below.
Design and Data Collection Procedures
A combination of landline and cellular random digit dial (RDD) samples was used to represent all teens and their parents in the continental United States who have access to either a landline or cellular telephone. Both samples were provided by Survey Sampling International, LLC (SSI) according to PSRAI specifications.
Both samples were disproportionately stratified to increase the incidence of blacks and Latinos. The same stratification scheme was used for both sample frames and was based on the estimated incidence of minority groups at the county level. All counties in the continental United States were divided into ten strata based on the estimated proportion of African American and Latino populations. Strata with higher minority densities were oversampled relative to strata with lower densities. Phone numbers were drawn with equal probabilities within strata. The disproportionate sample design was accounted for in the weighting and does not affect the representative nature of the sample.1
Interviews were conducted from April 19 to July 14, 2011. As many as 7 attempts were made to contact and interview a parent at every sampled telephone number. After the parent interview, if the teen was not immediately available, an additional 7 calls were made to interview an eligible teen. Sample was released for interviewing in replicates, which are representative subsamples of the larger sample. Using replicates to control the release of sample ensures that complete call procedures are followed for the entire sample. Calls were staggered over times of day and days of the week to maximize the chance of making contact with potential respondents. Each telephone number received at least one daytime call in an attempt to complete an interview.
Contact procedures were slightly different for the landline and cell samples. For the landline sample, interviewers first determined if the household had any 12 to 17 year-old residents. Households with no teens were screened-out as ineligible. In eligible households, interviewers first conducted a short parent interview with either the father/male guardian or mother/female guardian. The short parent interview asked some basic household demographic questions as well as questions about a particular teen in the household (selected at random if more than one teen lived in the house.)
For the cell phone sample, interviews first made sure that respondents were in a safe place (for example, not driving) to talk and that they were speaking with an adult. Calls made to minors were screened-out as ineligible. If the person was not in a safe place to talk a callback was scheduled. Interviewers then asked if any 12 to 17 year-olds lived in their household. Cases where no teens lived in the household were screened-out as ineligible. If there was an age-eligible teen in the household, the interviewers asked if the person on the cell phone was a parent of the child. Those who were parents went on to complete the parent interview. Those who were not parents were screened-out as ineligible.
For both samples, after the parent interview was complete an interview was completed with the target child. Data was kept only if the child interview was completed.2
Interviewers were given instructions to tell parents – if asked – that they should not remain on the phone with the child during the interview, but that if they were concerned they could sit nearby. The interviewer then coded whether or not the parent remained on the phone with the child. In this survey, 90 of the 799 interviews (or 11%) had a parent listening on the phone during the child’s interview. Parents who elected to remain on the phone while their child completed the interview were more likely to be listening to the interviews of girls and children age 12 and to a lesser extent, age 13. These parents were also more likely to be white. Teens whose parents listened to their interview were less likely to use the internet, use social network sites, or go online using a mobile phone. Among those teens whose parents attended their interview who do use social network sites, they were more likely to report using Facebook than teens whose parents did not listen in. We elected to retain these interviews as a part of our larger sample – first because there were very few statistically significant differences between the responses of teens whose parents listened in, and those whose parents did not. Second, in the places where we did see modest differences, understanding what might be behind those differences was difficult to tease out – the age of the child may have been a factor, or how the parent parented that child, or the fact that the parent was listening to the interview. So rather than introduce additional bias into the data, we elected to leave the cases in the data set, and note in the text where the parent’s listening made a statistically significant difference in the responses of the teen.
Weighting and analysis
Weighting is generally used in survey analysis to compensate for patterns of nonresponse and disproportionate sample designs that might bias survey estimates. This sample was weighted in three stages. The first stage of weighting corrected for the disproportionate RDD sample designs. For each stratum the variable SAMPWT was computed as the ratio of the size of the sample frame in the stratum divided by the amount of sample ordered in the stratum.
The second stage of weighting involved correcting for different probabilities of selection based on respondents’ phone use patterns. Respondents who have both a landline and a cell phone have a greater chance of being sampled than respondents with access to only one kind of phone. To correct for this we computed a variable called PUA (Phone Use Adjustment). Respondents with one kind of phone (either landline or cell) were assigned a PUA of 0.5 while respondents with both types of phones were assigned a PUA of 1.0. SAMPWT and PUA were then multiplied together to use as an input weight (WEIGHT1) for post-stratification raking
The interviewed sample was raked to match national parameters for both parent and child demographics. The parent demographics used for weighting were: sex; age; education; race; Hispanic origin; number of 12-17 year-olds in household; phone use and region (U.S. Census definitions). The child demographics used for weighting were gender and age. The parameters came from a special analysis of the Census Bureau’s 2010 Annual Social and Economic Supplement (ASEC) that included all households in the continental United States. The phone use parameter was derived from recent PSRAI survey data.
Raking was accomplished using Sample Balancing, a special iterative sample weighting program that simultaneously balances the distributions of all variables using a statistical technique called the Deming Algorithm. Weights were trimmed to prevent individual interviews from having too much influence on the final results. The use of these weights in statistical analysis ensures that the demographic characteristics of the sample closely approximate the demographic characteristics of the national population. Table 1 compares weighted and unweighted sample distributions to population parameters.
Effects of Sample Design on Statistical Inference
Post-data collection statistical adjustments require analysis procedures that reflect departures from simple random sampling. PSRAI calculates the effects of these design features so that an appropriate adjustment can be incorporated into tests of statistical significance when using these data. The so-called “design effect” or deffrepresents the loss in statistical efficiency that results from systematic non-response. The total sample design effect for this survey is 1.95.
PSRAI calculates the composite design effect for a sample of size n, with each case having a weight, wi as:
In a wide range of situations, the adjusted standard error of a statistic should be calculated by multiplying the usual formula by the square root of the design effect (√deff ). Thus, the formula for computing the 95% confidence interval around a percentage is:
where p is the sample estimate and n is the unweighted number of sample cases in the group being considered.
The survey’s margin of error is the largest 95% confidence interval for any estimated proportion based on the total sample— the one around 50%. For example, the margin of error for the entire sample is ±4.8%. This means that in 95 out every 100 samples drawn using the same methodology, estimated proportions based on the entire sample will be no more than 4.8 percentage points away from their true values in the population. It is important to remember that sampling fluctuations are only one possible source of error in a survey estimate. Other sources, such as respondent selection bias, questionnaire wording and reporting inaccuracy, may contribute additional error of greater or lesser magnitude.
Table 2 reports the disposition of all sampled callback telephone numbers ever dialed. The response rate estimates the fraction of all eligible respondents in the sample that were ultimately interviewed. At PSRAI it is calculated by taking the product of three component rates:3
- Contact rate – the proportion of working numbers where a request for interview was made4
- Cooperation rate – the proportion of contacted numbers where a consent for interview was at least initially obtained, versus those refused
- Completion rate – the proportion of initially cooperating and eligible interviews that agreed to the child interview and were completed
Thus the response rate for landline sample was 12 percent and the response rate for the cell sample was 7 percent.
Please see below for the sample disposition table.
- numoffset="53"For more information on oversampling, see the Pew Research Center for People and Press’s website and their discussion of the implications of this survey technique: http://www.people-press.org/methodology/sampling/oversamples/ ↩
- At the start of the field period, we used a modified screener that allowed us to complete a teen interview prior to a parent interview. After a few weeks in the field (April 19-June 1), it became clear that completing the teen interview first was not productive. Therefore the screener was modified to the one described here where a parent was always interviewed first. There are 16 “teen-first” interviews included in the overall sample. ↩
- PSRAI’s disposition codes and reporting are consistent with the American Association for Public Opinion Research standards. ↩
- PSRAI assumes that 75 percent of cases that result in a constant disposition of “No answer” or “Busy” are actually not working numbers. ↩