Performance of Subgroups
2.2MB
Loading Images

Although national monitoring has been designed primarily to present an overall national picture of student achievement, there is some provision for reporting on performance differences among subgroups of the sample. Eight demographic variables are available for creating subgroups, with students divided into subgroups on each variable, as detailed in Chapter 1.

Analyses of the relative performance of subgroups used the total score for each task, created as described in Chapter 1.

 School Variables [ top | back ]
Five of the demographic variables related to the schools the students attended. For these five variables, statistical significance testing was used to explore differences in task performance among the subgroups. Where only two subgroups were compared (for School Type), differences in task performance between the two subgroups were checked for statistical significance using t-tests. Where three subgroups were compared, one-way analysis of variance was used to check for statistically significant differences among the three subgroups.

Because the number of students included in each analysis was quite large (approximately 450), the statistical tests were quite sensitive to small differences. To reduce the likelihood of attention being drawn to unimportant differences, the critical level for statistical significance for tasks reporting results for individual students was set at p = .01 (so that differences this large or larger among the subgroups would not be expected by chance in more than one percent of cases). For tasks administered to teams or groups of students, p = .05 was used as the critical level to compensate for the smaller numbers of cases in the subgroups.
For the first four of the five school variables, statistically significant differences among the subgroups were found for less than 11 percent of the tasks for both year 4 and year 8. For the remaining variable, statistically significant differences were found on more than half of the tasks at both levels. In the detailed report below, all “differences” mentioned are statistically significant (to save space, the words “statistically significant” are omitted).
 School Size [ top | back ]

Results were compared from students in large, medium-sized and small schools (exact definitions were given in Chapter 1).

For year 4 students, there was a difference among the three subgroups on one of the 30 tasks, with students from small schools scoring lowest on Link Task 19 (p45). There were no differences on questions of the Information Skills Survey (p47).

For year 8 students, there were no differences on any of the 46 tasks, or on questions of the Information Skills Survey.

 Community Size [ top | back ]
Results were compared for students living in communities containing over 100,000 people (main centre), communities containing 10,000 to 100,000 people (provincial city) and communities containing less than 10,000 people (rural areas).

For year 4 students, there were differences among the three subgroups on two of the 30 tasks. Students from the main centres scored highest on Breakdancing and lowest on Library Search. There were no differences on questions of the Information Skills Survey.

For year 8 students, there were no differences on any of the 46 tasks, or on questions of the Information Skills Survey.
 School Type [ top | back ]

Results were compared for year 8 students attending full primary and intermediate (or middle) schools. There were no differences between these two subgroups on any of the 46 tasks, or on questions of the Information Skills Survey.

This year, for the first time, the NEMP samples included enough year 8 students attending year 7 to 13 high schools to permit comparisons between them and students attending intermediate schools. There were statistically significant differences (p < .01) on three of the 40 tasks attempted by individual students. Students from year 7 to 13 high schools scored higher on Hens, Atlas Y8 and Please, Mum!. There were no differences on questions of the Information Skills Survey.

 Zone [ top | back ]

Results achieved by students from Auckland, the rest of the North Island, and the South Island were compared.

For year 4 students, there were differences among the three sub-groups on one of the 30 tasks. Students from the rest of the North Island (excluding Auckland) scored highest on Library Search. There were no differences on questions of the Information Skills Survey.

For year 8 students, there were differences among the three subgroups on five of the 46 tasks: students from the South Island scored highest on Link Tasks 1, 2 and 3, City Mountains, and Please, Mum!. There were no differences on questions of the Information Skills Survey.

 Socio-Economic Index [ top | back ]
Schools are categorised by the Ministry of Education based on census data for the census mesh blocks where children attending the schools live. The resulting index takes into account household income levels and categories of employment. It uses 10 subdivisions, each containing 10 percent of schools (deciles 1 to 10). For our purposes, the bottom three deciles (1-3) formed the low decile group, the middle four deciles (4-7) formed the medium decile group, and the top three deciles (8-10) formed the high decile group. Results were compared for students attending schools in each of these three decile groups.

For year 4 students, there were differences among the three subgroups on 17 of the 30 tasks, spread across the three task chapters. Because of the number of tasks showing differences, they are not listed here. Students in high decile schools performed better than students in low decile schools on all 17 tasks, usually with larger gaps between low and medium decile schools than between medium and high decile schools. There was also a difference on one question of the Information Skills Survey: students from low decile schools were most positive about hunting for information (question 5)

For year 8 students, there were differences among the three subgroups on 25 of the 46 tasks, spread across the three task chapters but including 16 of the 21 year 8 tasks in Chapter 4. Because of the number of tasks showing differences, they are not listed here. Students in high decile schools performed better than students in low decile schools on all 25 tasks, usually with larger gaps between low and medium decile schools than between medium and high decile schools. There was also a difference on one question of the Information Skills Survey: students from low decile schools were most positive about writing down what they found out (question 8).
 Student Variables [ top | back ]

Three demographic variables related to the students themselves:

Gender: boys and girls
Ethnicity: Mäori, Pasifika and Pakeha (this term was used for all other students)
Language used predominantly at home: English and other.

During the previous cycle of the Project (1999–2002), special supplementary samples of students from schools with at least 15 percent Pasifika students enrolled were included. These allowed the results of Pasifika students to be compared with those of Mäori and Pakeha students attending these schools. By 2002, with Pasifika enrolments having increased nationally, it was decided that from 2003 onwards a better approach would be to compare the results of Pasifika students in the main NEMP samples with the corresponding results for Mäori and Pakeha students. This gives a nationally representative picture, with the results more stable because the numbers of Mäori and Pakeha students in the main samples are much larger than their numbers previously in the special samples.

The analyses reported compare the performances of boys and girls, Pakeha and Mäori students, Pakeha and Pasifika students, and students from predominantly English-speaking and non-English-speaking homes.

For each of these three comparisons, differences in task performance between the two subgroups are described using “effect sizes” and statistical significance.

For each task and each year level, the analyses began with a t-test comparing the performance of the two selected subgroups and checking for statistical significance of the differences. Then the mean score obtained by students in one subgroup was subtracted from the mean score obtained by students in the other subgroup, and the difference in means was divided by the pooled standard deviation of the scores obtained by the two groups of students. This computed effect size describes the magnitude of the difference between the two subgroups in a way that indicates the strength of the difference and is not affected by the sample size. An effect size of +.30, for instance, indicates that students in the first subgroup scored, on average, three tenths of a standard deviation higher than students in the second subgroup.

For each pair of subgroups at each year level, the effect sizes of all available tasks were averaged to produce a mean-effect size for the curriculum area and year level, giving an overall indication of the typical performance difference between the two subgroups.

 Gender [ top | back ]

Results achieved by male and female students were compared using the effect-size procedures.

For year 4 students, the mean-effect size across the 24 tasks was 0.14 (girls averaged 0.14 standard deviations higher than boys). This difference is small. There were statistically significant (p < .01) differences favouring girls on 6 of the 24 tasks: Hens, Link Task 3, Bats, Link Task 8, Oh Pussy Cat, Pussy Cat! and Link Task 16. There was also a difference on one question of the Information Skills Survey: girls were more positive than boys about writing down what they found out (question 8).

For year 8 students, the mean-effect size across the 40 tasks was 0.27 (girls averaged 0.27 standard deviations higher than boys): a moderate difference. There were statistically significant differences on 24 of the 40 tasks, with girls performing better on all 24 tasks, spread across the three task chapters. Because of the number of tasks showing differences, they are not listed here. There were also differences on five of the eight questions of the Information Skills Survey. Girls reported that they more often had a really interesting study for which they had to find information (question 2) and more often voluntarily looked up information (question 3). Girls also were more positive about hunting for information (question 5), about how good they thought they were at hunting for information (question 6), and about how much they liked writing down what they found out (question 8).

 Ethnicity [ top | back ]

Results achieved by Mäori, Pasifika and Pakeha (all other) students were compared using the effect-size procedures. First, the results for Pakeha students were compared to those for Mäori students. Second, the results for Pakeha students were compared to those for Pasifika students.

 Pakeha-Mäori Comparisons [ top | back ]

For year 4 students, the mean-effect size across the 24 tasks was 0.36 (Pakeha students averaged 0.36 standard deviations higher than Mäori students). This is a moderate difference. There were statistically significant differences (p <. 01) on 17 of the 24 tasks, spread across the three task chapters. Pakeha students scored higher than Mäori students on all 17 tasks. Because of the number of tasks showing differences, they are not listed here. There were no differences on questions of the Information Skills Survey.

For year 8 students, the picture was similar. The mean-effect size across the 40 tasks was 0.27 (Pakeha students averaged 0.27 standard deviations higher than Mäori students). This is a moderate difference. There were statistically significant differences on 18 of the 40 tasks, spread across the three task chapters. Pakeha students scored higher than Mäori students on all 18 tasks. Because of the number of tasks showing differences, they are not listed here. There were no differences on questions of the Information Skills Survey.

 Pakeha-Pasifika Comparisons [ top | back ]

Readers should note that only 30 to 45 Pasifika students were included in the analysis for each task. This is lower than normally preferred for NEMP subgroup analyses, but has been judged adequate for giving a useful indication, through the overall pattern of results, of the Pasifika students’ performance. Because of the relatively small numbers of Pasifika students, p = .05 has been used here as the critical level for statistical significance.

For year 4 students, the mean-effect size across the 24 tasks was 0.37 (Pakeha students averaged 0.37 standard deviations higher than Pasifika students). This is a moderate difference. There were statistically significant differences on 14 of the 24 tasks, spread across the three task chapters. Pakeha students scored higher on all 14 tasks. Because of the number of tasks showing differences, they are not listed here. There was also a difference on one question of the Information Skills Survey: Pasifika students reported having to find information for a study more frequently (question 1).

For year 8 students, the mean-effect size across the 40 tasks was 0.46 (Pakeha students averaged 0.46 standard deviations higher than Pasifika students). This is a large difference. There were statistically significant differences on 29 of the 40 tasks, spread across the three task chapters, but including all tasks in Chapter 3. Pakeha students scored higher on all 29 tasks. Because of the number of tasks showing differences, they are not listed here. There were also differences on four questions of the Information Skills Survey. Pasifika students reported having to find information for a study more frequently (question 1) and more often voluntarily looking up information (question 3). They also were more positive about hunting for information (question 5), and about how much they liked writing down what they found out (question 8).

 Home Language [ top | back ]

Results achieved by students who reported that English was the predominant language spoken at home were compared, using the effect-size procedures, with the results of students who reported predominant use of another language at home (most commonly an Asian or Pasifika language). Because of the relatively small numbers in the “other language” group, p = .05 has been used here as the critical level for statistical significance.

For year 4 students, the mean-effect size across the 24 tasks was 0.16 (students for whom English was the predominant language at home averaged 0.16 standard deviations higher than the other students). This is a small difference. There were statistically significant differences on three of the 24 tasks: students for whom English was the predominant language spoken at home scored higher on Link Task 1, Atlas Y4 and Oh Pussy Cat, Pussy Cat!. There was also a difference on one question of the Information Skills Survey. Students whose predominant language at home was not English reported that they more often had a really interesting study for which they had to find information (question 2).

For year 8 students, the mean-effect size across the 40 tasks was 0.18 (students for whom English was the predominant language at home averaged 0.18 standard deviations higher than the other students). This is a small difference. There were statistically significant differences on 16 of the 40 tasks: students for whom English was the predominant language spoken at home scored lower on Link Task 14, but higher on the other 15 tasks, spread across the three task chapters. Because of the number of tasks showing differences, they are not listed here. There were also differences on four questions of the Information Skills Survey. Students whose predominant language at home was not English reported that they more often voluntarily looked up information (question 3). They also were more positive about hunting for information (question 5), about how good they thought they were at hunting for information (question 6), and about how much they liked writing down what they found out (question 8)

 Summary, with Comparisons to Previous Information Skills Assessments [ top | back ]

School type (full primary, intermediate, or year 7 to 13 high school), school size, community size and geographic zone did not seem to be important factors predicting achievement on the information skills tasks. The same was true for the 2001 and 1997 assessments. However, there were statistically significant differences in the performance of students from low, medium and high decile schools on 57 percent of the tasks at year 4 level (compared to 43 percent in 2001 and 81 percent in 1997) and 54 percent of the tasks at year 8 level (compared to 71 percent in 2001 and 56 percent in 1997).

For the comparisons of boys with girls, Pakeha with Mäori, Pakeha with Pasifika students, and students for whom the predominant language at home was English with those for whom it was not, effect sizes were used. Effect size is the difference in mean (average) performance of the two groups, divided by the pooled standard deviation of the scores on the particular task. For this summary, these effect sizes were averaged across all tasks.

Year 4 girls averaged slightly higher than boys, with a mean effect size of 0.14 (compared to 0.06 in 2001). Year 8 girls averaged moderately higher than boys, with a mean effect size of 0.27 (compared to 0.15 in 2001). As was also true in 2001, the Information Skills Survey results at both year levels showed some evidence that girls were more positive than boys about information skills activities.

Pakeha students averaged moderately higher than Mäori students, with mean effect sizes of 0.36 for year 4 students and 0.27 for year 8 students (the corresponding figures in 2001 were 0.25 and 0.39).

Year 4 Pakeha students averaged moderately higher than Pasifika students, with a mean effect size of 0.37 (compared to 0.40 in 2001). Year 8 Pakeha students averaged substantially higher than Pasifika students, with a mean effect size of 0.48 (compared to 0.46 in 2001). The Information Skills Survey results showed that Pasifika students were more involved in and enthusiastic about some aspects of information skills.

Compared to students for whom the predominant language at home was English, students from homes where other languages predominated averaged slightly lower, with mean effect sizes of 0.16 for year 4 students and 0.18 for year 8 students. Comparative figures are not available for the assessments in 2001.

 
Loading Images