WLD_2009_PISA_v02_M
Programme for International Student Assessment 2009
Name | Country code |
---|---|
Albania | ALB |
Argentina | ARG |
Australia | AUS |
Austria | AUT |
Azerbaijan | AZE |
Belgium | BEL |
Brazil | BRA |
Bulgaria | BGR |
Canada | CAN |
Chile | CHL |
China | CHN |
Colombia | COL |
Costa Rica | CRI |
Croatia | HRV |
Czech Republic | CZE |
Denmark | DNK |
Estonia | EST |
Finland | FIN |
France | FRA |
Georgia | GEO |
Germany | DEU |
Greece | GRC |
Hong Kong SAR, China | HKG |
Hungary | HUN |
Iceland | ISL |
India | IND |
Indonesia | IDN |
Ireland | IRL |
Israel | ISR |
Italy | ITA |
Japan | JPN |
Jordan | JOR |
Kazakhstan | KAZ |
Kyrgyz Republic | KGZ |
Latvia | LVA |
Liechtenstein | LIE |
Lithuania | LTU |
Luxembourg | LUX |
Macao SAR, China | MAC |
Malaysia | MYS |
Malta | MLT |
Mauritius | MUS |
Mexico | MEX |
Montenegro | MNE |
Netherlands Antilles | ANT |
Netherlands | NLD |
New Zealand | NZL |
Norway | NOR |
Panama | PAN |
Peru | PER |
Poland | POL |
Portugal | PRT |
Qatar | QAT |
Romania | ROU |
Russian Federation | RUS |
Serbia | SRB |
Singapore | SGP |
Slovak Republic | SVK |
Slovenia | SVN |
Spain | ESP |
Sweden | SWE |
Switzerland | CHE |
Taiwan, China | TWN |
Thailand | THA |
Trinidad and Tobago | TTO |
Tunisia | TUN |
Turkiye | TUR |
United Arab Emirates | ARE |
United Kingdom | GBR |
United States | USA |
Uruguay | URY |
Venezuela, RB | VEN |
Vietnam | VNM |
PISA 2009 is the fourth round of PISA surveys. Following the detailed assessment of each of PISA’s three main subjects – reading, mathematics and science – in 2000, 2003 and 2006, the 2009 survey marks the beginning of a new round with a return to a focus on reading, but in ways that reflect the extent to which reading has changed since 2000, including the prevalence of digital texts.
PISA assesses the extent to which students near the end of compulsory education have acquired some of the knowledge and skills that are essential for full participation in modern societies, with a focus on reading, mathematics and science. PISA focuses on young people’s ability to use their knowledge and skills to meet real-life challenges. This orientation reflects a change in the goals and objectives of curricula themselves, which are increasingly concerned with what students can do with what they learn at school and not merely with whether they have mastered specific curricular content. PISA also aims to examine students' learning strategies, their competencies in areas such as problem solving that involves multiple disciplines and their interests in different topics. This kind of broader assessment started in PISA 2000, which asked students about their motivation and other aspects of their attitudes towards learning, their familiarity with computers and, under the heading "self-regulated learning", about their strategies for managing and monitoring their own education. The assessment of students' motivations and attitudes continued in PISA 2006, with special attention given to students' attitudes towards and interest in science. Returning to reading as the major subject of assessment, PISA 2009 focused on students' engagement in reading activities and their understanding about their own reading and learning strategies.
PISA’s unique features include:
Sample survey data [ssd]
The scope of the Programme for International Student Assessment 2009 includes:
Topic | Vocabulary |
---|---|
Education | World Bank |
75 countries and economies, including all 34 OECD countries and 41 partner countries and economies: Albania, Argentina, Australia, Austria, Azerbaijan, Belgium, Brazil, Bulgaria, Canada, Chile, Chinese Taipei, Colombia, Costa Rica, Croatia, Czech Republic, Denmark, Estonia, Finland, France, Georgia, Germany, Greece, Himachal Pradesh-India, Hong Kong-China, Hungary, Iceland, Indonesia, Ireland, Israel, Italy, Japan, Jordan, Kazakhstan Korea, Kyrgyzstan, Latvia, Liechtenstein, Lithuania, Luxembourg, Macao-China, Malaysia, Malta, Mauritius, Mexico, Miranda-Venezuela, Montenegro, Netherlands, Netherlands-Antilles, New Zealand, Norway, Panama, Peru, Poland, Portugal, Qatar, Romania, Russian Federation, Serbia, Shanghai-China, Singapore, Slovak Republic, Slovenia, Spain, Sweden, Switzerland, Tamil Nadu-India, Thailand, Trinidad and Tobago, Tunisia, Turkey, United Arab Emirates, United Kingdom, United States, Uruguay, Vietnam.
Name |
---|
Organisation for Economic Co-operation and Development (OECD) |
Name |
---|
Australian Council for Educational Research |
National Institute for Educational Measurement (Netherlands) |
Unité d’analyse des systèmes et pratiques d’enseignement (Belgium) |
cApStAn Linguistic Quality Control (Belgium) |
Deutches Institut für Internationale Pädagogische Forschung |
Westat (USA) |
Educational Testing Service (USA) |
National Institute for Educational Policy Research (Japan) |
Name |
---|
Organisation for Economic Co-operation and Development |
The specific sample design and size for each country aimed to maximise sampling efficiency for student-level estimates. In OECD countries, sample sizes ranged from 4,410 students in Iceland to 38,250 students in Mexico. Countries with large samples have often implemented PISA both at national and regional/state levels (e.g. Australia, Belgium, Canada, Italy, Mexico, Spain, Switzerland and the United Kingdom). The selection of samples was monitored internationally and adhered to rigorous standards for the participation rate, both among schools selected by the international contractor and among students within these schools, to ensure that the PISA results reflect the skills of the 15-year-old students in participating countries. Countries were also required to administer the test to students in identical ways to ensure that students receive the same information prior to and during the assessment.
When a school has been selected to participate in PISA, a school co-ordinator is appointed. The school coordinator compiles a list of all 15-year-olds in the school and sends this list to the PISA National Centre in the country, which randomly selects 35 students to participate. The school co-ordinator then contacts the students who have been selected for the sample and obtains the necessary permissions from parents. The testing session is usually conducted by a test administrator who is trained and employed by the National Centre. The test administrator contacts the school co-ordinator to schedule administration of the assessment. The school coordinator ensures that the students attend the testing sessions. This can sometimes be difficult because students may come from different grades and different classes. The test administrator's primary tasks are to ensure that each test booklet is distributed to the correct student and to introduce the tests to the students. After the test is over, the test administrator collects the test booklets and sends them to the National Centre for coding.
In PISA 2009, 13 different test booklets were used in each country. Each booklet had a different subset of PISA questions, so that students answered overlapping groups of questions, in order to produce a wide range of test items while limiting the test time for each student. With 13 different booklets, in each group of 35 students, no more than 3 students were given the same booklet. Booklets were allocated to individual students according to a random selection process. The test administrator's introduction came from a prescribed text so that all students in different schools and countries received exactly the same instructions. Before starting the test, the students were asked to do a practice question from their booklets. The testing session was divided into two parts: the two-hour-long test to assess their knowledge and skills, and the questionnaire session to collect data on their personal background, their learning habits, their attitudes towards reading, and their engagement and motivation. The length of the questionnaire session varied across countries, depending on the options chosen for inclusion, but generally was about 30 minutes. Students were usually given a short break half-way through the test and again before they did the questionnaire.
For more information on PISA's technical background, refer to the documents "PISA 2009 Results: Overcoming Social Background - Volume II", Annex A and "PISA 2009 Technical Report" that are provided as external resources.
School Response Rates: A response rate of 85% was required for initially selected schools. If the initial school response rate fell between 65% and 85%, an acceptable school response rate could still be achieved through the use of replacement schools. To compensate for a sampled school that did not participate, where possible, two potential replacement schools were identified. Furthermore, a school with a student participation rate between 25% and 50% was not considered as a participating school for the purposes of calculating and documenting response rates. However, data from such schools were included in the database and contributed to the estimates included in the initial PISA international report. Data from schools with a student participation rate of less than 25% were not included in the database, and such schools were regarded as non-respondents.
Student Response Rates: An overall response rate of 80% of selected students in participating schools was required. A student who had participated in the original or follow-up cognitive sessions was considered to be a participant. A minimum student response rate of 50% within each school was required for a school to be regarded as participating: the overall student response rate was computed using only students from schools with at least a 50% student response rate. Again, weighted student response rates were used for assessing this standard. Each student was weighted by the reciprocal of his/her sample selection probability.
For more detailed information on school and student response rates, refer to Chapter 4 in the technical document "PISA 2009 Technical Report" that is provided as an external resource.
Start | End | Cycle |
---|---|---|
2009 | 2009 | 1st Round |
2010 | 2010 | 2nd Round |
Use of the dataset must be acknowledged using a citation which would include:
Example:
Organisation for Economic Co-operation and Development. Programme for International Student Assessment (PISA) 2009. Ref. WLD_2009_PISA_v01_M. Dataset downloaded from [URL] on [date].
The user of the data acknowledges that the original collector of the data, the authorized distributor of the data, and the relevant funding agency bear no responsibility for use of the data or for interpretations or inferences based upon such uses.
Name | URL | |
---|---|---|
OECD PISA | edu.pisa@oecd.org | http://www.oecd.org/pisa/home/ |
DDI_WLD_2009_PISA_v03_M_WB
Name | Affiliation | Role |
---|---|---|
Development Economics Data Group | World Bank | Documentation of the study |
2011-05-09
DDI Document - Version 03 - (05/24/21)
This version is identical to DDI_WLD_2009_PISA_v02_M_WB but country field has been updated to capture all the countries covered by survey.
Version 02 (July 2014)