This chapter focuses on how the 2023 Survey of Adult Skills was designed, managed and conducted. It discusses the target population, exclusions from the survey, sample size and response rates, as well as how the overall quality of the data was assessed.
5. The methodology of the 2023 Survey of Adult Skills and the quality of data
Copy link to 5. The methodology of the 2023 Survey of Adult Skills and the quality of dataAbstract
The design and implementation of the 2023 Survey of Adult Skills was guided by associated technical standards and guidelines which were developed to ensure that the survey yielded high-quality and internationally comparable data. The Technical Standards and Guidelines (TSGs) for the 2023 Survey of Adult Skills articulate the standards participating countries and economies are expected to adhere to in implementing the survey and describes the steps they should follow in order to meet them1. It also makes recommendations for actions relating to the standards which were not mandatory but could help to produce high-quality data. Standards were established for 16 distinct aspects of the design and implementation of the survey (Table 5.1).
Table 5.1. Areas of activity covered by the 2023 Survey of Adult Skills Technical Standards and Guidelines
Copy link to Table 5.1. Areas of activity covered by the 2023 Survey of Adult Skills Technical Standards and Guidelines|
Standards |
|
|---|---|
|
Quality assurance and quality control |
Data collection staff training |
|
Ethics |
Data collection |
|
Survey planning |
Data capture |
|
Sample design |
Data file creation |
|
Survey instrument |
Confidentiality and data security |
|
Translation and adaptation |
Weighting |
|
Information technology |
Estimation |
|
Field management |
Documentation |
Box 5.1. How the survey was managed
Copy link to Box 5.1. How the survey was managedThe development and implementation of the 2023 Survey of Adult Skills was overseen by the PIAAC Board of Participating Countries (BPC). The BPC consisted of representatives from each of the participating countries and economies. Croatia participated as an observer. The BPC was responsible for making major decisions on budgets, developing and implementing the survey, reporting results, and monitoring the progress of the project. The BPC was supported in its work by the OECD Secretariat, which was responsible for providing advice to the BPC and managing the project on its behalf.
The OECD contracted an international consortium to undertake a range of tasks relating to the design and development of the assessment, implementation of the survey, and analysis of the resulting data. The consortium was responsible for developing questionnaires, assessment instruments and the computer-delivery platform; supporting survey operations; quality control; scaling; and database preparation.
Participating countries and economies were responsible for the national implementation of the survey, in particular for sampling, translation and adaptation of materials, data collection, and database production. In each country, all these activities were led and co-ordinated by a national project manager.
The TSG document is one element of a comprehensive process of quality assurance and control that was put in place to reduce potential sources of error, increase comparability and maximise the quality of the data produced by the 2023 Survey of Adult Skills. Participating countries and economies received assistance in meeting the standards in a variety of ways. Where relevant, manuals, training materials, testing plans and toolkits were produced. Training was provided to countries at appropriate stages of the project. In certain areas, such as sampling, translation and adaptation, and the operation of the computer-delivery platform, passage through the various stages of implementation was subject to a review of the steps completed, and sign-off was often required before moving to the next stage. Regular consultations were held with countries at project meetings and through bilateral contact. Compliance with the technical standards was monitored throughout the development and implementation phases through direct contact, the provision of evidence that required activities were completed and ongoing collection of data from countries about key aspects of implementation.
The quality of each country and economy’s data was reviewed prior to publication. The review was based on the analysis of the psychometric characteristics of the data and evidence of compliance with the technical standards. A data quality assessment was prepared for each country, and recommendations were made regarding release and, if necessary, any restrictions and/or qualifications that should apply to the release and publication. The Technical Advisory Group (TAG), comprised of independent experts, validated the approach to the review of data and the results of the analysis; the project’s steering body, the PIAAC Board of Participating Countries (BPC), made the final decision on release.
This chapter provides the background needed to correctly interpret the results of the reviews of data quality. It describes the following aspects of the design and the methodology of the 2023 Survey of Adult Skills:
survey and assessment design
sampling
translation and adaptation of survey instruments
survey administration
survey response rates and non-response bias analysis
the doorstep interview and literacy-related non-response
the overall assessment of data quality.
Survey and assessment design
Copy link to Survey and assessment designThe Survey of Adult Skills (PIAAC) collects data using a combination of a personal interview and a self-completed assessment. The 2023 Survey of Adult Skills has two components: a background questionnaire and a direct assessment.
The background questionnaire (BQ) is administered as a computer-aided personal interview. Trained interviewers ask the questions contained in the background questionnaire and record the answers using a tablet and a keyboard. The time needed to complete the background questionnaire varies depending on respondents’ characteristics (many questions in the background questionnaire only concern adults who are currently employed, for instance).
The second component is a direct assessment of literacy, numeracy and adaptive problem solving (APS) completed autonomously by the respondent. To complete the assessment, respondents use the same touchscreen tablet device used by the interviewer to administer the questionnaire. The tablet interface replicates the feeling of using paper-based instruments as much as possible. If they wish, respondents can also use a digital stylus to interact with the interface. The interviewers remain with the respondents to supervise them during the assessment. The interviewer is not supposed to help the respondent during the direct assessment but can assist if there are technical problems. Normally, the interviewer encourages respondents to persist through the assessment and to attempt to provide answers to the best of their ability. The direct assessment component is untimed, allowing respondents to take as much time as they need to complete it.
Figure 5.1 summarises the design of the survey and the various elements of the interview.
Figure 5.1. The 2023 Survey of Adult Skills assessment design
Copy link to Figure 5.1. The 2023 Survey of Adult Skills assessment design
The interview starts with the background questionnaire. The questionnaire collects detailed information on the socio-demographic characteristics of respondents (age, gender, migration status), as well as on their educational careers and labour-market outcomes (see Chapter 4 for more information). Respondents who are unable to complete the background questionnaire because of language barriers are administered the doorstep interview, discussed below.
Following completion of the background questionnaire, respondents first go through a short tablet tutorial to make sure they understand how to interact with the device and with the interface. The ease of use of the tablet, accompanied by a digital stylus, ensures that even adults with very low familiarity with digital devices can complete the assessment on the tablet.
The tablet-only design allows adaptive algorithms to be used to optimise the delivery of assessment items. These algorithms use information from the background questionnaire (such as age and educational attainment), as well as information from answers to previous assessment items, to select and administer assessment items that are neither too easy nor too difficult for the respondent.
The first step of this adaptive process is a locator test (also referred to as Stage 1), including eight literacy and eight numeracy items. Based on their answers to these items, respondents are sorted into three different paths:
Respondents who fail the locator test are directed to Path 1 and are administered an assessment of reading and numeracy component skills.
Respondents who pass the locator test with a low score are directed to Path 2. They take the reading and numeracy component skills assessment and then proceed to the direct assessment of literacy, numeracy and adaptive problem solving.
Respondents who pass the locator test with a high score are directed to Path 3. Some of these adults (12.5%) are randomly chosen to take the component skills assessment and then proceed to the direct assessment of literacy, numeracy and adaptive problem solving. The rest proceed directly to the assessment.
Table 5.2 reports the distribution of respondents across the different paths of the survey for each country and economy.
Table 5.2. Distribution of respondents across different survey paths
Copy link to Table 5.2. Distribution of respondents across different survey paths|
OECD countries and economies |
Doorstep Interview (%) |
Break-offs with a complete BQ (%) |
Path 1 (%) |
Path 2 (%) |
Path 3 (%) |
|---|---|---|---|---|---|
|
Austria |
2.8 |
4.2 |
1.6 |
9.9 |
81.4 |
|
Canada |
2.5 |
3.9 |
1.1 |
9.5 |
82.9 |
|
Chile |
0.0 |
3.0 |
7.1 |
26.4 |
63.4 |
|
Czechia |
4.5 |
0.0 |
0.5 |
9.3 |
85.6 |
|
Denmark |
5.3 |
4.3 |
0.5 |
4.5 |
85.4 |
|
Estonia |
1.2 |
1.1 |
1.0 |
9.0 |
87.6 |
|
Finland |
3.7 |
1.1 |
0.9 |
4.0 |
90.3 |
|
France |
0.6 |
6.2 |
2.2 |
12.3 |
78.7 |
|
Germany |
2.7 |
1.8 |
1.5 |
8.9 |
85.2 |
|
Hungary |
1.0 |
0.2 |
1.7 |
16.3 |
80.8 |
|
Ireland |
0.6 |
2.1 |
0.5 |
12.3 |
84.5 |
|
Israel |
0.8 |
1.3 |
3.4 |
19.1 |
75.4 |
|
Italy |
0.2 |
0.7 |
1.9 |
20.7 |
76.5 |
|
Japan |
1.1 |
0.2 |
0.9 |
5.4 |
92.4 |
|
Korea |
0.6 |
3.8 |
1.4 |
17.7 |
76.5 |
|
Latvia |
0.4 |
6.7 |
1.0 |
12.6 |
79.2 |
|
Lithuania |
2.6 |
1.6 |
1.3 |
15.3 |
79.3 |
|
Netherlands |
3.1 |
2.0 |
1.8 |
5.4 |
87.7 |
|
New Zealand |
0.1 |
3.9 |
4.1 |
10.6 |
81.3 |
|
Norway |
1.4 |
1.6 |
1.0 |
6.0 |
90.0 |
|
Poland |
0.1 |
0.2 |
7.1 |
15.2 |
77.4 |
|
Portugal |
3.0 |
1.2 |
3.3 |
20.2 |
72.2 |
|
Slovak Republic |
0.4 |
1.0 |
1.3 |
11.8 |
85.5 |
|
Spain |
2.1 |
0.5 |
1.1 |
15.8 |
80.4 |
|
Sweden |
- |
2.9 |
0.8 |
3.9 |
92.4 |
|
Switzerland |
3.3 |
0.4 |
2.1 |
8.0 |
86.2 |
|
United States |
0.9 |
1.1 |
4.8 |
15.1 |
78.1 |
|
Subnational entities |
|||||
|
England (UK) |
0.4 |
1.6 |
1.4 |
8.8 |
87.8 |
|
Flemish Region (Belgium) |
3.1 |
4.3 |
1.0 |
6.5 |
85.1 |
|
Partner countries |
|||||
|
Croatia |
0.3 |
2.4 |
2.0 |
13.7 |
81.5 |
|
Singapore |
- |
0.2 |
2.7 |
13.2 |
83.9 |
The reading and numeracy component skills assessment focuses on basic literacy and numeracy skills that are foundational to the more advanced skills measured in the direct assessment. The inclusion of component skills allows for a more precise estimation of literacy and numeracy at the bottom end of the skills distribution.
In the direct assessment, each respondent is only assessed in two out of the three domains (literacy, numeracy or adaptive problem solving). Respondents are first randomly allocated to one domain. Then, after they complete that assessment, they are randomly allocated to one of the two remaining domains.
The literacy and numeracy assessments use a hybrid multistage design with both an adaptive and a linear component. In both the adaptive and the linear path, respondents are administered one testlet in both Stage 2 and 3 (Stage 1 being the locator test). In the adaptive path, six different testlets are available in Stage 2, three of low difficulty and three of high difficulty. Which of these six is assigned to respondents depends on his or her performance in the locator test and personal characteristics collected in the background questionnaire (such as level of education). Stage 3 also features six testlets (two of low difficulty, two of medium difficulty and two of high difficulty). One of these six is administered to respondents based on how they performed in Stage 2 (with those that performed better having a higher chance of receiving a more difficult testlet). The linear (non-adaptive) path is used to ensure that each item is attempted by a sufficient number of respondents from a wide proficiency range. After the locator test, 25% of the respondents are randomly allocated to this path, where they take one of six possible testlets in both Stage 2 and 3.
The assessment of adaptive problem solving follows a balanced incomplete block design, where assessment items are divided into five clusters. Respondents who take the APS assessment are assigned to two randomly selected clusters of items.
A final set of questions after completion of the direct assessment asks respondents about the effort they put into completing the assessment (as compared to a high-stakes situation). Respondents are also asked how they thought they performed in the assessment.
Finally, interviewers complete a post-interview questionnaire in which they record their observations about the context, the environment and the conditions under which the interview took place.
Sampling
Copy link to SamplingTo maximise the comparability of results, countries and economies participating in the 2023 Survey of Adult Skills were expected to meet stringent standards relating to the target population, sample design, sample selection response rates and non-response bias analysis.
The target population and sampling frame
The target population for the survey consisted of the non-institutionalised population aged 16-65 years, residing in the country at the time of data collection, irrespective of nationality, citizenship or language status. The normal territorial unit covered by the survey was that of the country as a whole. However, in two countries, the sample frame covered only a portion of the national territory: in Belgium, only residents in the Flemish Region participated in the survey, and in the United Kingdom, only residents in England participated.
To draw a sample of respondents from the target population, a sampling frame is needed. Different countries and economies used different sampling frames and sampling strategies. The main distinction is whether frames are based on population registries, or whether they are based on lists of dwellings.
Population registries are based on administrative lists of residents maintained at either national or regional level, which contain useful variables for stratification, weighting and non-response bias analyses. The frames used by countries with population registry samples are shown in Table 5.3.
When population registries are not available, countries relied on a dwelling units registry or geographical clusters, which are based on lists of dwelling units or primary sampling units maintained at the national level for official surveys or a frame of geographical clusters formed by combining adjacent geographical areas, respecting their population sizes and taking into consideration travel distances for interviewers. When sampling frames are based on dwellings, interviewers need to visit the household to screen household members, randomly selecting members who are eligible to participate in the survey (because they are part of the target population). The frames used by countries and economies that relied on screeners are listed in Table 5.4.
Table 5.3. Sampling frames for countries and economies with population registry samples
Copy link to Table 5.3. Sampling frames for countries and economies with population registry samples|
OECD countries and economies |
Sampling frame |
||
|---|---|---|---|
|
Stage 1 |
Stage 2 |
||
|
Austria |
Population registry, 2022 |
||
|
Denmark |
Population registry, 2022 |
||
|
Estonia |
Population registry, 2022 |
||
|
Finland |
Statistics Finland’s Population database (based on the Central Population Register), 2022 |
||
|
France |
Centralised person registry from tax files updated yearly, 2018 (PSUs were formed in 2018 based on 2016 data) |
Person registry from taxation file, 2021 |
|
|
Germany |
List of municipalities from German Federal Bureau of Statistics, 2021 |
Population registries in selected municipalities, 2022 |
|
|
Hungary |
Register of localities from Hungarian Central Statistical Office, 2022 |
Register of persons from Ministry of Interior, 2022 |
|
|
Israel |
Big localities |
Population registry, 2022 |
|
|
Italy |
Small localities |
Population registry, 2022 |
|
|
List of municipalities from National Statistical Institute of Italy, 2022 |
Population registry, 2022 |
Population registry, 2022 |
|
|
Japan |
Resident registry, 2020 |
Local registries, 2022 |
|
|
Netherlands |
Population registry, 2022 |
||
|
Norway |
Population registry, 2022 |
||
|
Poland |
Population registry, 2022 |
Population registry, 2022 |
|
|
Sweden |
Population registry, 2022 |
||
|
Switzerland |
Population registry, 2022 |
||
|
Subnational entities |
|||
|
Flemish Region (Belgium) |
Population registry, 2022 |
||
|
Partner countries |
|||
|
Croatia |
Census of Population, Households and Dwellings, 2021 |
Population registry, 2022 |
|
|
Singapore |
Population registry, 2022 |
||
Note: The grey shading indicates that there is no such stage in the country’s sample design. PSU stands for “primary sampling unit”.
Table 5.4. Sampling frames for countries and economies using screeners
Copy link to Table 5.4. Sampling frames for countries and economies using screeners|
OECD countries and economies |
Sampling frame |
|||
|---|---|---|---|---|
|
Stage 1 |
Stage 2 |
Stage 3 |
Stage 4 |
|
|
Canada |
2021 Census |
2021 Population Census long-form returns |
Field enumeration |
|
|
Chile |
2017 National Census files, updated with more recent data whenever possible |
2017 National Census files, updated with more recent data whenever possible |
Pre-Census 2017 files and field listing for selected Secondary Sampling Units |
Field enumeration |
|
Czechia |
Register of Census Districts and Buildings (RSO), 2022 |
Register of Census Districts and Buildings (RSO), 2022 |
Field enumeration |
|
|
Ireland |
Census 2016 data, updated from Geo Directory, 2022 |
Census 2016 data, updated from Geo Directory, 2022 |
Field enumeration |
|
|
Korea |
2020 Register-based Population and Housing Census |
2020 Register-based Population and Housing Census, with updates from 2021 regional datasets and 2022 canvassing |
Field enumeration |
|
|
Latvia |
Demographic Statistics Data Processing System (using Population and Dwelling Register information), 2022 |
Demographic Statistics Data Processing System, 2022 |
Field enumeration |
|
|
Lithuania |
Address Register, 2022 |
Address Register, 2022 |
Address Register, 2022 |
Field enumeration |
|
New Zealand |
Statistics NZ’s household survey frame, 2019 |
Postal address file (PAF) and Māori Electoral Roll, 2022 |
Field enumeration |
|
|
Portugal |
National postal codes and addresses database, 2022 |
National postal codes and addresses database, 2022 |
Field enumeration |
|
|
Slovak Republic |
2021 Census |
2021 Census |
Field enumeration |
|
|
Spain* |
Population register, 2022 |
Population register, 2022 |
Field enumeration |
|
|
United States |
List of counties from US Census Bureau, 2020 |
Blocks defined by Census Bureau, 2020 |
List of addresses from the postal service, with field listing where necessary, 2022 |
Field enumeration |
|
Subnational entities |
||||
|
England (UK) |
List of MSOAs based on 2011 Census, with 2021 size measures |
Royal Mail Postcode Address File, 2022 |
Field listing |
Field enumeration |
Note: The grey shading indicates that there is no such stage in the country’s sample design. MSOA stands for “Middle layer Super Output Areas”, as defined by the UK Office for National Statistics.
* Spain is included in this table, even though it has a population registry. Spain used a population registry as the sampling frame for the first two sampling stages only and used a household screener for the third sampling stage.
Coverage of the target population
According to the TSGs, sampling frames were required to cover at least 95% of the target population. The exclusion (non-coverage) of groups in the target population had to be limited to the greatest extent possible and be based on operational or resource constraints, as in the case of populations located in remote and isolated regions. A complete list of exclusions for countries and economies using population registries is presented in Table 5.5; Table 5.6 includes a similar list for countries using screeners.
Table 5.5. Exclusions from the target population: countries and economies using population registries
Copy link to Table 5.5. Exclusions from the target population: countries and economies using population registries|
OECD countries and economies |
Percentage of target population not covered* (%) |
Groups not covered |
|---|---|---|
|
Austria |
2.0 |
Undocumented immigrants |
|
Denmark |
<=0.6 |
Undocumented immigrants, PIAAC field trial (FT) sampled persons, sampled persons from a recent national survey related to adult competencies |
|
Estonia |
1.9 |
People without a detailed address; undocumented immigrants (no estimate provided) |
|
Finland |
0.7 |
Asylum seekers, undocumented immigrants, people who have official security classification |
|
France |
2.0 |
Undocumented immigrants |
|
Germany |
0.6 |
Persons living in inaccessible areas, undocumented immigrants |
|
Hungary |
0.5 |
Individuals with no registered address; undocumented immigrants (negligible) |
|
Israel |
4.9 |
Foreign citizens with or without a permit; persons who have crossed the border illegally; Bedouin tribes and other persons living outside boundaries of localities; people who were selected to other surveys in the past three years; respondents to the Survey of Adult Skills field trial |
|
Italy |
1.3 |
Undocumented immigrants |
|
Japan |
0.1 |
Undocumented immigrants |
|
Netherlands |
2.6 |
Opt-outs, Waddeneilanden (small islands too difficult to reach), undocumented immigrants |
|
Norway |
0.5 |
Undocumented immigrants |
|
Poland |
0.3 |
Undocumented immigrants, foreigners who do not meet an obligation to register |
|
Sweden |
0.6 |
Asylum seekers, people with a residence permit valid for less than one year, undocumented immigrants |
|
Switzerland |
<4.7 |
People under guardianship, asylum seekers, diplomats, people in non-institutional collective dwelling units, undocumented immigrants |
|
Subnational entities |
||
|
Flemish Region (Belgium) |
1.0 |
Undocumented immigrants |
|
Partner countries |
||
|
Croatia |
4.9 |
Residents of remote islands and mountain areas; undocumented immigrants |
|
Singapore |
<0.1 |
Undocumented immigrants |
Note: The non-coverage rate accounts for excluded subpopulations, such as undocumented immigrants or non-institutionalised collective dwelling units, with the exception that the homeless are not being considered part of this rate. Other exclusions that occurred as a natural part of the survey process are not included in the expected non-coverage rate.
Table 5.6. Exclusions from the target population: countries and economies using screeners
Copy link to Table 5.6. Exclusions from the target population: countries and economies using screeners|
OECD countries and economies |
Percentage of target population not covered* (%) |
Groups not covered |
|---|---|---|
|
Canada |
3.3 |
Residents of reserves and other Indigenous settlements; residents of remote areas in provinces or sparsely populated regions; persons in non-institutional collective dwelling units; persons in the northern territories. |
|
Chile |
0.2 |
Areas sparsely populated and difficult to access |
|
Czechia |
2.9 |
Municipalities with less than 200 inhabitants; PSUs with less than 30 occupied dwellings |
|
Ireland |
0.2 |
Households on the islands surrounding Ireland |
|
Korea |
0.3 |
Small island residents (without land connection); redeveloping areas; natural and other disaster-affected areas |
|
Latvia |
1.5 |
Eligible dwelling units without any declared person on the frame |
|
Lithuania |
4.0 |
Residents of villages with 20 or fewer residents as age-eligible persons; residents of Neringa |
|
New Zealand |
2.0 |
People in non-private dwellings and private temporary dwellings; people living in off-shore islands and waterways (except Waiheke Island, which is included); Primary Sampling Units with less than nine occupied dwellings on 2018 Census night |
|
Portugal |
1.8 |
Group quarters, addresses in some smaller municipalities in the ultra-periphery regions of Portugal |
|
Slovak Republic |
4.4 |
Municipalities with fewer than 300 PIAAC-eligible persons |
|
Spain * |
0.5 |
Dangerous areas |
|
United States |
0.5 |
Non-locatable dwelling units |
|
Subnational entities |
||
|
England (UK) |
<0.7 |
Non-institutional collective dwelling units |
Note: The non-coverage rate accounts for excluded subpopulations, such as undocumented immigrants or non-institutionalised collective DUs, with the exception that the homeless are not being considered part of this rate. Other exclusions that occurred as a natural part of the survey process are not included in the expected non-coverage rate.
* Spain is included in this table, even though it has a population registry. Spain used a population registry as the sampling frame for the first two sampling stages only and used a household screener for the third sampling stage.
Sample size
The minimum sample size required for the 2023 Survey of Adult Skills depended on two factors: the sample design and the number of languages in which the assessment was administered. It ranged between 4 000 to 5 000 completed cases per reporting language, depending upon the sample design, as shown in Table 5.7. For a case to be counted as “completed”, the following conditions must be met: i) responses to key background questions, including age, gender, highest level of schooling, employment status and country of birth, have been collected; ii) the tablet tutorial Section has been attempted; and iii) the Locator has been attempted.
Table 5.7. Assignment of the standard minimum number of completed cases
Copy link to Table 5.7. Assignment of the standard minimum number of completed cases|
Design type |
Standard minimum number of completed cases (per reporting language) |
|---|---|
|
Unclustered – 1-stage (persons) registry with equal probabilities of selection |
4 000 |
|
Clustered – 2-stage (primary sampling units and persons), that is, an area sample with many PSUs; or 2-stage (DUs and persons) |
4 500 |
|
Clustered – 3-stage (PSUs, DUs and persons); 3-stage (PSUs, secondary sampling units (SSUs) and persons); or 4-stage (PSUs, SSUs, DUs and persons) |
5 000 |
Note: PSU stands for “primary sampling units”; SSU stands for “secondary sampling units”; DUs stands for “dwelling units”.
Countries that planned to report on general proficiency had to achieve the appropriate minimum completed sample size shown in Table 5.7 for their main language. Eight countries (Canada, Estonia, Finland, Israel, Latvia, the Slovak Republic, Spain and Switzerland) implemented the survey in multiple languages, but only Canada decided to provide separate results for the two assessment languages (English and French). Table 5.8 provides sampling information including the minimum based on the standards, and the expected and actual sample size for each participating country and economy. More detailed information about sampling and weighting can be found in the Technical Report of the 2023 Survey of Adult Skills (OECD, forthcoming[1]).
Table 5.8. Sample size information
Copy link to Table 5.8. Sample size information|
OECD countries and economies |
Standards minimum (Target population or Target population/ Country-specific samples) |
Expected number of completes (Target population or Target population/ Country-specific samples) |
Actual number of completed cases in the dataset |
Groups oversampled |
|---|---|---|---|---|
|
Austria |
4 000 |
4 500 |
4 565 |
Low-educated individuals, non-Austrians, and persons in certain interviewer regions |
|
Canada |
5000 / 10 000 |
7 000 / 10 000 |
11 697 |
An incomplete oversample of 424 individuals who were within the scope of the PIAAC target population (Indigenous population, the youth population (aged 16-30) living in Nova Scotia) which was incorporated into the main sample by matching to non-respondents in the main sample |
|
Chile |
5 000 |
5 000 / 5 100 |
4 726 |
|
|
Czechia |
5 000 |
5 061 |
5 057 |
|
|
Denmark |
4 000 |
4 925 |
5 067 |
Immigrants |
|
Estonia |
4 000 |
7 500 |
6 665 |
|
|
Finland |
4 000 |
4 200 |
4 061 |
|
|
France |
4 400 / 5 000 |
5 500 |
6 432 |
Small regions |
|
Germany |
4 500 |
5 000 |
4 793 |
|
|
Hungary |
4 500 |
4 500 |
4 564 |
Selected regions |
|
Ireland |
5 000 |
5 000 |
3 852 |
|
|
Israel |
4 060 |
6 250 / 6 280 |
6 092 |
The Arab population and Ultra-orthodox |
|
Italy |
4 500 |
7 500 |
4 847 |
Persons aged 16-29 and immigrant population (foreigners) |
|
Japan |
4 500 / 5 000 |
5 000 |
5 165 |
|
|
Korea |
5 000 |
5 000 |
6 198 |
|
|
Latvia |
5 000 |
7 692 |
6 563 |
|
|
Lithuania |
5 000 |
5 000 |
6 186 |
|
|
Netherlands |
4 000 |
4 000 / 5 267 |
3 513 |
|
|
New Zealand |
5 000 |
7 351 / 7 965 |
5 359 |
Persons of Māori and Pacific ethnicities Persons aged 16-24 years |
|
Norway |
4 000 |
4 000 |
4 053 |
|
|
Poland |
4 500 |
4 500 |
5 014 |
|
|
Portugal |
4 500 / 5000 |
5 000 |
3 160 |
|
|
Slovak Republic |
5 000 |
5 000 |
5 238 |
|
|
Spain |
5 000 |
6 000 |
5 871 |
|
|
Sweden |
4 000 |
4 000 |
3 710 |
|
|
Switzerland |
4 000 |
7 000 |
6 648 |
French-speaking and Italian-speaking language areas |
|
United States |
5 000 |
5 000 / 9 380 |
3 765 |
An incomplete oversample of small states was incorporated into the main sample through composite estimation |
|
Subnational entities |
||||
|
England (UK) |
5 000 |
5 000 |
4 941 |
|
|
Flemish Region (Belgium) |
4 000 |
4 000 |
3 909 |
|
|
Partner countries |
||||
|
Croatia |
4 500 |
4 500 |
4 316 |
|
|
Singapore |
4 000 |
4 000 |
5 011 |
Note: The PIAAC technical standard targets are for a self-weighting sample of the PIAAC target population. Additional samples are needed for country-specific samples outside of the target population or oversampled populations. Targets include multiple languages. Initial sample sizes and country targets for Austria, Canada, Chile, Denmark, France, Israel, Italy, the Netherlands, New Zealand, Switzerland and the United States include oversampled populations and country-specific samples outside the target population unless otherwise noted.
Translation and adaptation of survey instruments
Copy link to Translation and adaptation of survey instrumentsParticipating countries and economies were responsible for adapting and translating the assessment instruments, background questionnaires and survey materials for administration in their national languages. Any national adaptations of either the assessment instruments or the questionnaire were subject to strict guidelines, review and approval by the international consortium. The recommended procedure included a double translation from the English source version performed by professional staff, followed by a reconciliation.
All national versions of the instruments were subject to full linguistic quality-control procedures, which involved verification by the consortium of target versions submitted by each participating country/economy against the source versions, with reporting of residual errors and undocumented deviations and expert advice where corrective action was needed:
for questionnaire and assessment items newly developed for Cycle 2: full verification of all national materials
for trend items (used in the first cycle): focused verification of changes requested by countries and economies
for trend literacy units: verification of the scoring rules in literacy items.
Detailed information on the various aspects of translation, adaptation and verification of test and survey materials can be found in the Technical Report of the 2023 Survey of Adult Skills (OECD, forthcoming[1]).
Survey administration
Copy link to Survey administrationThe Survey of Adult Skills was administered under the supervision of trained interviewers either in the respondent’s home or in a location agreed upon between the respondent and the interviewer.
The background questionnaire, which was the first component of the interview, was administered by the interviewer. Respondents were able to seek assistance from others in the household in completing the questionnaire, for example, in translating questions and answers. Proxy respondents were not permitted.
Following completion of the background questionnaire, the respondent undertook the direct assessment on a tablet as described above. Respondents were also permitted to use calculators, notepads and a pen during the assessment. Interviewers administering the survey were required to be trained according to common standards. These covered the timing and duration of training, as well as its format and content. A full set of training materials was provided to countries. The persons responsible for organising training nationally attended training sessions organised by the international consortium.
The survey (background questionnaire plus direct assessment) was normally undertaken in a single session. However, in exceptional circumstances, a respondent could take the questionnaire in one session and the direct assessment in another. The direct assessment was required to be completed in one session. Respondents who did not complete the assessment within a single session for whatever reason were not permitted to finish it at a later time.
Data collection in the 2023 Survey of Adult Skills was scheduled from 1 September 2022 to 30 April 2023 (8 months or 242 days). Six countries completed data collection within this period (Estonia, France, Hungary, Japan, Korea and Poland). The pace of data collection in many countries was slower than projected due to unanticipated challenges such as staffing shortages from the outset, interviewer attrition or difficulties in reaching certain population subgroups (e.g. immigrants, younger respondents). As a result, the OECD and the consortium allowed some countries to extend the data collection period, with 5 countries completing it by the end of May 2023 (Austria, Germany, Lithuania, Singapore and Switzerland), 12 countries completing it by the end of June 2023 (Croatia, Denmark, England [UK], Finland, the Flemish Region [Belgium], Ireland, Latvia, Norway, Spain, the Slovak Republic, Sweden and the United States), 3 countries completing it by the end of July 2023 (Italy, Israel and Canada), and 5 countries completing it by 6 August 2023 (Chile, Czechia, the Netherlands, New Zealand and Portugal). Detailed information on the data collection timeline can be found in Chapter 19 of the Technical Report of the 2023 Survey of Adult Skills (OECD, forthcoming[1]).
Response rates and non-response bias analysis
Copy link to Response rates and non-response bias analysisIn all surveys, low response rates constitute a significant threat to the quality of the data, as they may introduce non-response bias. This happen when those who do not respond are systematically different from those who decide to participate in the survey. When this is the case, the results of the survey are no longer representative of the characteristics of the underlying target population. The 2023 Survey of Adult Skills TSGs require countries and economies to put in place a range of strategies to reduce the incidence and effects of non-response, to adjust for it when it occurs, and to evaluate the effectiveness of any weighting adjustments implemented to reduce non-response bias. In particular, countries and economies were expected to establish procedures during data collection to minimise non-response. These included pre-collection publicity, selecting high-quality interviewers, delivering training on methods to reduce and convert refusals, monitoring data collection closely to identify problem areas or groups, and directing resources to these particular groups. At least seven attempts were to be made to contact a selected individual or household before it could be classed as a non-contact. The overall rate of non-contact was to be kept below 3%.
Response rates were calculated for each stage of the assessment: 1) the screening questionnaire (for countries and economies that need to sample households before selecting respondents); 2) the background questionnaire; and 3) the direct assessment.
The overall response rate was calculated as the product of the response rates (complete cases/eligible cases) for the relevant stages of the assessment. For countries and economies with a screening questionnaire, the overall response rate was the product of the response rates for the screener, the background questionnaire and assessment; for those without a screener, it was the product of the response rates for the questionnaire and the assessment.
The computations at each stage are hierarchical, in that they depend on the response status from the previous data collection stage. A completed case thus involved completing the screener (if applicable), the background questionnaire and the direct assessment. In the case of the BQ, a completed case was defined as having responses to key background questions, including age, gender, highest level of schooling and employment status, or responses to age and gender for literacy-related non-respondents. For the direct assessment, a completed case was defined as having completed the locator stage 1 in Figure 5.1) and a literacy or numeracy module (stage 2 and 3 in Figure 5.1), or a case in which the locator was not completed for a literacy-related reason (for example because of a language difficulty or because the respondent was unable to read or write in any of a country’s test languages, or because of learning or mental disability.
Countries and economies using population registry-based sampling frames were able to treat some or all of the individuals in their samples who were untraceable as exclusions (i.e. as outside the target population) and exclude them from the numerator and denominator of the response-rate calculation (provided that the 5% threshold for exclusions was not exceeded).
The TSGs set a goal of a 70% response rate. However, countries faced challenges with declining response rates, which resulted in lower rates than encountered in the first cycle. Response and coverage rates are presented in Table 5.9.
Table 5.9. Achieved response rates and population coverage
Copy link to Table 5.9. Achieved response rates and population coverage|
OECD countries and economies |
Response rate (%) |
Coverage rate (%) |
|---|---|---|
|
Austria |
39 |
95 |
|
Canada |
28 |
97 |
|
Chile |
56 |
100 |
|
Czechia |
40 |
97 |
|
Denmark |
27 |
98 |
|
Estonia |
50 |
97 |
|
Finland |
34 |
99 |
|
France |
55 |
95 |
|
Germany |
45 |
97 |
|
Hungary |
59 |
95 |
|
Ireland |
47 |
100 |
|
Israel |
61 |
95 |
|
Italy |
29 |
95 |
|
Japan |
41 |
95 |
|
Korea |
73 |
100 |
|
Latvia |
28 |
99 |
|
Lithuania |
44 |
96 |
|
Netherlands |
40 |
95 |
|
New Zealand |
48 |
98 |
|
Norway |
41 |
99 |
|
Poland |
57 |
95 |
|
Portugal |
39 |
98 |
|
Slovak Republic |
70 |
96 |
|
Spain |
61 |
99 |
|
Sweden |
31 |
99 |
|
Switzerland |
30 |
95 |
|
United States |
28 |
100 |
|
Subnational entities |
||
|
England (UK) |
38 |
99 |
|
Flemish Region (Belgium) |
35 |
95 |
|
Partner countries |
||
|
Croatia |
36 |
95 |
|
Singapore |
62 |
99 |
Bias from non-response can arise if non-respondents are systematically different (in terms of skills proficiency, for example) from those who agree to participate in the survey. While low response rates introduce a potential source of bias, they do not necessarily mean that a bias is present. If the decision to participate in the survey is not related to skills proficiency, very low response rates will not result in any bias.
For this reason, the non-response bias analysis (NRBA) undertaken in the 2023 Survey of Adult Skills looked at a wide range of indicators, over and above response rates, to assess the extent to which results from the survey are susceptible to non-response bias. An extended NRBA considered further indicators related to the likelihood of non-response bias in the estimation of adults’ proficiency from the survey. These include:
comparison of estimates before and after weighting adjustments
comparison of weighted estimates to external totals
correlations between auxiliary variables used for weighting and proficiency estimates
comparison of estimates from alternative weighting adjustments
analysis of variables collected during data collection
level-of-effort analysis, looking at differences in proficiencies among respondents who required different number of contacts before agreeing to participate in the survey
the sensitivity of population estimates of proficiency to a range of assumptions on the proficiency of non-respondents.
Korea was the only country that achieved a response rate above 70% and was therefore not required to undertake the extended NRBA. The Slovak Republic, despite achieving a response rate of 70%, undertook the extended NRBA because it did not fully meet the sampling standards. In particular, in both Lithuania and the Slovak Republic, evidence was found that not all eligible persons in a household were given a chance of being selected to participate in the survey, which could lead to undercoverage bias. Measures were taken to reduce undercoverage bias (weight calibration). Although some additional caution should be used when analysing data from these countries, the outcomes of additional analysis, including the NRBA, suggest that the effects of this departure from the sampling standards are rather small.
When the decision to participate in the survey is linked to specific characteristics of the sampled adults, the distribution of these characteristics among participants in the Survey of Adult Skills may not match the true distribution in the population or that observed in other surveys. However, as all surveys are affected by some non-response, it is not clear that other sources are necessarily more accurate than the Survey of Adult Skills. Misalignments between the distribution of certain characteristics of adults in the sample and in the population, as well as with other sources, have been observed in both cycles of the Survey of Adult Skills. As a result, the changes in the composition of the population observed over the two cycles of the Survey of Adult Skills may also not match exactly the changes observed in other sources.
Sampling weights can be applied to bring the composition of the sample closer to the known distribution of characteristics in the population. In the Survey of Adult Skills, countries and economies agreed with the Secretariat and the OECD contractors on which variables to use for weighting. For variables that were not used in weighting, some discrepancies may remain after weighting.
Table 5.10 provides an overview of some misalignments between the (weighted) PIAAC sample and alternative data sources. Discrepancies are listed if they are statistically significant and over 1.5 percentage points. Some of these misalignments can be explained by differences in the definition or in other methodological aspects between the Survey of Adult Skills and other sources. Moreover, not all countries could verify the alignment across all characteristics due to data availability.
Alternative weighting schemes were considered to understand the extent to which the observed discrepancies might bias the estimated proficiency of the adult population in the 2023 Survey of Adult Skills. Significant differences in estimated proficiency were found in a few countries with these alternative weights (Table 5.11). However, the impact of applying alternative weighting schemes never exceeded four score points. This means that the fact that the distribution of some characteristics in the sample is not consistent with other sources does not seem to have a major impact on estimated proficiency because these alternative weighting adjustments resulted in only minor changes to the overall results.
Table 5.10. Discrepancies in the distribution of certain variables between the PIAAC weighted sample and an alternative source
Copy link to Table 5.10. Discrepancies in the distribution of certain variables between the PIAAC weighted sample and an alternative source|
OECD countries and economies |
Variables whose distribution does not match an alternative source |
Differences in coverage, timing or definition can partly explain the discrepancies |
|---|---|---|
|
OECD countries |
||
|
Austria |
- |
- |
|
Canada |
Education, Nativity, Language, Life satisfaction |
Yes |
|
Chile |
Education by gender |
No |
|
Czechia |
ISCO-08 broad skill levels (current occupation) |
Yes |
|
Denmark |
Population density |
No |
|
Estonia |
- |
- |
|
Finland |
- |
- |
|
France |
- |
- |
|
Germany |
- |
- |
|
Hungary |
Education |
Yes |
|
Israel |
Marital status by population group1, Type of locality2 |
No |
|
Italy |
- |
- |
|
Japan |
Education by gender, Education by age, Employment status by age |
Yes |
|
Korea |
Not applicable |
Not applicable |
|
Latvia |
Employment status |
Yes |
|
Lithuania |
Education, Employment status, Gender |
Yes |
|
Netherlands |
Income, Socio-economic status, Wealth |
No |
|
New Zealand |
Employment status, Ethnicity |
Yes |
|
Norway |
Employment status, Industry, Household composition |
No |
|
Poland |
Education by age |
Yes |
|
Portugal |
Region by age |
Yes |
|
Slovak Republic |
Education by region |
Yes |
|
Spain |
- |
- |
|
Sweden |
Education, Employment status, Occupation, Economic activity, Nativity |
Yes |
|
Switzerland |
- |
- |
|
United States |
Education, Employment status, Household composition, Race/ethnicity, Health insurance coverage |
Yes |
|
Subnational entities |
||
|
England (UK) |
Employment status by age |
Yes |
|
Flemish Region (Belgium) |
Employment status by age |
Yes |
|
Partner countries |
||
|
Croatia |
- |
- |
|
Singapore |
Nativity |
Yes |
1: The population groups are defined as follows: Jews, Not Ultra-Orthodox; Jews, Ultra-Orthodox; Arab Population.
2: Localities are classified as follow: Jewish/Arab by population density.
Table 5.11. Notable and significant differences in estimated proficiency from alternative weighting schemes
Copy link to Table 5.11. Notable and significant differences in estimated proficiency from alternative weighting schemes|
Countries and economies |
Score differences |
Variables used for reweighting |
||
|---|---|---|---|---|
|
Literacy |
Numeracy |
Adaptive problem solving |
||
|
Estonia |
2.54 |
2.58 |
1.82 |
County (5) * Education (3) |
|
Lithuania |
-3.55 |
-3.73 |
-2.34 |
Education (5) * Labour force status (4) |
|
Latvia |
2.86 |
2.91 |
2.23 |
Gender (2) * Education (7) |
|
Netherlands |
1.10 |
1.41 |
0.98 |
Socio-economic status |
|
Poland |
-1.88 |
-2.16 |
-1.09 |
Education (4) |
|
Singapore |
2.18 |
2.17 |
1.43 |
Gender (2) * Education (5) |
|
Spain |
-2.45 |
-2.46 |
-1.95 |
Country of birth (2) * Region (18) |
Note: The table reports the difference between the average proficiency estimated using the final PIAAC weights and the proficiency obtained using alternative weighting schemes. The table only reports countries and economies for which differences are statistically significant and notable (absolute value of the difference larger than the standard error of the differences). In a few other countries, statistically significant differences are found, but they are almost all smaller than 1 score point. These are not reported to save space but can be found in the Survey of Adult Skills 2023 Technical Report (OECD, forthcoming[1]). The number in parenthesis next to the variables used for reweighting indicates the number of categories for each variable.
Following the extended NRBA, countries and economies were classified into different categories, reflecting a holistic assessment of how susceptible their proficiency estimates were to non-response bias, and the corresponding level of caution that is advised when interpreting the results:
pass: meaning that the analysis provided no strong evidence of non-response bias
low caution: meaning that some caution should be taken in interpreting the results, as non-response bias may be present
medium caution: meaning that proficiency estimates are more susceptible to non-response bias and more caution should be exerted
high caution: meaning that the likelihood of non-response bias is higher.
Details on the indicators included in the analysis and on the criteria followed in classifying countries and economies can be found in the Technical Report of the 2023 Survey of Adult Skills (OECD, forthcoming[1]).
Results of the NRBA are presented in Table 5.12. In interpreting the outcomes of the NRBA and the resulting classification of countries, readers should note that:
The classification reflects an assessment of the likely existence of non-response bias, and not of its magnitude. In other words, one cannot conclude that countries in the “high caution” category necessarily have a larger non-response bias than countries in the “low caution” category.
The criteria and the thresholds used for the classification have been approved by consensus by the independent experts who are member of the Technical Advisory Group.
In any such classification, threshold effects will always exist, meaning that countries might be classified in different categories even though the differences in their underlying indicator are very small.
This classification reflects a judgement on the collected data, and not on the quality of the work done by national centres and data collection agencies, which all countries completed satisfactorily and in accordance with the requirements specified in the technical standards and guidelines.
Table 5.12. Outcomes of the non-response bias analysis
Copy link to Table 5.12. Outcomes of the non-response bias analysis|
Countries and economies |
Response rate (%) |
Outcome of the NRBA analysis |
|---|---|---|
|
Korea |
73 |
Not applicable |
|
Slovak Republic |
70 |
Pass |
|
Singapore |
62 |
Pass |
|
Israel |
61 |
Pass |
|
Spain |
61 |
Pass |
|
Estonia |
50 |
Pass |
|
France |
55 |
Pass |
|
Hungary |
59 |
Low caution |
|
Poland |
57 |
Low caution |
|
Chile |
56 |
Low caution |
|
New Zealand |
48 |
Low caution |
|
Ireland |
47 |
Low caution |
|
Germany |
45 |
Low caution |
|
Lithuania |
44 |
Low caution |
|
Norway |
41 |
Low caution |
|
Czechia |
40 |
Low caution |
|
Austria |
39 |
Low caution |
|
Finland |
34 |
Low caution |
|
Sweden |
31 |
Low caution |
|
Denmark |
27 |
Low caution |
|
Japan |
41 |
Medium caution |
|
Netherlands |
40 |
Medium caution |
|
Portugal |
39 |
Medium caution |
|
England (UK) |
38 |
Medium caution |
|
Croatia |
36 |
Medium caution |
|
Flemish Region (Belgium) |
35 |
Medium caution |
|
Switzerland |
30 |
Medium caution |
|
United States |
28 |
Medium caution |
|
Canada |
28 |
Medium caution |
|
Italy |
29 |
High caution |
|
Latvia |
28 |
High caution |
Note: The extended NRBA was not required for countries with response rates above 70%. For this reason, results of the NRBA are not applicable for Korea. The extended NRBA was conducted for the Slovak Republic, despite a response rate of 70%, because the country did not fully meet the sampling standards.
The doorstep interview and literacy-related non-response
Copy link to The doorstep interview and literacy-related non-responseIn most participating countries and economies, a proportion of respondents are unable to undertake the assessment for literacy-related reasons, such as being unable to speak or read the test language(s), having difficulty reading or writing, or having a learning or mental disability. Some of these respondents may be able to complete the background questionnaire or key parts of it, presumably with the assistance of an interviewer who spoke the respondent’s language, a family member or another person. This form of non-response could introduce bias since it is systematically concentrated among those with low literacy proficiency in the survey language (presumably migrants or people with very poor reading skills). In the first cycle, the share of such non-respondents amounted to less than 2% in most countries but exceeded 4% in four countries and economies (OECD, 2019[2]).
To reduce the bias induced by such literacy-related non-response, the 2023 Survey of Adult Skills introduced a new instrument called the doorstep interview. This is a short questionnaire offered in 43 languages which collects basic background information: gender, age, years of schooling, employment status and country of origin. This questionnaire can be easily completed by individuals who do not speak the language(s) of the assessment and are, therefore, unable to answer the regular background questionnaire and the direct skills assessment. The information collected through the doorstep interview was used to estimate the literacy and numeracy proficiency of these non-respondents. This innovation allowed the survey results to cover the entire target population.
The share of doorstep interview cases in each country and economy are shown in Table 5.13.
Table 5.13. Doorstep interview cases across participating countries and economies
Copy link to Table 5.13. Doorstep interview cases across participating countries and economies|
OECD countries and economies |
Number of cases |
Share among all interviews (% - unweighted) |
Share of the represented population (% - weighted) |
Share among foreign-born adults (% - weighted) |
|---|---|---|---|---|
|
Austria |
182 |
4.0 |
2.8 |
9.9 |
|
Canada |
90 |
0.8 |
2.5 |
7.3 |
|
Chile |
2 |
0.0 |
0.0 |
0.4 |
|
Czechia |
72 |
1.4 |
4.5 |
52.8 |
|
Denmark |
887 |
17.5 |
5.3 |
29.5 |
|
Estonia |
110 |
1.7 |
1.2 |
10.4 |
|
Finland |
143 |
3.5 |
3.7 |
32.4 |
|
France |
36 |
0.6 |
0.6 |
3.8 |
|
Germany |
118 |
2.5 |
2.7 |
11.1 |
|
Hungary |
48 |
1.1 |
1.0 |
20.0 |
|
Ireland |
11 |
0.3 |
0.6 |
1.9 |
|
Israel |
20 |
0.3 |
0.8 |
4.2 |
|
Italy |
32 |
0.7 |
0.2 |
1.7 |
|
Japan |
43 |
0.8 |
1.1 |
43.1 |
|
Korea |
35 |
0.6 |
0.6 |
12.9 |
|
Latvia |
4 |
0.1 |
0.4 |
4.0 |
|
Lithuania |
88 |
1.4 |
2.6 |
40.7 |
|
Netherlands |
97 |
2.8 |
3.1 |
15.3 |
|
New Zealand |
2 |
0.0 |
0.1 |
0.1 |
|
Norway |
64 |
1.6 |
1.5 |
6.2 |
|
Poland |
1 |
0.0 |
0.1 |
16.8 |
|
Portugal |
57 |
1.8 |
3.0 |
11.6 |
|
Slovak Republic |
8 |
0.2 |
0.4 |
17.5 |
|
Spain |
84 |
1.4 |
2.1 |
9.6 |
|
Sweden |
0 |
0.0 |
0.0 |
- |
|
Switzerland |
217 |
3.3 |
3.3 |
9.0 |
|
United States |
7 |
0.2 |
0.9 |
4.4 |
|
Subnational entities |
||||
|
England (UK) |
5 |
0.1 |
0.4 |
1.4 |
|
Flemish Region (Belgium) |
209 |
5.4 |
3.1 |
16.2 |
|
Partner countries |
||||
|
Croatia |
22 |
0.5 |
0.3 |
2.8 |
|
Singapore |
0 |
0.0 |
0.0 |
- |
Note: The high number of unweighted cases in Denmark is due to the decision to oversample the immigrant population.
While clearly an improvement with respect to the first cycle of the Survey of Adult Skills, the introduction of the doorstep interview poses a challenge for comparing estimates over time, as the sampled populations are no longer fully comparable: adults who completed the doorstep interview in the second cycle would have been handled as literacy-related non-respondents in the first cycle. The OECD recommends excluding cases who only completed the doorstep interview in all analysis comparing results from the second and first cycles of the Survey of Adult Skills.
Overall assessment of data quality
Copy link to Overall assessment of data qualityThe data from participating countries and economies were subject to a process of adjudication to determine whether they were of sufficient quality to be reported and released to the public. Data adjudication can be seen as the culmination of the quality assurance and quality-control arrangements put in place to ensure that the survey produces reliable and valid data. These had included establishing the Technical Standards and Guidelines (TSGs) covering all aspects of the implementation of the survey and collecting information to monitor compliance with them, as well as identifying problems as they emerged and recommending corrective action. The international consortium also provided guidance, training, assistance and tools to help countries comply with the requirements of the TSGs.
The adjudication process used a broad definition of quality: “fitness for use”. Although countries’ compliance with the TSG was an important component of the quality assessment, the goal was to go beyond compliance to assess whether the data produced were of sufficient quality in terms of their intended uses or applications.
In assessing overall data quality, the focus was on four key areas:
sampling
survey operations and interviewer training standards
instrumentation
data output and the operation of the delivery platform.
In each of these areas, countries and economies were assessed against a set of quality indicators which reflected the major requirements of the TSGs. All countries and economies either fully met the required quality standards or met them to a degree that was believed not to compromise the overall quality of the data. The data from all participating countries and economies were determined to have met the quality standards required for reporting and public release. The project’s Technical Advisory Group reviewed the assessments of the quality of all national data before submitting them to the Board of Participating Countries.
In some countries, there were some specific concerns because unusual response patterns were identified, suggesting that some respondents may not have exerted a reasonable level of effort in answering the literacy, numeracy and adaptive problem solving assessment. This may call into question whether their responses accurately reflect their proficiency. The OECD Secretariat therefore conducted additional quality checks on the data, expanding on the quality-control procedures set in the Technical Standards and Guidelines. To identify such cases, the OECD relied on the following criteria: a very short time spent on the assessment, a high share of very rapid responses, a high share of missing answers, and locator failure (i.e. failure to answer a set of easy questions) from highly educated, native-born respondents. Anomalies were mostly found in the responses to the cognitive assessment; the pattern of responses to the background questionnaire did not raise particular concerns.
Disengaged respondents will always exist in surveys, and it is difficult to establish objective criteria to assess whether a reasonable level of effort was exerted. In some countries (Israel, Lithuania, New Zealand, Poland, the Slovak Republic, and Spain), it was found that many of these respondents were clustered around a few interviewers, suggesting that the problem may stem from such interviewers not following the PIAAC protocols. In particular, interviewers were identified for which a high share of their cases met at least two of the criteria mentioned above. This cast doubts on the quality of all data collected by such interviewers.
In Lithuania, New Zealand, the Slovak Republic and Spain, all cases from the identified interviewers (406 in Lithuania, 301 in New Zealand, 356 in the Slovak Republic, and 385 in Spain) have been excluded from the data used to estimate the population model, which establishes the relationship between the variables from the background questionnaire and performance on the direct assessment to generate proficiency estimates (plausible values; see the Survey of Adult Skills 2023 Technical Report (OECD, forthcoming[1]) for more detail). This exclusion enhances the robustness of the model, by ensuring it is estimated based only on the cases considered to be of sufficient quality. In the absence of definitive evidence of data falsification or other forms of interviewer misconduct, the responses to the cognitive assessment items still contributed to the estimation of plausible values for these and all respondents.
In Israel, six interviewers were identified as having a relatively large share of cases with unusual response patterns, using the same criteria that led to the identification of cases in the other countries. All data from these interviewers (748 cases in total) have been excluded from the data used to estimate the population model, which establishes the relationship between the variables from the background questionnaire and performance on the direct assessment to generate proficiency estimates (OECD, forthcoming[1]). This exclusion enhances the robustness of the model, by ensuring it is estimated based only on the cases considered to be of sufficient quality.
Moreover, stronger evidence was collected that three of these interviewers breached data collection protocols throughout the survey or were implausibly productive (conducting a very large number of interviews in a relatively short period). As this raised additional concerns about the quality of data, the responses to the cognitive assessment items for all cases of these interviewers were excluded from the database (572 in total). Plausible values for these cases were then estimated using only their responses to the background questionnaire (for which no unusual patterns were detected) and the parameters estimated by the population model.
In Poland, nine interviewers were identified as having a relatively large share of cases with unusual response patterns of respondents, using the same criteria that led to the identification of cases in the other countries. All data from these interviewers (774 cases in total) have been excluded from the data used to estimate the population model, which establishes the relationship between the variables from the background questionnaire and performance on the direct assessment to generate proficiency estimates (OECD, forthcoming[1]). This exclusion enhances the robustness of the model, by ensuring it is estimated based only on the cases considered to be of sufficient quality.
Moreover, stronger evidence was collected that six of these interviewers in Poland breached data collection protocols throughout the survey. For instance, some of these interviewers were implausibly productive, conducting many interviews on a single day. Others did not record interviews or obtain respondents’ phone numbers, which made validation of interviews more difficult. Yet another interviewer was found to have falsified seven cases during data collection (cases which were immediately removed from the dataset as part of the quality-control process and are not included in the 774 cases under consideration in this note). Twenty-seven other cases collected from this interviewer were, however, validated and remained in the dataset. Since these factors raise concerns about the quality of all cases completed by these six interviewers, the responses to the cognitive assessment items for all cases of these six interviewers were excluded from the database (559 in total). Plausible values for these cases were then estimated using only their responses to the background questionnaire (for which no unusual patterns were detected) and the parameters estimated by the population model.
In Poland, other cases with unusual response patterns that could suggest possible disengagement or lack of a reasonable level of effort during the assessment were identified. As these cases were not clustered within any particular interviewer, they were left in the dataset and treated as all other cases, given the difficulty of establishing objective criteria to determine whether reasonable effort was exerted, and whether the results of the assessment truly reflect the proficiency of respondents. While similar cases are present in all countries, the number of such cases in Poland can potentially have a significant impact on the estimated proficiency of the overall population. This should be kept in mind when interpreting Poland’s results. For this reason, in OECD (2024[3]) results for Poland are flagged with an asterisk.
References
[3] OECD (2024), Do Adults Have the Skills They Need to Thrive in a Changing World?: Survey of Adult Skills 2023, OECD Skills Studies, OECD Publishing, Paris, https://doi.org/10.1787/b263dc5d-en.
[2] OECD (2019), The Survey of Adult Skills: Reader’s Companion, Third Edition, OECD Skills Studies, OECD Publishing, Paris, https://doi.org/10.1787/f70238c7-en.
[1] OECD (forthcoming), Survey of Adult Skills 2023 Technical Report, OECD Publishing, Paris.
Note
Copy link to Note← 1. The Technical Standards and Guidelines are available at https://www.oecd.org/en/about/programmes/piaac/piaac-data.html#manuals