Neidio i’r prif gynnwys
English Cymraeg
Food and You 2: Wave 6 Technical report

FY2 Wave 6 Technical report: Appendices

All appendices for the Food and You 2 Wave 6 report.

Diweddarwyd ddiwethaf: 26 July 2023
Diweddarwyd ddiwethaf: 26 July 2023

List of appendices

The online questionnaire has been included as an Appendix to the technical report but is available as a separate PDF and a table of methodological differences between waves has been presented overleaf. 

The rest of the documentation (listed below) will be uploaded onto the UK Data Archive:

Food and You 2 Wave 6 online questionnaire
Food and You 2 Wave 6 postal questionnaires

  • Eating at Home (England)
  • Eating at Home (Northern Ireland)
  • Eating at Home Wales (English)
  • Eating at Home Wales (Welsh)
  • Eating Out (England)
  • Eating Out (Northern Ireland)
  • Eating Out Wales (English)
  • Eating Out Wales (Welsh)

Food and You 2 Wave6 invitation and reminder letters

  • Invitation letters (sent to main sample addresses)
  • First reminder letters (sent to main sample addresses)
  • Second reminder letter (sent to main sample addresses only)
  • Final reminder letter (sent to main sample addresses only)

Food and You 2 Wave 6 full SPSS data
Food and You 2 Wave 6 SPSS user guide
Food and You 2 Wave 6 full data tables (and user guide) for England, Wales and Northern Ireland combined 
Food and You 2 Wave 6 individual country data tables (and user guide) for England, Wales and Northern Ireland

Methodological differences between Waves 1 to 6

Table 21 compares the differences between Waves 1-6 in sample size and household participation rates. The largest sample size was in Wave 1 (21,053), dropping to 13,922 in Wave 2 and remaining at 14,115 for Waves 3 and 4. For Wave 5, the inclusion of the reserve addresses meant that the sample size was larger than in Waves 2-4, with 16,115 addresses. Reserve addresses were not used for Wave 6 but the main sample size for Wave 6 was increased slightly to 14,500 compared to 14,115 issued in Waves 3 and 4. As up to two adults per household can participate the number of individual returns overall is always higher than the number of households participating. The number of individual returns overall was 9,319 (Wave 1), 5,900 (Wave 2), 6,271 (Wave 3), 5,796 (Wave 4), 6,770 (Wave 5), and 5,991 (Wave 6).  The number of household returns overall is show in Table 19. The highest overall response rate was achieved in Wave 3: 30.7%.

Table 19: Comparing differences between Wave 1 to 6 in sample size and participation rates

Wave Sample size (issued addresses) Number of household returns overall Overall response rate
1 21,053 6,408 30.4%
2 13,922 3,955 28.4%
3 14,115 4,338 30.7%
4 14,115 4,026 28.5%
5 16,115 4,727 29.3%
6 14,500 4,217 29.1%

The average length of the online questionnaire was different in each wave:

  • Wave 1 was 29 minutes and 58 seconds
  • Wave 2 was 36 minutes and 27 seconds
  • Wave 3 was 30 minutes and 13 seconds
  • Wave 4 was 37 minutes and 14 seconds
  • Wave 5 was 27 minutes and 08 seconds
  • Wave 6 was 25 minutes and 46 seconds 

To maximise response and reduce participant burden the recommended survey length for the online questionnaire is no more than 30 minutes. Waves 1, 5, and 6 were at or within the target 30-minute average length.

Table 20 shows the number of versions of the postal questionnaire used, the fieldwork period and the proportion of the outstanding sample who were sent the final reminder. In all waves other than Wave 3, two versions of the postal questionnaire was used. In Wave 3, the response rate was high enough after the second reminder for the final reminder to be sent to just two-thirds of the non-responding sample. In all other waves the final reminder was sent to all outstanding non-responding households. 

Table 20: Comparing differences between Wave 1 to 6, for postal questionnaire versions, fieldwork period and mailing sample

Wave Number of different versions of postal questionnaires for each country Fieldwork period Proportion of available sample sent final mailing
1 Two (Version A and Version B) 29 to 6 October 2020 (about 10 weeks)

100% of non-responding households in Wales and Northern Ireland. 

50% of non-responding households in England

2 Two ('Eating Out' and 'Eating at Home') 20 November 2020 to 21 January 2021 (about nine weeks) 100% of non-responding households across the sample
3 One (with different questions in Northern Ireland than for England and Wales) 28 April to 25 June 2021 (about 8 weeks) 66.6% of non-responding households across the sample
4 Two ('Eating Out' and 'Eating at Home') 18 October 2021 and 10 January 2022 (about 12 weeks) 100% of non-responding households across the sample
5 Two (Version A and Version B) 26 April 2022 and 24 July 2022 (about 13 weeks) 100% of non-responding households across the main sample, 0% of households from the reserve sample
6 Two ('Eating Out' and 'Eating at Home') 12 October 2022 to 10 January 2023 (about 13 weeks) 100% of non-responding households across the sample

Questionnaire development

In Wave 1 a prolonged period of questionnaire development took place which involved an extensive review of questions from previous FSA surveys (Food and You and Public Attitudes Tracker). After all relevant questions were compiled a workshop with the Food and You 2 advisory group was held to discuss key priorities for the questionnaire. This was followed by a second workshop with key internal stakeholders to discuss their priorities for the questionnaire and provide Ipsos with direction regarding questionnaire content. 

Following this, draft questionnaire modules were compiled based on questions from previous FSA surveys. Numerous alterations to the wording, ordering, format and content of the questions were made in the process based on survey design best practice, with additional questions designed based on stakeholder needs. The questionnaire development stages for Waves 2-6 were much shorter as core questions and materials had been developed in Wave 1.  

Cognitive testing 

Ahead of Waves 1-4 and 6, cognitive testing was conducted to examine participant comprehension of new or potentially challenging questions. Participants for cognitive testing were recruited from Ipsos MORI’s iOmnibus recontact database and for Wave 2 onwards, via an external Ipsos approved supplier. In Wave 1, 26 cognitive interviews were completed. In Wave 2, 14 interviews were completed, and 20 interviews were conducted when developing the questionnaire for Waves 3-4.  Cognitive testing was not used for Wave 5 as only one straightforward question was added. In Wave 6, 19 interviews were completed.

Usability testing

Prior to Wave 1 fieldwork, usability testing was also undertaken to identify areas where improvements could be made in the form and format of the questions on the online survey across different commonly used devices (for example, mobile phone, tablet, computer). Interviews were conducted over online video conferencing software, with interviewers observing participants journey through the online questionnaire (using screen share technology) and asking questions where relevant. Eleven interviews were undertaken at this stage. This helped identify formatting and layout issues with the online questionnaire which were amended ahead of the pilot survey. Usability testing was not conducted again ahead of Waves 2-6 as the online questionnaire took the same format as the Wave 1 questionnaire.

Pilot

Prior to the main stage fieldwork for Wave 1 a pilot was conducted on the full questionnaire to understand the time it took for participants to complete the questionnaire and each individual module within it. The questionnaire was tested over four days with 390 members of Ipsos MORI’s online access panel. The questionnaire took participants on average 26 mins and 48 seconds to complete and it was believed that no alterations were needed to the length of the questionnaire, in order for it to fall within the desired 30 minutes. Pilots were not conducted in Waves 2-6 as the expected completion time was estimated from earlier waves. 

Differences in the questionnaire

Due to the modular design of Food and You 2, some questions (core modules) are asked in every wave, whereas other questions are only present in certain waves. For some questions, the base will vary between waves. This is due to changes in the questions available for filtering, and/or their inclusion in the postal questionnaire. Please see the Wave 6 Tables User Guide for details.

The table below notes which modules were present in each wave of the survey, though note the content of each module varied somewhat between waves, as outlined above. 

Table 21: Questionnaire module content of each survey wave

Full list of modules from Wave 1 to 6 Waves included
About you and your household 1, 2, 3, 4, 5, 6
Food Concerns (core) 1, 2, 3, 4, 5, 6
Food you can trust (core) 1, 2, 3, 4, 5, 6
Household Food Security (Core) 1, 2, 3, 4, 5, 6
Eating at Home (core questions) 2, 4, 6
Eating at Home (full module) 1,5
Food Shopping 1, 3, 5
Defra Questions 1, 3, 5
Eating Out 2, 4, 6
Online Food Platforms 3, 5
Food Hypersensitivities (core questions) 1, 3, 4, 5
Food Hypersensitivities (full module) 2, 6
Healthy Eating (Northern Ireland only) 3
Emerging Issues 4

Differences in fieldwork

Fieldwork dates

The Food and You 2 survey should take place every six months. However, the length of the initial questionnaire development led to a later start in its first year. The fieldwork dates of each wave are as follows:

  • Wave 1: 29th July 2020 to 6th October 2020 (about 10 weeks)
  • Wave 2: 20th November 2020 to 21st January 2021 (about nine weeks)
  • Wave 3: 28th April to 25th June 2021 (about eight weeks)
  • Wave 4: 18th October 2021 and 10th January 2022 (about 12 weeks)
  • Wave 5: 26th April 2022 and 24th July 2022 (about 13 weeks)
  • Wave 6: 12th October 2022 - 10th January 2023 (about 13 weeks)

Sample sizes 

There were just over 21,000 addresses issued in Wave 1, leading to 9,319 individual returns. Since this was much higher than the target of 5,600 individual returns, only 13,922 addresses were issued in Wave 2, and 14,115 addresses issued in Waves 3 and 4. For Wave 5, 16,115 addresses were issued (14,115 from the main sample and 2,000 from the reserve sample). 14,500 addresses were issued for Wave 6. 

Vouchers 

As an experiment, each adult who completed the questionnaire in Wave 1 received either a £15 online voucher, £10 online or paper voucher and £5 online or paper voucher. Based on the results, respondents in Waves 2-6 received only the £10 voucher. The experiment process and results were summarised in an article published on the Social Research Association (SRA) website, Volume 11, Summer 2021

Postal questionnaires

When postal questionnaires were sent out in Waves 1, 2 and 4-6, the version was assigned to person one and person two in the household on a quasi-random basis. This meant half contained questions from one module and the rest contained questions from another module. However, in Wave 3, one of the modules was only relevant to residents of Northern Ireland. Therefore, the content of the postal questions varied on a country basis rather than randomly.

Reminders

In Wave 1, the final reminder was sent to all outstanding non-responding households in Wales and Northern Ireland, and to 50% of those in England. In Wave 3, the response rate was high enough after Mailing 3 for the final reminder to be sent to just two-thirds of the non-responding sample. In Waves 2, 4, 5, and 6, all non-responding sample received a final reminder. In Wave 5, main sample cases received up to three reminders whereas reserve sample cases received up to two – the final reminder was not issued to reserve sample addresses due to an increase in response.
 

Differences in weighting

Overall, the same weighting approach was taken in Waves 1-6. However, in each wave, some additional weights are needed for those questions that are not asked to all postal respondents. These additional weights will vary between waves depending on which questions are included.

Waves 4-6 have “Welsh and Welsh-England” weights to easily compare Welsh respondents against an English population calibrated to have similar demographic characteristics. In Waves 1 to 3, the weights for the calibrated English population were called “Welsh England” weights. The corresponding weights for Welsh respondents were formally part of the individual country level weights.

Differences in data validation and management

In Waves 1 and 2, the tables were created from the underlying data independently of the SPSS dataset. From Wave 3 onwards, syntax produced the derived variables in SPSS, and this was used to produce the tables in Quantum. As part of this change, the data validation procedures were reviewed and the following improvements made: 

  • in all waves, back editing and forwarding editing was applied to inconsistencies in the postal data, with a smaller amount of back editing applied to the data for Waves 3-6 than in other waves. Back editing meant that if a filtered question was answered but the filter origin question contradicted that answer (blank or different), then the origin question was changed to be the answer for the filter question. Whereas forward edited meant that if a participant answered a question but did not follow the routing to answer the next filtered question they were assigned a code of -99 “Not stated”.
  • in Waves 1 and 2, if a question was incorrectly answered as a multi-code question when only one answer should have been selected, then a digit from the participant ID was used to randomly select an answer. In Waves 3-6, the responses were set to -99 “Not stated”.
  • from Wave 3 onwards, an edit was introduced to correct the number of adults when participants from a multiple response household answered that only one adult lived in that household.

Ipsos standards and accreditations

Ipsos’s standards and accreditations provide our clients with the peace of mind that they can always depend on us to deliver reliable, sustainable findings. Moreover, our focus on quality and continuous improvement means we have embedded a ‘right first time’ approach throughout our organisation.

Market Research ISO 20252 Certificate

This is the international market research specific standard that supersedes BS 7911/MRQSA and incorporates IQCS (Interviewer Quality Control Scheme). It covers the five stages of a Market Research project. Ipsos UK was the first company in the world to gain this accreditation. 

Information Security ISO 27001 Certificate

This is the international standard for information security designed to ensure the selection of adequate and proportionate security controls. Ipsos UK was the first research company in the UK to be awarded this in August 2008.

Company Quality ISO 9001 Certificate

This is the international general company standard with a focus on continual improvement through quality management systems. In 1994, we became one of the early adopters of the ISO 9001 business standard.

Market Research Society (MRS) Company Partners Certificate

By being an MRS Company Partner, Ipsos endorse and support the core MRS brand values of professionalism, research excellence and business effectiveness, and commits to comply with the MRS Code of Conduct throughout the organisation. Ipsos were the first company to sign our organisation up to the requirements & self regulation of the MRS Code; more than 350 companies have followed our lead.
The UK General Data Protection Regulation (UK GDPR) & the UK Data protection Act 2018 (DPA) 
Ipsos UK is required to comply with the UK General Data Protection Regulation and the UK Data Protection Act; it covers the processing of personal data and the protection of privacy.

Cyber Essentials Certificate

A government backed and key deliverable of the UK’s National Cyber Security Programme. Ipsos UK was first assessment validated for certification in 2016 and this is renewed annually. Cyber Essentials defines a set of controls which, when properly implemented, provide organisations with basic protection from the most prevalent forms of threat coming from the internet. 

Fair Data 

Ipsos UK is signed up as a ‘Fair Data’ Company by agreeing to adhere to ten core principles. The principles support and complement other standards such as ISOs, and the requirements of Data Protection legislation.