Cost-of-living crisis: ONS to explore web scraping data to track supermarket prices

Written by Sam Trendall on 4 May 2022 in News

Agency’s dedicated data-science facility looks to make a big contribution to government’s response to rising costs

Credit: Jernej Furman/CC BY 2.0 

The Office for National Statistics is to explore the use of web scraping technology to keep tabs on the cost of goods at the UK’s major supermarkets.

The statistical agency has previously experimented with using the tech, in which automated systems trawl the web and extract data from sites. From 2014 to 2017, as part of a wider exploration of the big data sources, the ONS trialled the use of web scraping to collect information on grocery pricing. 

This work largely predated the creation in spring 2017 of the agency’s dedicated Data Science Campus, the remit of which is to explore the use of potential new and innovative information sources. In its work to date, the campus has used data on the quarterly VAT returns and the movement of ships in and out of the UK ports to provide indicators of the country’s economic performance.

As it moves into its second half decade, it has ambitions to use data science to support some of government’s biggest policy objectives and its response to society’s biggest challenges.

Related content

Acting director Arthur Turrell told PublicTechnology: “I want us to make a major contribution to the cost-of-living crisis and levelling up.”

This will partly involve expanding the range of data gathered on the shipping industry, in which worldwide GPS information is used to track the movement of ships. During the pandemic, this helped to identify supply-chain issues, according to Turrell.

“We knew that the ships that were bringing goods into the country were backed up outside ports,” he said.

The ambition is now to derive information from bills of lading – the documents issued by shipping firms to importers and containing information on the cargo being carried.

Turrell says that this could enable better understanding of levels of supply of goods, and help spot patterns and possible shortages.

“We know where the ships are, but what we would like to know is what is on them… and what does it mean if a ship is, for example, 70% trousers and 30% iPhones? And how many units does this equate to?,” he adds.

At the other end of the supply chain, the campus also intends to explore the potential of web scraping to help gather information on consumer prices.

“ONS has a bit of a history of using web-scraped data,” Turrell said. “We are looking at the prices in supermarkets and whether we can use web scraping.”

Keep an eye out in the coming days for PublicTechnology’s full interview with the Data Science Campus chief, including lots of insights into how the unit supported the government’s response to the pandemic, and its ambitions for furthering the use of data in the months and years ahead.

About the author

Sam Trendall is editor of PublicTechnology. He can be reached on

Share this page




Please login to post a comment or register for a free account.

Related Articles

HMRC lifts lid on £200m programme to achieve data-protection compliance
27 September 2022

Recently released information provides details of three-year project to minimise risk and improve use of data

IPO seeks new chief exec to lead digital and data transformation
22 September 2022

BEIS-sponsored agency floats £100k leadership post

ONS signs £15m partner for consumer price index
14 September 2022

Data analytics firm will help statisticians select products for ‘basket’ of goods

Cabinet Office guest speakers to be subject to social-media vetting
16 August 2022

Officials will reportedly be asked to look over five years of posts to check there is no ‘potentially problematic material’

Related Sponsored Articles

Rewiring government: improving outcome management
6 September 2022

Paul Pick-Aluas, Strategy & Transformation, Public Sector at Salesforce, explains how governments can use technology innovation to improve how it can analyse outcomes