Chrome has a plugin called WebScraper that makes it easier to scrape websites and to bypass any coding. Follow the instructions on the website to download the plug in and it might be good to watch the intro video on the same page. This will allow you to get a feel for the plug in and how it works.
Once the plug in is installed, if you right click anywhere on the page (doesn’t have to be the one you want to scrape) and click on “Inspect (Element)”. There should be a tab at the top of the inspector that is called WebScraper. This is the plug in! Next click on Create new sitemap > Import Sitemap, then you can paste the sitemap the DigiLab has created, or any other sitemap that has already been created here. This is the place where the digilab will upload its sitemaps for public use. They will be named based on the website they are scraping.
The last step before we scrape is you need to change the URL in the sitemap before you import it. So after you paste the sitemap into the webscraper, change what is inside “startUrl”:[“PUT URL IN HERE BETWEEN QUOTATION MARKS”]. Don’t change anything else to the sitemap besides the URL itself. Once this is imported, you can click on Sitemaps [name of sitemap] at the top of the inspector, and click Scrape. You will be asked to put request intervals, 2000 is plenty of time for both request and page load. This is just to show the browser that you are a human and not a robot.
Let the plug in do it’s thing, and you might have to click on the refresh data button once it is done. Then you can export the table as a csv and now you have your data!