Crawling: This is the first step in web scraping, it starts at the source of the data, a website or webpage and then scans the websites for links to other useful data that match the ought after content.
Scraping: As the name implies, this process involves the actual collection of data from the websites by the crawler, in this process, the specific data gotten from the website is copied out on a separate platform.
Extracting: This process involves sorting through the scrapped data and extracting meaningful information. The extractor could either be extracting names, phone numbers, prices, job descriptions, image information or video details, etc.
Formatting: Once the data has been extracted, it is then fed into a user application in a bid to reach the final user. Some of the common formats used in presenting this data are JSON, CSV, XML, etc.