You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Before we get started, this tutorial was run on Python 3.10.10 in Anaconda. The purpose of this package
is to utilize Scrapy. Scrapy is a Python program that enables the user to send, sniff and dissect and forge
network packets. This capability allows construction of tools that can probe, scan or attack networks. For
your knowledge a network packet in simple terms is a unit of data transferred over a computer network.
In this tutorial I will walk you through utilizing scrapy to walk through a website and gather data from
that website by creating a Spider. For this specific tutorial we will be walking through a chocolate
website, chocolate.co.uk. However, this tutorial can be applied to numerous websites like Walmart,
Amazon, Liquor stores, and much more.
Languages and Utilities Used
Python
Anaconda
Environments Used
Windows 11 (21H2)
Program Reflection & Code:
Creating Environment:
Creating Environment & Project File:
Creating Spider:
Start Using Spider:
Accessing Website:
Extracting Data:
Extracting Data (contd.):
Extracting Data (contd.):
Saving File & Sources (contd.):
About
This tutorial uses Python 3.10.10 and Anaconda to demonstrate Scrapy, a tool for sending, sniffing, and forging network packets. You'll learn to create a Spider to scrape data from websites, with a practical example using chocolate.co.uk, though the method is applicable to other sites like Walmart and Amazon.