Skip to content

claudioandriaan/Python_web_scraping_test

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

3 Commits
 
 
 
 

Repository files navigation

Web Scraping Project

This project is designed to scrape listing and detail pages from a website and store the extracted data into Google Sheets. Additionally, it downloads images from the detail pages and uploads them to an FTP server.

Setup

  1. Clone the Repository: git clone https://github.com/claudioandriaan/Python_web_scraping_test.git

  2. Install Dependencies:

  pip install BeautifulSoup
  
  pip install selenium 
  
  pip install gspread 
  
  pip install ftputil 
  
  pip install requests 

  1. Set Up Google API Credentials:
  • Obtain Google API credentials (JSON file) and save it as dot.json in the project directory.
  1. Configure FTP Credentials:
  • Update the FTP server details (host, username, password) in the scrape_data_from_link() function in the script.
  1. Run the Script:
python spiders.py -d <output_directory>

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages