Skip to content

Scrapy to scrape/crawl website and get data to store and analyze

License

Notifications You must be signed in to change notification settings

sachin-s-joshi/Scrapy

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

29 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Scrapy

Scrapy to scrape/crawl website and get data to store and analyze

Command for run

Setup virtual environment with – pyhton3 -e venv nameofenv , a folder is created with name in root directory example (bot)

Activate virtual environment with – source /virtual_environment_folder/bin/activate

After activation, install all required packages using python install -r requirements.txt

And , now under subdirectory demo, run – scrapy crawl me

Running with Custom URLs:

After the project is set setup create .env file at root level.

Following variable are set now:

  • site - for website url

This .env file is served as environment variable to specify urls you want to crawl

Command to run:

  • Setup virtual environment with – pyhton3 -e venv nameofenv , a folder is created with name in root directory example (bot)
  • Activate virtual environment with – source /virtual_environment_folder/bin/activate
  • After activation, install all required packages using python install -r requirements.txt
  • And , now under subdirectory demo, run – scrapy crawl me

Running with Docker images:

  • Run docker build -t <image_name> .
  • docker run crawler:custom crawl me

About

Scrapy to scrape/crawl website and get data to store and analyze

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published