stobasa / webpage-crawler

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Team-Granite-Backend

A Web Crawler Parser

=========================

  • Run the App:
  • Fork this repository
  • Clone to your local machine
  • cd into the repository and create a virtual environment with pipenv library
  • activate the environmment with
pipenv shell
  • Then run
pipenv install -r requirements.txt

This will download all the dependencies for this application.

Using the service

  • We have Scraper as part of the dependencies.
  • We have BeautifulSoup as part of the dependencies.
  • We have Pandas as part of the dependencies.
  • Make sure you import pandas. You can import pandas as pd.
  • To read the excel run, use this function
 pd.read_excel('name of file.xlsx')

Please make sure you read the file into a variable, acceptable variable names include d_frame, df, data_frame, etc. To read the data frame into another another file, use this function d_frame.to_json(name of file.csv) For more information on how to manipulate data using pandas visit https://pandas.pydata.org/pandas-docs/stable/user_guide/io.html

Contribute guide

If you're in team-granite-backend:

About


Languages

Language:Python 100.0%