swx0 / linkedin-jobs-scraper

Python script to extract data of Linkedin job openings using Selenium and BeautifulSoup. Storing the data using Google Sheets API.

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Linkedin Jobs Scraper 🔍

About

Linkedin Jobs Scraper extracts the following data from Linkedin search results:

  • Job data-id

  • Location

  • Company Name

  • Position

  • Job description

Afterwhich, the data obtained would be written and saved in Google sheets using Google Sheets API

demo

Setup

To start scraping job information from Linkedin:

  1. Install these libraries:
pip install selenium
pip install beautifulsoup4
pip install gspread
  1. Find the respective WebDriver for your Chrome version. Save this chromedriver.exe in the same directory as the .py script

  2. Create Google Cloud Platform account (if you haven't). Follow this guide to authorize and authenticate access to the sheets. You should obtain the JSON file with credentials, rename this file as 'client_key.json' and save it in the same directory as the .py script.

✔️ Remember to share the created Google sheets with the client_email found in the JSON file.

  1. Obtain a link from Linkedin that indicates your target position and location. The following is a sample link "https://www.linkedin.com/jobs/search/?keywords=Software%20Engineer&location=United%20States"

  2. Read the comments in .py script, and edit:

    • string as the title name of your Google Sheets
    sheet = client.open('rename sheets title here').sheet1
    
    • Link from step 4)
    driver.get('https://www.linkedin.com/jobs/search/?keywords=Software%20Engineer&location=United%20States')
    
  3. Run .py script

About

Python script to extract data of Linkedin job openings using Selenium and BeautifulSoup. Storing the data using Google Sheets API.

License:Apache License 2.0


Languages

Language:Python 100.0%