Selenium is a free (open-source) automated testing framework used to validate web applications across different browsers and platforms.
https://medium.com/@kurumert/web-scraping-linkedin-job-page-with-selenium-python-e0b6183a5954
In this project, I have used Selenium library to automate the scraping process, navigate between pages, fill out forms (such as login credentials), click links, etc. Briefly, besides the data scraping part, Selenium was used to automate the following steps mentioned below to manage some actions until we access the pages where we will scrape the data. I have scraped the job offers and scraped their information by visiting 14 pages with each page having 25 job offers.
Automated Steps
- Enter to Linkedin login page
- Spot the cookies pop-up and click Accept cookies
- Fill E-Mail Address and Password areas and click Login
- Click on the Jobs from the section above
- Search for job positions Junior Data Analyst in Spain
- Scroll down till the end of the page while collecting the link of each displayed Job offer
- Go to the next page when it is the end of the page while keep collecting links
- After all links are collected, go to the each link
- Click the "See More" button the expand the job description text.
- Scrape the desired data
Items Scraped:
- Job Title
- Company Name
- Company Location
- Job Description
- Work Method (Hybrid, Remote, On-Site)
- Experience Level
- Post Date
Extra:
You can check my medium post where I have analyzed the data and found some useful insights: https://medium.com/@kurumert/the-insights-behind-the-data-field-job-offers-on-linkedin-4edb5203525b