Python Code
Overview
The project addresses the need for job seekers to identify trending roles and their locations. It aims to offer personalized and current insights using LinkedIn data. The process involves web scraping, data cleaning, transformation, and visualization for a comprehensive job market overview.
Project Steps
- User input & URL generation: The project begins by taking user input for the desired job title and generating a corresponding URL for web scraping.
- Web Scraping: To scrape the data from the URL, pythons requests library was used to retrieve the HTML content from the server and store data in a Python object. Beautiful Soup was used to parse the lengthy code and make it more accessible to extract the relevant information.
- Create Data frame: Job data, including titles, companies, locations, and levels, is stored in a structured DataFrame.
- Data Cleaning: The location field is split into city and state, and entries with missing values are removed for accuracy.
- Data Analysis and Visualisation: Utilizing packages seaborn and matplotlib, further analysis and visualisation was carried out to determine the most common job levels being recruited for, understanding the distribution of job levels across different states, identifying the industries that are actively hiring, exploring the relationship between job levels and industries, and examining the association between job functions and job levels.
Conclusion
By offering insights into job market trends, this project empowers job seekers and employers with valuable decision-making information. It combines web scraping and data analysis to provide a clear and informative job landscape view.