The Resume Matcher takes your resume and job descriptions as input, parses them using Python, and mimics the functionalities of an ATS, providing you with insights and suggestions to make your resume ATS-friendly.
The process is as follows:
-
Parsing: The system uses Python to parse both your resume and the provided job description, just like an ATS would. Parsing is critical as it transforms your documents into a format the system can readily analyze.
-
Keyword Extraction: The tool uses advanced machine learning algorithms to extract the most relevant keywords from the job description. These keywords represent the skills, qualifications, and experiences the employer seeks.
-
Key Terms Extraction: Beyond keyword extraction, the tool uses textacy to identify the main key terms or themes in the job description. This step helps in understanding the broader context of what the resume is about.
-
Vector Similarity Using Qdrant: The tool uses Qdrant, a highly efficient vector similarity search tool, to measure how closely your resume matches the job description. This process is done by representing your resume and job description as vectors in a high-dimensional space and calculating their cosine similarity. The more similar they are, the higher the likelihood that your resume will pass the ATS screening.
On top of that, there are various data visualizations that I've added to help you get started.
π¬ Feel free to join in the Discord to discuss anything.
π» Create an issue.
π©βπ» Contribute.
πParticipate in the Discussions/Announcements.
π§ͺ Please check the Landing Page. PRs are also welcomed over there.
Follow these steps to set up the environment and run the application:
- Clone the Repository:
git clone https://github.com/srbhr/Resume-Matcher.git
cd Resume-Matcher
- Create a Python Virtual Environment:
python -m venv env
- Activate the Virtual Environment:
-
On Windows:
env\Scripts\activate
-
On macOS and Linux:
source env/bin/activate
- Install Dependencies:
pip install -r requirements.txt
- Prepare Data:
- Resumes: Place your resumes in PDF format in the
Data/Resumes
folder. Remove any existing contents in this folder. - Job Descriptions: Place your job descriptions in PDF format in the
Data/JobDescription
folder. Remove any existing contents in this folder.
- Parse Resumes to JSON:
python run_first.py
- Run the Application:
streamlit run streamlit_app.py
Note: For local versions, you do not need to run "streamlit_second.py" as it is specifically for deploying to Streamlit servers.
Additional Note: The Vector Similarity part is precomputed to optimize performance due to the resource-intensive nature of sentence encoders that require significant GPU and RAM resources. If you are interested in leveraging this feature in a Google Colab environment for free, refer to the upcoming blog (link to be provided) for further guidance.
- Build image and start application
docker-compose up
- Open
localhost:80
on your browser
Thanks for the support π this is an ongoing project that I want to build with open source community. There are many ways in which this tool can be upgraded. This includes (not limited to):
- Create a better dashboard instead of Streamlit.
- Add more features like uploading of resumes and parsing.
- Add a docker image for easy usage.
- Contribute to better parsing algorithm.
- Contribute to on a blog to how to make this work.
- Contribute to the landing page maybe re-create in React/Vue/etc.