AiCruit is a full-stack AI-powered recruitment platform that streamlines the hiring process.
- Figma Designs:
- Backend: Ruby on Rails (API-only)
- Frontend: Next.js
- Microservices: Python
Clone the repository:
git clone git@github.com:joshsoftware/aicruit.git
aicruit/
├── aicruit-api # Ruby on Rails backend (API-only)
├── app # Next.js frontend
└── service # Python-based microservices
Please use the dev
branch for all development work.
Make sure to branch off from dev
and open pull requests into dev
.
- Ruby version: 3.3.2 (RVM recommended)
- Rails version: 7.2.2
- PostgreSQL: 12+
-
Navigate to the backend directory:
cd aicruit/aicruit-api
-
Create
.env
file (refer to.env.example
for required variables) -
Install dependencies:
bundle install
-
Set up the database:
rails db:create rails db:migrate rails db:seed
-
Start the Rails server:
rails s
The Rails API will be available at http://localhost:3000
- Node.js 18+
- npm
-
Navigate to the frontend directory:
cd aicruit/app
-
Create
.env.local
file (refer to.env.example
if available) -
Install dependencies:
npm install
-
Run the development server:
npm run dev
The frontend will be available at http://localhost:5173
-
Open a terminal window
-
Edit the hosts file with root permissions:
sudo nano /etc/hosts
-
Add the following line at the end of the file:
127.0.0.1 joshsoftware.aicruit.com
-
Save and exit the editor
-
Test the configuration:
ping joshsoftware.aicruit.com
-
Access the frontend using the custom domain:
http://joshsoftware.aicruit.com:5173
This setup allows you to access your local development environment using the custom domain, which is essential for testing multitenancy features. For the frontend application, replace http://localhost:5173 with http://joshsoftware.aicruit.com:5173 in your browser.
Pre-requisite: Note: Following instructions a for linux, python 3.8.1 or above
sudo apt-get update
sudo apt-get install python3.8.1
ffmpeg
sudo apt update && sudo apt install ffmpeg
Ollam
For Linux:
curl -fsSL https://ollama.com/install.sh | sh
For Mac:
https://ollama.com/download/Ollama-darwin.zip
For Windows:
https://ollama.com/download/OllamaSetup.exe
Llama 3.2 model
ollama run llama3.2
Setup Clone this github repository git clone
Create python virtual environment
python3 -m venv lingo .
Activate the virtual environment
source lingo/bin/activate
Install dependencies
pip install -r requirements.txt
uvicorn main:app --host localhost --port 8000 --reload