The data collector consists of two parts:
crawler.py
script which walks through commits indata/atom
repository- node.js script
index.js
, which launches a server and waits for requests fromcrawler.py
and returns serialized json with collected metrics
To run data collection:
mkdir data
and clone atom repository therenpm install
- Run node.js server:
node index.js
- Run crawler:
python3 crawler.py
To install and run jupyter notebook:
sudo -H pip3 install --upgrade pip
sudo -H pip3 install virtualenv
virtualenv jupyter_env
source jupyter_env/bin/activate
pip install jupyter
pip install pandas
pip install matplotlib
pip install sklearn
jupyter notebook
To run the app in docker do the following:
- Install docker
- Run docker container:
docker-compose run --rm metrics /bin/bash
- Run any commands