Content and scripts from this repository can be freely reused for academic purposes, provided that you cite the following paper in your work:
F. Calefato, F. Lanubile, and B. Vasilescu (2019) “A large-scale, in-depth analysis of developers’
personalities in the Apache ecosystem.” Information and Software Technology, Vol. 114, Oct.,
pp. 1-20, DOI: 10.1016/j.infsof.2019.05.012.
The final results of the scripts (i.e., developers' monthly scores per project) are stored here (see the files in CSV format). Instead, the entire MySQL database, containing the data scraped from the Apache website, the email archives, and the code metadata obtained from GitHub, is stored here.
The dump can be imported into a pre-existing db named apache as follows:
$ mysql -u <username> -p<PlainPassword> apache < apachebig5.sqlRepeat the instruction above for all the .sql files provided.
$ git clone https://github.com/collab-uniba/personality.git --recursiveEdit the following configuration files:
src/python/db/cfg/setup.yml- MySQL database configuration
mysql:
host: 127.0.0.1
user: root
passwd: *******
db: apachesrc/python/big5_personality/personality_insights/cfg/watson.yml- IBM Watson Personality Insights (you will need to register and get your personal username and password)
personality:
username: secret-user
password: secret-password
version: 2017-10-13src/python/big5_personality/liwc/cfg/receptiviti.yaml- Receptiviti (you will need to register and get your personal api key and api secret key)
receptiviti:
baseurl: https://api-v3.receptiviti.com
api_key: *****
api_secret_key: *****- Setup:
First, install the library libgit2 on your system. Then, use a Python 3 environment and install the required packages from
src/python/requirements.txt - Execution:
From directory
src/python/apache_crawlerrun:
$ scrapy apache_crawler -t (json|csv) -o apache-projects.(json|csv) [-L DEBUG --logfile apache.log]- Setup:
Use Python 2 environment and install packages from
src/python/ml_downloader/requirements.txt. Then, recreate database schema as follows:
$ mysql -u<user> -p<password> apache < submodules/mlminer/db/data_model_mysql.sql- Execution:
From directory
src/python/ml_downloaderrun:
$ sh run.sh- Setup: Use Python 3 environment as described in Step 3.
- Execution:
From directory
src/python/git_clonerrun:
$ sh run.shProjects will be cloned into the subfolder apache_repos.
- Setup:
- In MySql command line enter following instruction:
set character set utf8mb4;
- Use Python 3 environment as described in Step 3.
- Add a new file
github-api-tokens.txtand enter a GitHub API access token
- In MySql command line enter following instruction:
- Execution:
From directory
src/python/github_users_locationrun:
$ sh run.sh [reset]where:
- reset: to empty db table containing github users location
- Setup:
- Use Python 3 environment as described in Step 2.
- At first run, execute from dicrectory
src/python/unmasking:$ python nltk_download.py
- Execution:
From directory
src/python/unmaskingrun:
$ sh run.sh- Setup: Use Python 3 environment as described in Step 3.
- Execution:
From directory
src/python/commit_analyzerrun:
$ sh run.sh - Setup:
- Install NLoN package as described here
- Use Python 3 environment as described in Step 3.
- Execution:
From directory
src/python/big5_personalityrun:
$ sh run.sh <tool> [reset]where:
- tool: tool name, either
liwc15orp_insights - reset: to empty the db tables containing personality data before new computing
- Setup: Use Python 3 environment as described in Step 3.
- Execution:
From directory
src/python/export_resultsrun:
$ sh run.sh <tool>where:
- tool: tool name, with values in {
liwc07,liwc15,p_insights}
Results are stored in files personality_liwc.csv and personality_p_insights.csv.