diff options
Diffstat (limited to 'utils/test/vnfcatalogue')
-rw-r--r-- | utils/test/vnfcatalogue/cronjobs/README.md | 24 | ||||
-rw-r--r-- | utils/test/vnfcatalogue/cronjobs/database.js | 14 | ||||
-rw-r--r-- | utils/test/vnfcatalogue/cronjobs/github.js | 129 |
3 files changed, 167 insertions, 0 deletions
diff --git a/utils/test/vnfcatalogue/cronjobs/README.md b/utils/test/vnfcatalogue/cronjobs/README.md new file mode 100644 index 000000000..cf27ff8c7 --- /dev/null +++ b/utils/test/vnfcatalogue/cronjobs/README.md @@ -0,0 +1,24 @@ +# CRONJOB Directory + +## Helper to setup cronjob to fill the vnf table + +There are two important parameters that need to be set in github.js +before running cronjob. + + +``` + access_token : generate an access token from github account for accessing + the github apis. This is necessary as the non access token limit tends to + be 50 api calls per hour. + delta : the threshold between the last update of the row of the vnf table + and current time. It is measured in seconds. +``` + +Enter the details namely username and password in the **database.js**. +Then setup the cronjob by putting the following line in the crontab + +In the crontab + +```bash + node github +``` diff --git a/utils/test/vnfcatalogue/cronjobs/database.js b/utils/test/vnfcatalogue/cronjobs/database.js new file mode 100644 index 000000000..a1d926e44 --- /dev/null +++ b/utils/test/vnfcatalogue/cronjobs/database.js @@ -0,0 +1,14 @@ +var mysql = require('mysql'); + +var pool = mysql.createPool({ + host: 'localhost', + user: 'myuser', + password: 'mypassword', + database: 'vnf_catalogue', + connectionLimit: 50, + supportBigNumbers: true, + multipleStatements: true, + dateStrings: 'date' +}); + +exports.pool = pool; diff --git a/utils/test/vnfcatalogue/cronjobs/github.js b/utils/test/vnfcatalogue/cronjobs/github.js new file mode 100644 index 000000000..05cc6c155 --- /dev/null +++ b/utils/test/vnfcatalogue/cronjobs/github.js @@ -0,0 +1,129 @@ +// Important Add your access token here default rate of github is limited to 60 API calls per hour +var access_token = '*'; +// Important set the delta threshold for repo details updation. For instance if the threshold is +// set to 1 day(60 * 60 * 24), the cronjob will only update the row if the difference between current +// time and last_updated time stamp of a repo is greater than one day +var delta = 60 * 60 * 24; + + +var github = require('octonode'); +db_pool = require('./database').pool; +async = require('async'); + +var current_time = Math.floor(new Date().getTime() / 1000);//toISOString().slice(0, 19).replace('T', ' '); +console.log(current_time); + +var get_val_from_header = function(header_link) { + // small hack by parsing the header and setting per_page = 1, hence no pagination fetch required + result_intermediate = header_link.split(';'); + result_intermediate = result_intermediate[result_intermediate.length - 2]; + var reg = /&page=([0-9].*)>/g; + var match = reg.exec(result_intermediate); + return parseInt(match[1]); +} + +var get_stargazers = function(result, ghrepo, primary_callback, cb) { + ghrepo.stargazers({per_page: 1}, function(err, data, headers) { + //console.log(JSON.stringify(data)); + try { + result['no_of_stars'] = get_val_from_header(headers['link']); + cb(null, result, ghrepo, primary_callback); + } catch(err) { + result['no_of_stars'] = null; + cb(null, result, ghrepo, primary_callback); + } + }); +} + +var get_branches = function(result, ghrepo, primary_callback, cb) { + ghrepo.branches({per_page: 1}, function(err, data, headers) { + try { + result['versions'] = get_val_from_header(headers['link']); + cb(null, result, ghrepo, primary_callback); + } catch(err) { + result['versions'] = null; + cb(null, result, ghrepo, primary_callback); + } + }); +} + +var get_contributors = function(result, ghrepo, primary_callback, cb) { + ghrepo.contributors({per_page: 1}, function(err, data, headers) { + try { + result['no_of_developers'] = get_val_from_header(headers['link']); + cb(null, result, primary_callback); + } catch(err) { + result['no_of_developers'] = null; + cb(null, result, primary_callback); + + } + }); +} + +var get_lines_of_code = function(result, cb) { + // #TODO +} + +var secondary_callback = function (err, result, primary_callback) { + console.log(result); + if((result['last_updated'] == null) || (current_time - result['last_updated'] > delta)) { + db_pool.getConnection(function(err, connection) { + //Use the connection + var last_updated = current_time; + var no_of_stars = result['no_of_stars']; + var versions = result['versions']; + var no_of_developers = result['no_of_developers']; + sql_query = 'update vnf set last_updated = FROM_UNIXTIME(' + last_updated; + sql_query += '), no_of_stars = ' + no_of_stars + ', versions = ' + versions; + sql_query += ', no_of_developers = ' + no_of_developers + ' where vnf_id = '; + sql_query += result['vnf_id']; + console.log(sql_query); + connection.query(sql_query, function (error, results, fields) { + if (error) throw error; + //And done with the connection. + primary_callback(null, result['vnf_id'] + ' updated'); + connection.release(); + // Handle error after the release. + // Don't use the connection here, it has been returned to the pool. + }); + }); + } else { + primary_callback(null, result['vnf_id'] + ' not updated'); + } +} + +var get_stats = function(vnf_details, callback) { + repo = vnf_details['repo_url']; + repo = repo.split("/"); + github_id = repo[repo.length - 2] + '/' + repo[repo.length - 1]; + + var async = require('async'); + var client = github.client(access_token); + var ghrepo = client.repo(github_id); + + result = {} + result['vnf_id'] = vnf_details['vnf_id']; + result['last_updated'] = vnf_details['last_updated']; + + async.waterfall([ + async.apply(get_stargazers, result, ghrepo, callback), + get_branches, + get_contributors, + //get_lines_of_code, + ], secondary_callback); +} + +db_pool.getConnection(function(err, connection) { + sql_query = 'select vnf_id, repo_url, UNIX_TIMESTAMP(last_updated) last_updated from vnf'; + console.log(sql_query); + connection.query(sql_query, function (error, results, fields) { + if (error) throw error; + async.map(results, get_stats, function(error, results) { + //console.log(results); + console.log(results); + process.exit(); + + }); + }); +}); + |