| var fs |
| try { |
| fs = require('graceful-fs') |
| } catch (er) { |
| fs = require('fs') |
| } |
| |
| var path = require('path') |
| |
| var glob = require('glob') |
| var normalizeData = require('normalize-package-data') |
| var safeJSON = require('json-parse-better-errors') |
| var util = require('util') |
| var slash = require('slash') |
| |
| module.exports = readJson |
| |
| // put more stuff on here to customize. |
| readJson.extraSet = [ |
| bundleDependencies, |
| gypfile, |
| serverjs, |
| scriptpath, |
| authors, |
| readme, |
| mans, |
| bins, |
| githead |
| ] |
| |
| var typoWarned = {} |
| var cache = {} |
| |
| function readJson (file, log_, strict_, cb_) { |
| var log, strict, cb |
| for (var i = 1; i < arguments.length - 1; i++) { |
| if (typeof arguments[i] === 'boolean') { |
| strict = arguments[i] |
| } else if (typeof arguments[i] === 'function') { |
| log = arguments[i] |
| } |
| } |
| |
| if (!log) log = function () {} |
| cb = arguments[ arguments.length - 1 ] |
| |
| readJson_(file, log, strict, cb) |
| } |
| |
| function readJson_ (file, log, strict, cb) { |
| fs.readFile(file, 'utf8', function (er, d) { |
| parseJson(file, er, d, log, strict, cb) |
| }) |
| } |
| |
| function stripBOM (content) { |
| // Remove byte order marker. This catches EF BB BF (the UTF-8 BOM) |
| // because the buffer-to-string conversion in `fs.readFileSync()` |
| // translates it to FEFF, the UTF-16 BOM. |
| if (content.charCodeAt(0) === 0xFEFF) content = content.slice(1) |
| return content |
| } |
| |
| function jsonClone (obj) { |
| if (obj == null) { |
| return obj |
| } else if (Array.isArray(obj)) { |
| var newarr = new Array(obj.length) |
| for (var ii in obj) { |
| newarr[ii] = obj[ii] |
| } |
| } else if (typeof obj === 'object') { |
| var newobj = {} |
| for (var kk in obj) { |
| newobj[kk] = jsonClone[kk] |
| } |
| } else { |
| return obj |
| } |
| } |
| |
| function parseJson (file, er, d, log, strict, cb) { |
| if (er && er.code === 'ENOENT') { |
| return fs.stat(path.dirname(file), function (err, stat) { |
| if (!err && stat && !stat.isDirectory()) { |
| // ENOTDIR isn't used on Windows, but npm expects it. |
| er = Object.create(er) |
| er.code = 'ENOTDIR' |
| return cb(er) |
| } else { |
| return indexjs(file, er, log, strict, cb) |
| } |
| }) |
| } |
| if (er) return cb(er) |
| |
| if (cache[d]) return cb(null, jsonClone(cache[d])) |
| |
| var data |
| |
| try { |
| data = safeJSON(stripBOM(d)) |
| } catch (er) { |
| data = parseIndex(d) |
| if (!data) return cb(parseError(er, file)) |
| } |
| |
| extrasCached(file, d, data, log, strict, cb) |
| } |
| |
| function extrasCached (file, d, data, log, strict, cb) { |
| extras(file, data, log, strict, function (err, data) { |
| if (!err) { |
| cache[d] = jsonClone(data) |
| } |
| cb(err, data) |
| }) |
| } |
| |
| function indexjs (file, er, log, strict, cb) { |
| if (path.basename(file) === 'index.js') return cb(er) |
| |
| var index = path.resolve(path.dirname(file), 'index.js') |
| fs.readFile(index, 'utf8', function (er2, d) { |
| if (er2) return cb(er) |
| |
| if (cache[d]) return cb(null, cache[d]) |
| |
| var data = parseIndex(d) |
| if (!data) return cb(er) |
| |
| extrasCached(file, d, data, log, strict, cb) |
| }) |
| } |
| |
| readJson.extras = extras |
| function extras (file, data, log_, strict_, cb_) { |
| var log, strict, cb |
| for (var i = 2; i < arguments.length - 1; i++) { |
| if (typeof arguments[i] === 'boolean') { |
| strict = arguments[i] |
| } else if (typeof arguments[i] === 'function') { |
| log = arguments[i] |
| } |
| } |
| |
| if (!log) log = function () {} |
| cb = arguments[i] |
| |
| var set = readJson.extraSet |
| var n = set.length |
| var errState = null |
| set.forEach(function (fn) { |
| fn(file, data, then) |
| }) |
| |
| function then (er) { |
| if (errState) return |
| if (er) return cb(errState = er) |
| if (--n > 0) return |
| final(file, data, log, strict, cb) |
| } |
| } |
| |
| function scriptpath (file, data, cb) { |
| if (!data.scripts) return cb(null, data) |
| var k = Object.keys(data.scripts) |
| k.forEach(scriptpath_, data.scripts) |
| cb(null, data) |
| } |
| |
| function scriptpath_ (key) { |
| var s = this[key] |
| // This is never allowed, and only causes problems |
| if (typeof s !== 'string') return delete this[key] |
| |
| var spre = /^(\.[/\\])?node_modules[/\\].bin[\\/]/ |
| if (s.match(spre)) { |
| this[key] = this[key].replace(spre, '') |
| } |
| } |
| |
| function gypfile (file, data, cb) { |
| var dir = path.dirname(file) |
| var s = data.scripts || {} |
| if (s.install || s.preinstall) return cb(null, data) |
| |
| glob('*.gyp', { cwd: dir }, function (er, files) { |
| if (er) return cb(er) |
| if (data.gypfile === false) return cb(null, data) |
| gypfile_(file, data, files, cb) |
| }) |
| } |
| |
| function gypfile_ (file, data, files, cb) { |
| if (!files.length) return cb(null, data) |
| var s = data.scripts || {} |
| s.install = 'node-gyp rebuild' |
| data.scripts = s |
| data.gypfile = true |
| return cb(null, data) |
| } |
| |
| function serverjs (file, data, cb) { |
| var dir = path.dirname(file) |
| var s = data.scripts || {} |
| if (s.start) return cb(null, data) |
| glob('server.js', { cwd: dir }, function (er, files) { |
| if (er) return cb(er) |
| serverjs_(file, data, files, cb) |
| }) |
| } |
| |
| function serverjs_ (file, data, files, cb) { |
| if (!files.length) return cb(null, data) |
| var s = data.scripts || {} |
| s.start = 'node server.js' |
| data.scripts = s |
| return cb(null, data) |
| } |
| |
| function authors (file, data, cb) { |
| if (data.contributors) return cb(null, data) |
| var af = path.resolve(path.dirname(file), 'AUTHORS') |
| fs.readFile(af, 'utf8', function (er, ad) { |
| // ignore error. just checking it. |
| if (er) return cb(null, data) |
| authors_(file, data, ad, cb) |
| }) |
| } |
| |
| function authors_ (file, data, ad, cb) { |
| ad = ad.split(/\r?\n/g).map(function (line) { |
| return line.replace(/^\s*#.*$/, '').trim() |
| }).filter(function (line) { |
| return line |
| }) |
| data.contributors = ad |
| return cb(null, data) |
| } |
| |
| function readme (file, data, cb) { |
| if (data.readme) return cb(null, data) |
| var dir = path.dirname(file) |
| var globOpts = { cwd: dir, nocase: true, mark: true } |
| glob('{README,README.*}', globOpts, function (er, files) { |
| if (er) return cb(er) |
| // don't accept directories. |
| files = files.filter(function (file) { |
| return !file.match(/\/$/) |
| }) |
| if (!files.length) return cb() |
| var fn = preferMarkdownReadme(files) |
| var rm = path.resolve(dir, fn) |
| readme_(file, data, rm, cb) |
| }) |
| } |
| |
| function preferMarkdownReadme (files) { |
| var fallback = 0 |
| var re = /\.m?a?r?k?d?o?w?n?$/i |
| for (var i = 0; i < files.length; i++) { |
| if (files[i].match(re)) { |
| return files[i] |
| } else if (files[i].match(/README$/)) { |
| fallback = i |
| } |
| } |
| // prefer README.md, followed by README; otherwise, return |
| // the first filename (which could be README) |
| return files[fallback] |
| } |
| |
| function readme_ (file, data, rm, cb) { |
| var rmfn = path.basename(rm) |
| fs.readFile(rm, 'utf8', function (er, rm) { |
| // maybe not readable, or something. |
| if (er) return cb() |
| data.readme = rm |
| data.readmeFilename = rmfn |
| return cb(er, data) |
| }) |
| } |
| |
| function mans (file, data, cb) { |
| var m = data.directories && data.directories.man |
| if (data.man || !m) return cb(null, data) |
| m = path.resolve(path.dirname(file), m) |
| glob('**/*.[0-9]', { cwd: m }, function (er, mans) { |
| if (er) return cb(er) |
| mans_(file, data, mans, cb) |
| }) |
| } |
| |
| function mans_ (file, data, mans, cb) { |
| var m = data.directories && data.directories.man |
| data.man = mans.map(function (mf) { |
| return path.resolve(path.dirname(file), m, mf) |
| }) |
| return cb(null, data) |
| } |
| |
| function bins (file, data, cb) { |
| if (Array.isArray(data.bin)) return bins_(file, data, data.bin, cb) |
| |
| var m = data.directories && data.directories.bin |
| if (data.bin || !m) return cb(null, data) |
| |
| m = path.resolve(path.dirname(file), m) |
| glob('**', { cwd: m }, function (er, bins) { |
| if (er) return cb(er) |
| bins_(file, data, bins, cb) |
| }) |
| } |
| |
| function bins_ (file, data, bins, cb) { |
| var m = (data.directories && data.directories.bin) || '.' |
| data.bin = bins.reduce(function (acc, mf) { |
| if (mf && mf.charAt(0) !== '.') { |
| var f = path.basename(mf) |
| acc[f] = slash(path.join(m, mf)) |
| } |
| return acc |
| }, {}) |
| return cb(null, data) |
| } |
| |
| function bundleDependencies (file, data, cb) { |
| var bd = 'bundleDependencies' |
| var bdd = 'bundledDependencies' |
| // normalize key name |
| if (data[bdd] !== undefined) { |
| if (data[bd] === undefined) data[bd] = data[bdd] |
| delete data[bdd] |
| } |
| if (data[bd] === false) delete data[bd] |
| else if (data[bd] === true) { |
| data[bd] = Object.keys(data.dependencies || {}) |
| } else if (data[bd] !== undefined && !Array.isArray(data[bd])) { |
| delete data[bd] |
| } |
| return cb(null, data) |
| } |
| |
| function githead (file, data, cb) { |
| if (data.gitHead) return cb(null, data) |
| var dir = path.dirname(file) |
| var head = path.resolve(dir, '.git/HEAD') |
| fs.readFile(head, 'utf8', function (er, head) { |
| if (er) return cb(null, data) |
| githead_(file, data, dir, head, cb) |
| }) |
| } |
| |
| function githead_ (file, data, dir, head, cb) { |
| if (!head.match(/^ref: /)) { |
| data.gitHead = head.trim() |
| return cb(null, data) |
| } |
| var headRef = head.replace(/^ref: /, '').trim() |
| var headFile = path.resolve(dir, '.git', headRef) |
| fs.readFile(headFile, 'utf8', function (er, head) { |
| if (er || !head) { |
| var packFile = path.resolve(dir, '.git/packed-refs') |
| return fs.readFile(packFile, 'utf8', function (er, refs) { |
| if (er || !refs) { |
| return cb(null, data) |
| } |
| refs = refs.split('\n') |
| for (var i = 0; i < refs.length; i++) { |
| var match = refs[i].match(/^([0-9a-f]{40}) (.+)$/) |
| if (match && match[2].trim() === headRef) { |
| data.gitHead = match[1] |
| break |
| } |
| } |
| return cb(null, data) |
| }) |
| } |
| head = head.replace(/^ref: /, '').trim() |
| data.gitHead = head |
| return cb(null, data) |
| }) |
| } |
| |
| /** |
| * Warn if the bin references don't point to anything. This might be better in |
| * normalize-package-data if it had access to the file path. |
| */ |
| function checkBinReferences_ (file, data, warn, cb) { |
| if (!(data.bin instanceof Object)) return cb() |
| |
| var keys = Object.keys(data.bin) |
| var keysLeft = keys.length |
| if (!keysLeft) return cb() |
| |
| function handleExists (relName, result) { |
| keysLeft-- |
| if (!result) warn('No bin file found at ' + relName) |
| if (!keysLeft) cb() |
| } |
| |
| keys.forEach(function (key) { |
| var dirName = path.dirname(file) |
| var relName = data.bin[key] |
| if (typeof relName !== 'string') { |
| var msg = 'Bin filename for ' + key + |
| ' is not a string: ' + util.inspect(relName) |
| warn(msg) |
| delete data.bin[key] |
| handleExists(relName, true) |
| return |
| } |
| var binPath = path.resolve(dirName, relName) |
| fs.stat(binPath, (err) => handleExists(relName, !err)) |
| }) |
| } |
| |
| function final (file, data, log, strict, cb) { |
| var pId = makePackageId(data) |
| |
| function warn (msg) { |
| if (typoWarned[pId]) return |
| if (log) log('package.json', pId, msg) |
| } |
| |
| try { |
| normalizeData(data, warn, strict) |
| } catch (error) { |
| return cb(error) |
| } |
| |
| checkBinReferences_(file, data, warn, function () { |
| typoWarned[pId] = true |
| cb(null, data) |
| }) |
| } |
| |
| function makePackageId (data) { |
| var name = cleanString(data.name) |
| var ver = cleanString(data.version) |
| return name + '@' + ver |
| } |
| |
| function cleanString (str) { |
| return (!str || typeof (str) !== 'string') ? '' : str.trim() |
| } |
| |
| // /**package { "name": "foo", "version": "1.2.3", ... } **/ |
| function parseIndex (data) { |
| data = data.split(/^\/\*\*package(?:\s|$)/m) |
| |
| if (data.length < 2) return null |
| data = data[1] |
| data = data.split(/\*\*\/$/m) |
| |
| if (data.length < 2) return null |
| data = data[0] |
| data = data.replace(/^\s*\*/mg, '') |
| |
| try { |
| return safeJSON(data) |
| } catch (er) { |
| return null |
| } |
| } |
| |
| function parseError (ex, file) { |
| var e = new Error('Failed to parse json\n' + ex.message) |
| e.code = 'EJSONPARSE' |
| e.file = file |
| return e |
| } |