mirror of https://github.com/status-im/bip39.git
clean output dir before build
This commit is contained in:
parent
0bc19393b0
commit
d84e2821f3
|
@ -5,7 +5,8 @@
|
|||
"main": "src/index.js",
|
||||
"types": "./types/index.d.ts",
|
||||
"scripts": {
|
||||
"build": "tsc -p tsconfig.json",
|
||||
"build": "npm run clean && tsc -p tsconfig.json",
|
||||
"clean": "rm -rf src",
|
||||
"coverage": "nyc --branches 100 --functions 100 --check-coverage npm run unit",
|
||||
"format": "npm run prettier -- --write",
|
||||
"format:ci": "npm run prettier -- --check",
|
||||
|
|
|
@ -1,169 +0,0 @@
|
|||
"use strict";
|
||||
Object.defineProperty(exports, "__esModule", { value: true });
|
||||
const createHash = require("create-hash");
|
||||
const pbkdf2_1 = require("pbkdf2");
|
||||
const randomBytes = require("randombytes");
|
||||
// use unorm until String.prototype.normalize gets better browser support
|
||||
const unorm = require("unorm");
|
||||
// import CHINESE_SIMPLIFIED_WORDLIST from '../wordlists/chinese_simplified.json';
|
||||
const CHINESE_SIMPLIFIED_WORDLIST = require("../wordlists/chinese_simplified.json");
|
||||
const CHINESE_TRADITIONAL_WORDLIST = require("../wordlists/chinese_traditional.json");
|
||||
const ENGLISH_WORDLIST = require("../wordlists/english.json");
|
||||
const FRENCH_WORDLIST = require("../wordlists/french.json");
|
||||
const ITALIAN_WORDLIST = require("../wordlists/italian.json");
|
||||
const JAPANESE_WORDLIST = require("../wordlists/japanese.json");
|
||||
const KOREAN_WORDLIST = require("../wordlists/korean.json");
|
||||
const SPANISH_WORDLIST = require("../wordlists/spanish.json");
|
||||
const DEFAULT_WORDLIST = ENGLISH_WORDLIST;
|
||||
const INVALID_MNEMONIC = 'Invalid mnemonic';
|
||||
const INVALID_ENTROPY = 'Invalid entropy';
|
||||
const INVALID_CHECKSUM = 'Invalid mnemonic checksum';
|
||||
function lpad(str, padString, length) {
|
||||
while (str.length < length)
|
||||
str = padString + str;
|
||||
return str;
|
||||
}
|
||||
function binaryToByte(bin) {
|
||||
return parseInt(bin, 2);
|
||||
}
|
||||
function bytesToBinary(bytes) {
|
||||
return bytes
|
||||
.map(function (x) {
|
||||
return lpad(x.toString(2), '0', 8);
|
||||
})
|
||||
.join('');
|
||||
}
|
||||
function deriveChecksumBits(entropyBuffer) {
|
||||
var ENT = entropyBuffer.length * 8;
|
||||
var CS = ENT / 32;
|
||||
var hash = createHash('sha256')
|
||||
.update(entropyBuffer)
|
||||
.digest();
|
||||
return bytesToBinary([].slice.call(hash)).slice(0, CS);
|
||||
}
|
||||
function salt(password) {
|
||||
return 'mnemonic' + (password || '');
|
||||
}
|
||||
function mnemonicToSeed(mnemonic, password) {
|
||||
const mnemonicBuffer = Buffer.from(unorm.nfkd(mnemonic), 'utf8');
|
||||
const saltBuffer = Buffer.from(salt(unorm.nfkd(password)), 'utf8');
|
||||
return pbkdf2_1.pbkdf2Sync(mnemonicBuffer, saltBuffer, 2048, 64, 'sha512');
|
||||
}
|
||||
function mnemonicToSeedHex(mnemonic, password) {
|
||||
return mnemonicToSeed(mnemonic, password).toString('hex');
|
||||
}
|
||||
function mnemonicToSeedAsync(mnemonic, password) {
|
||||
return new Promise(function (resolve, reject) {
|
||||
try {
|
||||
var mnemonicBuffer = Buffer.from(unorm.nfkd(mnemonic), 'utf8');
|
||||
var saltBuffer = Buffer.from(salt(unorm.nfkd(password)), 'utf8');
|
||||
}
|
||||
catch (error) {
|
||||
return reject(error);
|
||||
}
|
||||
pbkdf2_1.pbkdf2(mnemonicBuffer, saltBuffer, 2048, 64, 'sha512', function (err, data) {
|
||||
if (err)
|
||||
return reject(err);
|
||||
else
|
||||
return resolve(data);
|
||||
});
|
||||
});
|
||||
}
|
||||
function mnemonicToSeedHexAsync(mnemonic, password) {
|
||||
return mnemonicToSeedAsync(mnemonic, password).then(function (buf) {
|
||||
return buf.toString('hex');
|
||||
});
|
||||
}
|
||||
function mnemonicToEntropy(mnemonic, wordlist) {
|
||||
wordlist = wordlist || DEFAULT_WORDLIST;
|
||||
var words = unorm.nfkd(mnemonic).split(' ');
|
||||
if (words.length % 3 !== 0)
|
||||
throw new Error(INVALID_MNEMONIC);
|
||||
// convert word indices to 11 bit binary strings
|
||||
var bits = words
|
||||
.map(function (word) {
|
||||
var index = wordlist.indexOf(word);
|
||||
if (index === -1)
|
||||
throw new Error(INVALID_MNEMONIC);
|
||||
return lpad(index.toString(2), '0', 11);
|
||||
})
|
||||
.join('');
|
||||
// split the binary string into ENT/CS
|
||||
var dividerIndex = Math.floor(bits.length / 33) * 32;
|
||||
var entropyBits = bits.slice(0, dividerIndex);
|
||||
var checksumBits = bits.slice(dividerIndex);
|
||||
// calculate the checksum and compare
|
||||
var entropyBytes = entropyBits.match(/(.{1,8})/g).map(binaryToByte);
|
||||
if (entropyBytes.length < 16)
|
||||
throw new Error(INVALID_ENTROPY);
|
||||
if (entropyBytes.length > 32)
|
||||
throw new Error(INVALID_ENTROPY);
|
||||
if (entropyBytes.length % 4 !== 0)
|
||||
throw new Error(INVALID_ENTROPY);
|
||||
var entropy = Buffer.from(entropyBytes);
|
||||
var newChecksum = deriveChecksumBits(entropy);
|
||||
if (newChecksum !== checksumBits)
|
||||
throw new Error(INVALID_CHECKSUM);
|
||||
return entropy.toString('hex');
|
||||
}
|
||||
function entropyToMnemonic(entropy, wordlist) {
|
||||
if (!Buffer.isBuffer(entropy))
|
||||
entropy = Buffer.from(entropy, 'hex');
|
||||
wordlist = wordlist || DEFAULT_WORDLIST;
|
||||
// 128 <= ENT <= 256
|
||||
if (entropy.length < 16)
|
||||
throw new TypeError(INVALID_ENTROPY);
|
||||
if (entropy.length > 32)
|
||||
throw new TypeError(INVALID_ENTROPY);
|
||||
if (entropy.length % 4 !== 0)
|
||||
throw new TypeError(INVALID_ENTROPY);
|
||||
var entropyBits = bytesToBinary([].slice.call(entropy));
|
||||
var checksumBits = deriveChecksumBits(entropy);
|
||||
var bits = entropyBits + checksumBits;
|
||||
var chunks = bits.match(/(.{1,11})/g);
|
||||
var words = chunks.map(function (binary) {
|
||||
var index = binaryToByte(binary);
|
||||
return wordlist[index];
|
||||
});
|
||||
return wordlist === JAPANESE_WORDLIST
|
||||
? words.join('\u3000')
|
||||
: words.join(' ');
|
||||
}
|
||||
function generateMnemonic(strength, rng, wordlist) {
|
||||
strength = strength || 128;
|
||||
if (strength % 32 !== 0)
|
||||
throw new TypeError(INVALID_ENTROPY);
|
||||
rng = rng || randomBytes;
|
||||
return entropyToMnemonic(rng(strength / 8), wordlist);
|
||||
}
|
||||
function validateMnemonic(mnemonic, wordlist) {
|
||||
try {
|
||||
mnemonicToEntropy(mnemonic, wordlist);
|
||||
}
|
||||
catch (e) {
|
||||
return false;
|
||||
}
|
||||
return true;
|
||||
}
|
||||
module.exports = {
|
||||
mnemonicToSeed: mnemonicToSeed,
|
||||
mnemonicToSeedAsync: mnemonicToSeedAsync,
|
||||
mnemonicToSeedHex: mnemonicToSeedHex,
|
||||
mnemonicToSeedHexAsync: mnemonicToSeedHexAsync,
|
||||
mnemonicToEntropy: mnemonicToEntropy,
|
||||
entropyToMnemonic: entropyToMnemonic,
|
||||
generateMnemonic: generateMnemonic,
|
||||
validateMnemonic: validateMnemonic,
|
||||
wordlists: {
|
||||
EN: ENGLISH_WORDLIST,
|
||||
JA: JAPANESE_WORDLIST,
|
||||
chinese_simplified: CHINESE_SIMPLIFIED_WORDLIST,
|
||||
chinese_traditional: CHINESE_TRADITIONAL_WORDLIST,
|
||||
english: ENGLISH_WORDLIST,
|
||||
french: FRENCH_WORDLIST,
|
||||
italian: ITALIAN_WORDLIST,
|
||||
japanese: JAPANESE_WORDLIST,
|
||||
korean: KOREAN_WORDLIST,
|
||||
spanish: SPANISH_WORDLIST,
|
||||
},
|
||||
};
|
Loading…
Reference in New Issue