fs-db.js 10 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339
  1. "use strict";
  2. var bodec = require('bodec');
  3. var inflate = require('../lib/inflate');
  4. var deflate = require('../lib/deflate');
  5. var codec = require('../lib/object-codec');
  6. var parsePackEntry = require('../lib/pack-codec').parseEntry;
  7. var applyDelta = require('../lib/apply-delta');
  8. var sha1 = require('git-sha1');
  9. var pathJoin = require('path').join;
  10. // The fs object has the following interface:
  11. // - readFile(path) => binary
  12. // Must also call callback() with no arguments if the file does not exist.
  13. // - readChunk(path, start, end) => binary
  14. // Must also call callback() with no arguments if the file does not exist.
  15. // - writeFile(path, binary) =>
  16. // Must also make every directory up to parent of path.
  17. // - readDir(path) => array<paths>
  18. // Must also call callback() with no arguments if the file does not exist.
  19. // The repo is expected to have a rootPath property that points to
  20. // the .git folder within the filesystem.
  21. module.exports = function (repo, fs) {
  22. var cachedIndexes = {};
  23. repo.loadAs = loadAs;
  24. repo.saveAs = saveAs;
  25. repo.loadRaw = loadRaw;
  26. repo.saveRaw = saveRaw;
  27. repo.readRef = readRef;
  28. repo.updateRef = updateRef;
  29. repo.hasHash = hasHash;
  30. repo.init = init;
  31. repo.setShallow = setShallow;
  32. function init(ref, callback) {
  33. if (!callback) return init.bind(null, ref);
  34. ref = ref || "refs/heads/master";
  35. var path = pathJoin(repo.rootPath, "HEAD");
  36. fs.writeFile(path, "ref: " + ref, callback);
  37. }
  38. function setShallow(ref, callback) {
  39. if (!callback) return setShallow.bind(null, ref);
  40. var path = pathJoin(repo.rootPath, "shallow");
  41. fs.writeFile(path, ref, callback);
  42. }
  43. function updateRef(ref, hash, callback) {
  44. if (!callback) return updateRef.bind(repo, ref, hash);
  45. var path = pathJoin(repo.rootPath, ref);
  46. var lock = path + ".lock";
  47. fs.writeFile(lock, bodec.fromRaw(hash + "\n"), function(err) {
  48. if(err) return callback(err);
  49. fs.rename(lock, path, callback);
  50. });
  51. }
  52. function readRef(ref, callback) {
  53. if (!callback) return readRef.bind(repo, ref);
  54. var path = pathJoin(repo.rootPath, ref);
  55. fs.readFile(path, function (err, binary) {
  56. if (err) return callback(err);
  57. if (binary === undefined) {
  58. return readPackedRef(ref, callback);
  59. }
  60. var hash;
  61. try { hash = bodec.toRaw(binary).trim(); }
  62. catch (err) { return callback(err); }
  63. callback(null, hash);
  64. });
  65. }
  66. function readPackedRef(ref, callback) {
  67. var path = pathJoin(repo.rootPath, "packed-refs");
  68. fs.readFile(path, function (err, binary) {
  69. if (binary === undefined) return callback(err);
  70. var hash;
  71. try {
  72. var text = bodec.toRaw(binary);
  73. var index = text.indexOf(ref);
  74. if (index >= 0) {
  75. hash = text.substring(index - 41, index - 1);
  76. }
  77. }
  78. catch (err) {
  79. return callback(err);
  80. }
  81. callback(null, hash);
  82. });
  83. }
  84. function saveAs(type, body, callback) {
  85. if (!callback) return saveAs.bind(repo, type, body);
  86. var raw, hash;
  87. try {
  88. raw = codec.frame({
  89. type: type,
  90. body: codec.encoders[type](body)
  91. });
  92. hash = sha1(raw);
  93. }
  94. catch (err) { return callback(err); }
  95. saveRaw(hash, raw, function (err) {
  96. if (err) return callback(err);
  97. callback(null, hash);
  98. });
  99. }
  100. function saveRaw(hash, raw, callback) {
  101. if (!callback) return saveRaw.bind(repo, hash, raw);
  102. var buffer, path;
  103. try {
  104. if (sha1(raw) !== hash) {
  105. throw new Error("Save data does not match hash");
  106. }
  107. buffer = deflate(raw);
  108. path = hashToPath(hash);
  109. }
  110. catch (err) { return callback(err); }
  111. // Try to read the object first.
  112. loadRaw(hash, function (err, data) {
  113. // If it already exists, we're done
  114. if (data) return callback();
  115. // Otherwise write a new file
  116. var tmp = path.replace(/[0-9a-f]+$/, 'tmp_obj_' + Math.random().toString(36).substr(2))
  117. fs.writeFile(tmp, buffer, function(err) {
  118. if(err) return callback(err);
  119. fs.rename(tmp, path, callback);
  120. });
  121. });
  122. }
  123. function loadAs(type, hash, callback) {
  124. if (!callback) return loadAs.bind(repo, type, hash);
  125. loadRaw(hash, function (err, raw) {
  126. if (raw === undefined) return callback(err);
  127. var body;
  128. try {
  129. raw = codec.deframe(raw);
  130. if (raw.type !== type) throw new TypeError("Type mismatch");
  131. body = codec.decoders[raw.type](raw.body);
  132. }
  133. catch (err) { return callback(err); }
  134. callback(null, body);
  135. });
  136. }
  137. function hasHash(hash, callback) {
  138. if (!callback) return hasHash.bind(repo, hash);
  139. loadRaw(hash, function (err, body) {
  140. if (err) return callback(err);
  141. return callback(null, !!body);
  142. });
  143. }
  144. function loadRaw(hash, callback) {
  145. if (!callback) return loadRaw.bind(repo, hash);
  146. var path = hashToPath(hash);
  147. fs.readFile(path, function (err, buffer) {
  148. if (err) return callback(err);
  149. if (buffer) {
  150. var raw;
  151. try { raw = inflate(buffer); }
  152. catch (err) { return callback(err); }
  153. return callback(null, raw);
  154. }
  155. return loadRawPacked(hash, callback);
  156. });
  157. }
  158. function loadRawPacked(hash, callback) {
  159. var packDir = pathJoin(repo.rootPath, "objects/pack");
  160. var packHashes = [];
  161. fs.readDir(packDir, function (err, entries) {
  162. if (!entries) return callback(err);
  163. entries.forEach(function (name) {
  164. var match = name.match(/pack-([0-9a-f]{40}).idx/);
  165. if (match) packHashes.push(match[1]);
  166. });
  167. start();
  168. });
  169. function start() {
  170. var packHash = packHashes.pop();
  171. var offsets;
  172. if (!packHash) return callback();
  173. if (!cachedIndexes[packHash]) loadIndex(packHash);
  174. else onIndex();
  175. function loadIndex() {
  176. var indexFile = pathJoin(packDir, "pack-" + packHash + ".idx" );
  177. fs.readFile(indexFile, function (err, buffer) {
  178. if (!buffer) return callback(err);
  179. try {
  180. cachedIndexes[packHash] = parseIndex(buffer);
  181. }
  182. catch (err) { return callback(err); }
  183. onIndex();
  184. });
  185. }
  186. function onIndex() {
  187. var cached = cachedIndexes[packHash];
  188. var packFile = pathJoin(packDir, "pack-" + packHash + ".pack" );
  189. var index = cached.byHash[hash];
  190. if (!index) return start();
  191. offsets = cached.offsets;
  192. loadChunk(packFile, index.offset, callback);
  193. }
  194. function loadChunk(packFile, start, callback) {
  195. var index = offsets.indexOf(start);
  196. if (index < 0) {
  197. var error = new Error("Can't find chunk starting at " + start);
  198. return callback(error);
  199. }
  200. var end = index + 1 < offsets.length ? offsets[index + 1] : -20;
  201. fs.readChunk(packFile, start, end, function (err, chunk) {
  202. if (!chunk) return callback(err);
  203. var raw;
  204. try {
  205. var entry = parsePackEntry(chunk);
  206. if (entry.type === "ref-delta") {
  207. return loadRaw.call(repo, entry.ref, onBase);
  208. }
  209. else if (entry.type === "ofs-delta") {
  210. return loadChunk(packFile, start - entry.ref, onBase);
  211. }
  212. raw = codec.frame(entry);
  213. }
  214. catch (err) { return callback(err); }
  215. callback(null, raw);
  216. function onBase(err, base) {
  217. if (!base) return callback(err);
  218. var object = codec.deframe(base);
  219. var buffer;
  220. try {
  221. object.body = applyDelta(entry.body, object.body);
  222. buffer = codec.frame(object);
  223. }
  224. catch (err) { return callback(err); }
  225. callback(null, buffer);
  226. }
  227. });
  228. }
  229. }
  230. }
  231. function hashToPath(hash) {
  232. return pathJoin(repo.rootPath, "objects", hash.substring(0, 2), hash.substring(2));
  233. }
  234. };
  235. function parseIndex(buffer) {
  236. if (readUint32(buffer, 0) !== 0xff744f63 ||
  237. readUint32(buffer, 4) !== 0x00000002) {
  238. throw new Error("Only v2 pack indexes supported");
  239. }
  240. // Get the number of hashes in index
  241. // This is the value of the last fan-out entry
  242. var hashOffset = 8 + 255 * 4;
  243. var length = readUint32(buffer, hashOffset);
  244. hashOffset += 4;
  245. var crcOffset = hashOffset + 20 * length;
  246. var lengthOffset = crcOffset + 4 * length;
  247. var largeOffset = lengthOffset + 4 * length;
  248. var checkOffset = largeOffset;
  249. var indexes = new Array(length);
  250. for (var i = 0; i < length; i++) {
  251. var start = hashOffset + i * 20;
  252. var hash = bodec.toHex(bodec.slice(buffer, start, start + 20));
  253. var crc = readUint32(buffer, crcOffset + i * 4);
  254. var offset = readUint32(buffer, lengthOffset + i * 4);
  255. if (offset & 0x80000000) {
  256. offset = largeOffset + (offset &0x7fffffff) * 8;
  257. checkOffset = Math.max(checkOffset, offset + 8);
  258. offset = readUint64(buffer, offset);
  259. }
  260. indexes[i] = {
  261. hash: hash,
  262. offset: offset,
  263. crc: crc
  264. };
  265. }
  266. var packChecksum = bodec.toHex(bodec.slice(buffer, checkOffset, checkOffset + 20));
  267. var checksum = bodec.toHex(bodec.slice(buffer, checkOffset + 20, checkOffset + 40));
  268. if (sha1(bodec.slice(buffer, 0, checkOffset + 20)) !== checksum) {
  269. throw new Error("Checksum mistmatch");
  270. }
  271. var byHash = {};
  272. indexes.sort(function (a, b) {
  273. return a.offset - b.offset;
  274. });
  275. indexes.forEach(function (data) {
  276. byHash[data.hash] = {
  277. offset: data.offset,
  278. crc: data.crc,
  279. };
  280. });
  281. var offsets = indexes.map(function (entry) {
  282. return entry.offset;
  283. }).sort(function (a, b) {
  284. return a - b;
  285. });
  286. return {
  287. offsets: offsets,
  288. byHash: byHash,
  289. checksum: packChecksum
  290. };
  291. }
  292. function readUint32(buffer, offset) {
  293. return (buffer[offset] << 24 |
  294. buffer[offset + 1] << 16 |
  295. buffer[offset + 2] << 8 |
  296. buffer[offset + 3] << 0) >>> 0;
  297. }
  298. // Yes this will lose precision over 2^53, but that can't be helped when
  299. // returning a single integer.
  300. // We simply won't support packfiles over 8 petabytes. I'm ok with that.
  301. function readUint64(buffer, offset) {
  302. var hi = (buffer[offset] << 24 |
  303. buffer[offset + 1] << 16 |
  304. buffer[offset + 2] << 8 |
  305. buffer[offset + 3] << 0) >>> 0;
  306. var lo = (buffer[offset + 4] << 24 |
  307. buffer[offset + 5] << 16 |
  308. buffer[offset + 6] << 8 |
  309. buffer[offset + 7] << 0) >>> 0;
  310. return hi * 0x100000000 + lo;
  311. }