gdal-async
Version:
Bindings to GDAL (Geospatial Data Abstraction Library) with full async support
231 lines • 7.86 kB
JavaScript
var __importDefault = (this && this.__importDefault) || function (mod) {
return (mod && mod.__esModule) ? mod : { "default": mod };
};
Object.defineProperty(exports, "__esModule", { value: true });
exports.replace = void 0;
// tar -r
const fs_minipass_1 = require("@isaacs/fs-minipass");
const node_fs_1 = __importDefault(require("node:fs"));
const node_path_1 = __importDefault(require("node:path"));
const header_js_1 = require("./header.js");
const list_js_1 = require("./list.js");
const make_command_js_1 = require("./make-command.js");
const options_js_1 = require("./options.js");
const pack_js_1 = require("./pack.js");
// starting at the head of the file, read a Header
// If the checksum is invalid, that's our position to start writing
// If it is, jump forward by the specified size (round up to 512)
// and try again.
// Write the new Pack stream starting there.
const replaceSync = (opt, files) => {
const p = new pack_js_1.PackSync(opt);
let threw = true;
let fd;
let position;
try {
try {
fd = node_fs_1.default.openSync(opt.file, 'r+');
}
catch (er) {
if (er?.code === 'ENOENT') {
fd = node_fs_1.default.openSync(opt.file, 'w+');
}
else {
throw er;
}
}
const st = node_fs_1.default.fstatSync(fd);
const headBuf = Buffer.alloc(512);
POSITION: for (position = 0; position < st.size; position += 512) {
for (let bufPos = 0, bytes = 0; bufPos < 512; bufPos += bytes) {
bytes = node_fs_1.default.readSync(fd, headBuf, bufPos, headBuf.length - bufPos, position + bufPos);
if (position === 0 &&
headBuf[0] === 0x1f &&
headBuf[1] === 0x8b) {
throw new Error('cannot append to compressed archives');
}
if (!bytes) {
break POSITION;
}
}
const h = new header_js_1.Header(headBuf);
if (!h.cksumValid) {
break;
}
const entryBlockSize = 512 * Math.ceil((h.size || 0) / 512);
if (position + entryBlockSize + 512 > st.size) {
break;
}
// the 512 for the header we just parsed will be added as well
// also jump ahead all the blocks for the body
position += entryBlockSize;
if (opt.mtimeCache && h.mtime) {
opt.mtimeCache.set(String(h.path), h.mtime);
}
}
threw = false;
streamSync(opt, p, position, fd, files);
}
finally {
if (threw) {
try {
node_fs_1.default.closeSync(fd);
}
catch (er) { }
}
}
};
const streamSync = (opt, p, position, fd, files) => {
const stream = new fs_minipass_1.WriteStreamSync(opt.file, {
fd: fd,
start: position,
});
p.pipe(stream);
addFilesSync(p, files);
};
const replaceAsync = (opt, files) => {
files = Array.from(files);
const p = new pack_js_1.Pack(opt);
const getPos = (fd, size, cb_) => {
const cb = (er, pos) => {
if (er) {
node_fs_1.default.close(fd, _ => cb_(er));
}
else {
cb_(null, pos);
}
};
let position = 0;
if (size === 0) {
return cb(null, 0);
}
let bufPos = 0;
const headBuf = Buffer.alloc(512);
const onread = (er, bytes) => {
if (er || typeof bytes === 'undefined') {
return cb(er);
}
bufPos += bytes;
if (bufPos < 512 && bytes) {
return node_fs_1.default.read(fd, headBuf, bufPos, headBuf.length - bufPos, position + bufPos, onread);
}
if (position === 0 &&
headBuf[0] === 0x1f &&
headBuf[1] === 0x8b) {
return cb(new Error('cannot append to compressed archives'));
}
// truncated header
if (bufPos < 512) {
return cb(null, position);
}
const h = new header_js_1.Header(headBuf);
if (!h.cksumValid) {
return cb(null, position);
}
/* c8 ignore next */
const entryBlockSize = 512 * Math.ceil((h.size ?? 0) / 512);
if (position + entryBlockSize + 512 > size) {
return cb(null, position);
}
position += entryBlockSize + 512;
if (position >= size) {
return cb(null, position);
}
if (opt.mtimeCache && h.mtime) {
opt.mtimeCache.set(String(h.path), h.mtime);
}
bufPos = 0;
node_fs_1.default.read(fd, headBuf, 0, 512, position, onread);
};
node_fs_1.default.read(fd, headBuf, 0, 512, position, onread);
};
const promise = new Promise((resolve, reject) => {
p.on('error', reject);
let flag = 'r+';
const onopen = (er, fd) => {
if (er && er.code === 'ENOENT' && flag === 'r+') {
flag = 'w+';
return node_fs_1.default.open(opt.file, flag, onopen);
}
if (er || !fd) {
return reject(er);
}
node_fs_1.default.fstat(fd, (er, st) => {
if (er) {
return node_fs_1.default.close(fd, () => reject(er));
}
getPos(fd, st.size, (er, position) => {
if (er) {
return reject(er);
}
const stream = new fs_minipass_1.WriteStream(opt.file, {
fd: fd,
start: position,
});
p.pipe(stream);
stream.on('error', reject);
stream.on('close', resolve);
addFilesAsync(p, files);
});
});
};
node_fs_1.default.open(opt.file, flag, onopen);
});
return promise;
};
const addFilesSync = (p, files) => {
files.forEach(file => {
if (file.charAt(0) === '@') {
(0, list_js_1.list)({
file: node_path_1.default.resolve(p.cwd, file.slice(1)),
sync: true,
noResume: true,
onReadEntry: entry => p.add(entry),
});
}
else {
p.add(file);
}
});
p.end();
};
const addFilesAsync = async (p, files) => {
for (let i = 0; i < files.length; i++) {
const file = String(files[i]);
if (file.charAt(0) === '@') {
await (0, list_js_1.list)({
file: node_path_1.default.resolve(String(p.cwd), file.slice(1)),
noResume: true,
onReadEntry: entry => p.add(entry),
});
}
else {
p.add(file);
}
}
p.end();
};
exports.replace = (0, make_command_js_1.makeCommand)(replaceSync, replaceAsync,
/* c8 ignore start */
() => {
throw new TypeError('file is required');
}, () => {
throw new TypeError('file is required');
},
/* c8 ignore stop */
(opt, entries) => {
if (!(0, options_js_1.isFile)(opt)) {
throw new TypeError('file is required');
}
if (opt.gzip ||
opt.brotli ||
opt.file.endsWith('.br') ||
opt.file.endsWith('.tbr')) {
throw new TypeError('cannot append to compressed archives');
}
if (!entries?.length) {
throw new TypeError('no paths specified to add/replace');
}
});
//# sourceMappingURL=replace.js.map
;