231 lines
		
	
	
		
			7.7 KiB
		
	
	
	
		
			JavaScript
		
	
	
			
		
		
	
	
			231 lines
		
	
	
		
			7.7 KiB
		
	
	
	
		
			JavaScript
		
	
	
"use strict";
 | 
						|
var __importDefault = (this && this.__importDefault) || function (mod) {
 | 
						|
    return (mod && mod.__esModule) ? mod : { "default": mod };
 | 
						|
};
 | 
						|
Object.defineProperty(exports, "__esModule", { value: true });
 | 
						|
exports.replace = void 0;
 | 
						|
// tar -r
 | 
						|
const fs_minipass_1 = require("@isaacs/fs-minipass");
 | 
						|
const node_fs_1 = __importDefault(require("node:fs"));
 | 
						|
const node_path_1 = __importDefault(require("node:path"));
 | 
						|
const header_js_1 = require("./header.js");
 | 
						|
const list_js_1 = require("./list.js");
 | 
						|
const make_command_js_1 = require("./make-command.js");
 | 
						|
const options_js_1 = require("./options.js");
 | 
						|
const pack_js_1 = require("./pack.js");
 | 
						|
// starting at the head of the file, read a Header
 | 
						|
// If the checksum is invalid, that's our position to start writing
 | 
						|
// If it is, jump forward by the specified size (round up to 512)
 | 
						|
// and try again.
 | 
						|
// Write the new Pack stream starting there.
 | 
						|
const replaceSync = (opt, files) => {
 | 
						|
    const p = new pack_js_1.PackSync(opt);
 | 
						|
    let threw = true;
 | 
						|
    let fd;
 | 
						|
    let position;
 | 
						|
    try {
 | 
						|
        try {
 | 
						|
            fd = node_fs_1.default.openSync(opt.file, 'r+');
 | 
						|
        }
 | 
						|
        catch (er) {
 | 
						|
            if (er?.code === 'ENOENT') {
 | 
						|
                fd = node_fs_1.default.openSync(opt.file, 'w+');
 | 
						|
            }
 | 
						|
            else {
 | 
						|
                throw er;
 | 
						|
            }
 | 
						|
        }
 | 
						|
        const st = node_fs_1.default.fstatSync(fd);
 | 
						|
        const headBuf = Buffer.alloc(512);
 | 
						|
        POSITION: for (position = 0; position < st.size; position += 512) {
 | 
						|
            for (let bufPos = 0, bytes = 0; bufPos < 512; bufPos += bytes) {
 | 
						|
                bytes = node_fs_1.default.readSync(fd, headBuf, bufPos, headBuf.length - bufPos, position + bufPos);
 | 
						|
                if (position === 0 &&
 | 
						|
                    headBuf[0] === 0x1f &&
 | 
						|
                    headBuf[1] === 0x8b) {
 | 
						|
                    throw new Error('cannot append to compressed archives');
 | 
						|
                }
 | 
						|
                if (!bytes) {
 | 
						|
                    break POSITION;
 | 
						|
                }
 | 
						|
            }
 | 
						|
            const h = new header_js_1.Header(headBuf);
 | 
						|
            if (!h.cksumValid) {
 | 
						|
                break;
 | 
						|
            }
 | 
						|
            const entryBlockSize = 512 * Math.ceil((h.size || 0) / 512);
 | 
						|
            if (position + entryBlockSize + 512 > st.size) {
 | 
						|
                break;
 | 
						|
            }
 | 
						|
            // the 512 for the header we just parsed will be added as well
 | 
						|
            // also jump ahead all the blocks for the body
 | 
						|
            position += entryBlockSize;
 | 
						|
            if (opt.mtimeCache && h.mtime) {
 | 
						|
                opt.mtimeCache.set(String(h.path), h.mtime);
 | 
						|
            }
 | 
						|
        }
 | 
						|
        threw = false;
 | 
						|
        streamSync(opt, p, position, fd, files);
 | 
						|
    }
 | 
						|
    finally {
 | 
						|
        if (threw) {
 | 
						|
            try {
 | 
						|
                node_fs_1.default.closeSync(fd);
 | 
						|
            }
 | 
						|
            catch (er) { }
 | 
						|
        }
 | 
						|
    }
 | 
						|
};
 | 
						|
const streamSync = (opt, p, position, fd, files) => {
 | 
						|
    const stream = new fs_minipass_1.WriteStreamSync(opt.file, {
 | 
						|
        fd: fd,
 | 
						|
        start: position,
 | 
						|
    });
 | 
						|
    p.pipe(stream);
 | 
						|
    addFilesSync(p, files);
 | 
						|
};
 | 
						|
const replaceAsync = (opt, files) => {
 | 
						|
    files = Array.from(files);
 | 
						|
    const p = new pack_js_1.Pack(opt);
 | 
						|
    const getPos = (fd, size, cb_) => {
 | 
						|
        const cb = (er, pos) => {
 | 
						|
            if (er) {
 | 
						|
                node_fs_1.default.close(fd, _ => cb_(er));
 | 
						|
            }
 | 
						|
            else {
 | 
						|
                cb_(null, pos);
 | 
						|
            }
 | 
						|
        };
 | 
						|
        let position = 0;
 | 
						|
        if (size === 0) {
 | 
						|
            return cb(null, 0);
 | 
						|
        }
 | 
						|
        let bufPos = 0;
 | 
						|
        const headBuf = Buffer.alloc(512);
 | 
						|
        const onread = (er, bytes) => {
 | 
						|
            if (er || typeof bytes === 'undefined') {
 | 
						|
                return cb(er);
 | 
						|
            }
 | 
						|
            bufPos += bytes;
 | 
						|
            if (bufPos < 512 && bytes) {
 | 
						|
                return node_fs_1.default.read(fd, headBuf, bufPos, headBuf.length - bufPos, position + bufPos, onread);
 | 
						|
            }
 | 
						|
            if (position === 0 &&
 | 
						|
                headBuf[0] === 0x1f &&
 | 
						|
                headBuf[1] === 0x8b) {
 | 
						|
                return cb(new Error('cannot append to compressed archives'));
 | 
						|
            }
 | 
						|
            // truncated header
 | 
						|
            if (bufPos < 512) {
 | 
						|
                return cb(null, position);
 | 
						|
            }
 | 
						|
            const h = new header_js_1.Header(headBuf);
 | 
						|
            if (!h.cksumValid) {
 | 
						|
                return cb(null, position);
 | 
						|
            }
 | 
						|
            /* c8 ignore next */
 | 
						|
            const entryBlockSize = 512 * Math.ceil((h.size ?? 0) / 512);
 | 
						|
            if (position + entryBlockSize + 512 > size) {
 | 
						|
                return cb(null, position);
 | 
						|
            }
 | 
						|
            position += entryBlockSize + 512;
 | 
						|
            if (position >= size) {
 | 
						|
                return cb(null, position);
 | 
						|
            }
 | 
						|
            if (opt.mtimeCache && h.mtime) {
 | 
						|
                opt.mtimeCache.set(String(h.path), h.mtime);
 | 
						|
            }
 | 
						|
            bufPos = 0;
 | 
						|
            node_fs_1.default.read(fd, headBuf, 0, 512, position, onread);
 | 
						|
        };
 | 
						|
        node_fs_1.default.read(fd, headBuf, 0, 512, position, onread);
 | 
						|
    };
 | 
						|
    const promise = new Promise((resolve, reject) => {
 | 
						|
        p.on('error', reject);
 | 
						|
        let flag = 'r+';
 | 
						|
        const onopen = (er, fd) => {
 | 
						|
            if (er && er.code === 'ENOENT' && flag === 'r+') {
 | 
						|
                flag = 'w+';
 | 
						|
                return node_fs_1.default.open(opt.file, flag, onopen);
 | 
						|
            }
 | 
						|
            if (er || !fd) {
 | 
						|
                return reject(er);
 | 
						|
            }
 | 
						|
            node_fs_1.default.fstat(fd, (er, st) => {
 | 
						|
                if (er) {
 | 
						|
                    return node_fs_1.default.close(fd, () => reject(er));
 | 
						|
                }
 | 
						|
                getPos(fd, st.size, (er, position) => {
 | 
						|
                    if (er) {
 | 
						|
                        return reject(er);
 | 
						|
                    }
 | 
						|
                    const stream = new fs_minipass_1.WriteStream(opt.file, {
 | 
						|
                        fd: fd,
 | 
						|
                        start: position,
 | 
						|
                    });
 | 
						|
                    p.pipe(stream);
 | 
						|
                    stream.on('error', reject);
 | 
						|
                    stream.on('close', resolve);
 | 
						|
                    addFilesAsync(p, files);
 | 
						|
                });
 | 
						|
            });
 | 
						|
        };
 | 
						|
        node_fs_1.default.open(opt.file, flag, onopen);
 | 
						|
    });
 | 
						|
    return promise;
 | 
						|
};
 | 
						|
const addFilesSync = (p, files) => {
 | 
						|
    files.forEach(file => {
 | 
						|
        if (file.charAt(0) === '@') {
 | 
						|
            (0, list_js_1.list)({
 | 
						|
                file: node_path_1.default.resolve(p.cwd, file.slice(1)),
 | 
						|
                sync: true,
 | 
						|
                noResume: true,
 | 
						|
                onReadEntry: entry => p.add(entry),
 | 
						|
            });
 | 
						|
        }
 | 
						|
        else {
 | 
						|
            p.add(file);
 | 
						|
        }
 | 
						|
    });
 | 
						|
    p.end();
 | 
						|
};
 | 
						|
const addFilesAsync = async (p, files) => {
 | 
						|
    for (let i = 0; i < files.length; i++) {
 | 
						|
        const file = String(files[i]);
 | 
						|
        if (file.charAt(0) === '@') {
 | 
						|
            await (0, list_js_1.list)({
 | 
						|
                file: node_path_1.default.resolve(String(p.cwd), file.slice(1)),
 | 
						|
                noResume: true,
 | 
						|
                onReadEntry: entry => p.add(entry),
 | 
						|
            });
 | 
						|
        }
 | 
						|
        else {
 | 
						|
            p.add(file);
 | 
						|
        }
 | 
						|
    }
 | 
						|
    p.end();
 | 
						|
};
 | 
						|
exports.replace = (0, make_command_js_1.makeCommand)(replaceSync, replaceAsync, 
 | 
						|
/* c8 ignore start */
 | 
						|
() => {
 | 
						|
    throw new TypeError('file is required');
 | 
						|
}, () => {
 | 
						|
    throw new TypeError('file is required');
 | 
						|
}, 
 | 
						|
/* c8 ignore stop */
 | 
						|
(opt, entries) => {
 | 
						|
    if (!(0, options_js_1.isFile)(opt)) {
 | 
						|
        throw new TypeError('file is required');
 | 
						|
    }
 | 
						|
    if (opt.gzip ||
 | 
						|
        opt.brotli ||
 | 
						|
        opt.file.endsWith('.br') ||
 | 
						|
        opt.file.endsWith('.tbr')) {
 | 
						|
        throw new TypeError('cannot append to compressed archives');
 | 
						|
    }
 | 
						|
    if (!entries?.length) {
 | 
						|
        throw new TypeError('no paths specified to add/replace');
 | 
						|
    }
 | 
						|
});
 | 
						|
//# sourceMappingURL=replace.js.map
 |