mirror of https://git.tuxpa.in/a/code-server.git
Fix install from VSIX for TAR and ZIP formats (#245)
* Fix install from VSIX for TAR and ZIP formats * Parse TAR before ZIP, when installing from VSIX
This commit is contained in:
parent
75435be949
commit
18f395b853
|
@ -12,6 +12,11 @@
|
||||||
"xmlhttprequest": "1.8.0"
|
"xmlhttprequest": "1.8.0"
|
||||||
},
|
},
|
||||||
"jest": {
|
"jest": {
|
||||||
|
"globals": {
|
||||||
|
"ts-jest": {
|
||||||
|
"diagnostics": false
|
||||||
|
}
|
||||||
|
},
|
||||||
"moduleFileExtensions": [
|
"moduleFileExtensions": [
|
||||||
"ts",
|
"ts",
|
||||||
"tsx",
|
"tsx",
|
||||||
|
@ -26,7 +31,9 @@
|
||||||
"@coder/ide/src/fill/evaluation": "<rootDir>/ide/src/fill/evaluation",
|
"@coder/ide/src/fill/evaluation": "<rootDir>/ide/src/fill/evaluation",
|
||||||
"@coder/ide/src/fill/client": "<rootDir>/ide/src/fill/client",
|
"@coder/ide/src/fill/client": "<rootDir>/ide/src/fill/client",
|
||||||
"@coder/(.*)/test": "<rootDir>/$1/test",
|
"@coder/(.*)/test": "<rootDir>/$1/test",
|
||||||
"@coder/(.*)": "<rootDir>/$1/src"
|
"@coder/(.*)": "<rootDir>/$1/src",
|
||||||
|
"vs/(.*)": "<rootDir>/../lib/vscode/src/vs/$1",
|
||||||
|
"vszip": "<rootDir>/../lib/vscode/src/vs/base/node/zip.ts"
|
||||||
},
|
},
|
||||||
"transform": {
|
"transform": {
|
||||||
"^.+\\.tsx?$": "ts-jest"
|
"^.+\\.tsx?$": "ts-jest"
|
||||||
|
|
|
@ -1 +1,2 @@
|
||||||
bin
|
bin
|
||||||
|
test/.test*
|
|
@ -4,11 +4,11 @@
|
||||||
*--------------------------------------------------------------------------------------------*/
|
*--------------------------------------------------------------------------------------------*/
|
||||||
|
|
||||||
import * as nls from "vs/nls";
|
import * as nls from "vs/nls";
|
||||||
|
import * as vszip from "vszip";
|
||||||
import * as fs from "fs";
|
import * as fs from "fs";
|
||||||
import * as path from "path";
|
import * as path from "path";
|
||||||
import * as tarStream from "tar-stream";
|
import * as tarStream from "tar-stream";
|
||||||
import { promisify } from "util";
|
import { promisify } from "util";
|
||||||
import { ILogService } from "vs/platform/log/common/log";
|
|
||||||
import { CancellationToken } from "vs/base/common/cancellation";
|
import { CancellationToken } from "vs/base/common/cancellation";
|
||||||
import { mkdirp } from "vs/base/node/pfs";
|
import { mkdirp } from "vs/base/node/pfs";
|
||||||
|
|
||||||
|
@ -16,8 +16,8 @@ export interface IExtractOptions {
|
||||||
overwrite?: boolean;
|
overwrite?: boolean;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Source path within the ZIP archive. Only the files contained in this
|
* Source path within the TAR/ZIP archive. Only the files
|
||||||
* path will be extracted.
|
* contained in this path will be extracted.
|
||||||
*/
|
*/
|
||||||
sourcePath?: string;
|
sourcePath?: string;
|
||||||
}
|
}
|
||||||
|
@ -28,11 +28,15 @@ export interface IFile {
|
||||||
localPath?: string;
|
localPath?: string;
|
||||||
}
|
}
|
||||||
|
|
||||||
export function zip(tarPath: string, files: IFile[]): Promise<string> {
|
/**
|
||||||
return new Promise<string>((c, e) => {
|
* Override the standard VS Code behavior for zipping
|
||||||
|
* extensions to use the TAR format instead of ZIP.
|
||||||
|
*/
|
||||||
|
export const zip = (tarPath: string, files: IFile[]): Promise<string> => {
|
||||||
|
return new Promise<string>((c, e): void => {
|
||||||
const pack = tarStream.pack();
|
const pack = tarStream.pack();
|
||||||
const chunks: Buffer[] = [];
|
const chunks: Buffer[] = [];
|
||||||
const ended = new Promise<Buffer>((res, rej) => {
|
const ended = new Promise<Buffer>((res): void => {
|
||||||
pack.on("end", () => {
|
pack.on("end", () => {
|
||||||
res(Buffer.concat(chunks));
|
res(Buffer.concat(chunks));
|
||||||
});
|
});
|
||||||
|
@ -56,17 +60,112 @@ export function zip(tarPath: string, files: IFile[]): Promise<string> {
|
||||||
e(ex);
|
e(ex);
|
||||||
});
|
});
|
||||||
});
|
});
|
||||||
|
};
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Override the standard VS Code behavior for extracting
|
||||||
|
* archives, to first attempt to process the archive as a TAR
|
||||||
|
* and then fallback on the original implementation, for processing
|
||||||
|
* ZIPs.
|
||||||
|
*/
|
||||||
|
export const extract = (archivePath: string, extractPath: string, options: IExtractOptions = {}, token: CancellationToken): Promise<void> => {
|
||||||
|
return new Promise<void>((c, e): void => {
|
||||||
|
extractTar(archivePath, extractPath, options, token).then(c).catch((ex) => {
|
||||||
|
if (!ex.toString().includes("Invalid tar header")) {
|
||||||
|
e(ex);
|
||||||
|
|
||||||
|
return;
|
||||||
}
|
}
|
||||||
|
vszip.extract(archivePath, extractPath, options, token).then(c).catch(e);
|
||||||
|
});
|
||||||
|
});
|
||||||
|
};
|
||||||
|
|
||||||
export async function extract(tarPath: string, targetPath: string, options: IExtractOptions = {}, token: CancellationToken): Promise<void> {
|
/**
|
||||||
const sourcePathRegex = new RegExp(options.sourcePath ? `^${options.sourcePath}` : '');
|
* Override the standard VS Code behavior for buffering
|
||||||
|
* archives, to first process the Buffer as a TAR and then
|
||||||
|
* fallback on the original implementation, for processing ZIPs.
|
||||||
|
*/
|
||||||
|
export const buffer = (targetPath: string, filePath: string): Promise<Buffer> => {
|
||||||
|
return new Promise<Buffer>((c, e): void => {
|
||||||
|
let done: boolean = false;
|
||||||
|
extractAssets(targetPath, new RegExp(filePath), (assetPath: string, data: Buffer) => {
|
||||||
|
if (path.normalize(assetPath) === path.normalize(filePath)) {
|
||||||
|
done = true;
|
||||||
|
c(data);
|
||||||
|
}
|
||||||
|
}).then(() => {
|
||||||
|
if (!done) {
|
||||||
|
e("couldn't find asset " + filePath);
|
||||||
|
}
|
||||||
|
}).catch((ex) => {
|
||||||
|
if (!ex.toString().includes("Invalid tar header")) {
|
||||||
|
e(ex);
|
||||||
|
|
||||||
return new Promise<void>(async (c, e) => {
|
return;
|
||||||
|
}
|
||||||
|
vszip.buffer(targetPath, filePath).then(c).catch(e);
|
||||||
|
});
|
||||||
|
});
|
||||||
|
};
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Override the standard VS Code behavior for extracting assets
|
||||||
|
* from archive Buffers to use the TAR format instead of ZIP.
|
||||||
|
*/
|
||||||
|
export const extractAssets = (tarPath: string, match: RegExp, callback: (path: string, data: Buffer) => void): Promise<void> => {
|
||||||
|
return new Promise<void>(async (c, e): Promise<void> => {
|
||||||
|
try {
|
||||||
const buffer = await promisify(fs.readFile)(tarPath);
|
const buffer = await promisify(fs.readFile)(tarPath);
|
||||||
const extractor = tarStream.extract();
|
const extractor = tarStream.extract();
|
||||||
extractor.once('error', e);
|
extractor.once("error", e);
|
||||||
extractor.on('entry', (header, stream, next) => {
|
extractor.on("entry", (header, stream, next) => {
|
||||||
const rawName = header.name;
|
const name = header.name;
|
||||||
|
if (match.test(name)) {
|
||||||
|
extractData(stream).then((data) => {
|
||||||
|
callback(name, data);
|
||||||
|
next();
|
||||||
|
}).catch(e);
|
||||||
|
stream.resume();
|
||||||
|
} else {
|
||||||
|
stream.on("end", () => {
|
||||||
|
next();
|
||||||
|
});
|
||||||
|
stream.resume();
|
||||||
|
}
|
||||||
|
});
|
||||||
|
extractor.on("finish", () => {
|
||||||
|
c();
|
||||||
|
});
|
||||||
|
extractor.write(buffer);
|
||||||
|
extractor.end();
|
||||||
|
} catch (ex) {
|
||||||
|
e(ex);
|
||||||
|
}
|
||||||
|
});
|
||||||
|
};
|
||||||
|
|
||||||
|
const extractData = (stream: NodeJS.ReadableStream): Promise<Buffer> => {
|
||||||
|
return new Promise<Buffer>((c, e): void => {
|
||||||
|
const fileData: Buffer[] = [];
|
||||||
|
stream.on("data", (data) => fileData.push(data));
|
||||||
|
stream.on("end", () => {
|
||||||
|
const fd = Buffer.concat(fileData);
|
||||||
|
c(fd);
|
||||||
|
});
|
||||||
|
stream.on("error", e);
|
||||||
|
});
|
||||||
|
};
|
||||||
|
|
||||||
|
const extractTar = (tarPath: string, targetPath: string, options: IExtractOptions = {}, token: CancellationToken): Promise<void> => {
|
||||||
|
return new Promise<void>(async (c, e): Promise<void> => {
|
||||||
|
try {
|
||||||
|
const sourcePathRegex = new RegExp(options.sourcePath ? `^${options.sourcePath}` : "");
|
||||||
|
const buffer = await promisify(fs.readFile)(tarPath);
|
||||||
|
const extractor = tarStream.extract();
|
||||||
|
extractor.once("error", e);
|
||||||
|
extractor.on("entry", (header, stream, next) => {
|
||||||
|
const rawName = path.normalize(header.name);
|
||||||
|
|
||||||
const nextEntry = (): void => {
|
const nextEntry = (): void => {
|
||||||
stream.resume();
|
stream.resume();
|
||||||
|
@ -81,107 +180,40 @@ export async function extract(tarPath: string, targetPath: string, options: IExt
|
||||||
return nextEntry();
|
return nextEntry();
|
||||||
}
|
}
|
||||||
|
|
||||||
const fileName = rawName.replace(sourcePathRegex, '');
|
const fileName = rawName.replace(sourcePathRegex, "");
|
||||||
|
|
||||||
const targetFileName = path.join(targetPath, fileName);
|
const targetFileName = path.join(targetPath, fileName);
|
||||||
if (/\/$/.test(fileName)) {
|
if (/\/$/.test(fileName)) {
|
||||||
stream.resume();
|
stream.resume();
|
||||||
mkdirp(targetFileName).then(() => {
|
mkdirp(targetFileName).then(() => {
|
||||||
next();
|
next();
|
||||||
}, e);
|
}, e);
|
||||||
|
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
const dirName = path.dirname(fileName);
|
const dirName = path.dirname(fileName);
|
||||||
const targetDirName = path.join(targetPath, dirName);
|
const targetDirName = path.join(targetPath, dirName);
|
||||||
if (targetDirName.indexOf(targetPath) !== 0) {
|
if (targetDirName.indexOf(targetPath) !== 0) {
|
||||||
e(nls.localize('invalid file', "Error extracting {0}. Invalid file.", fileName));
|
e(nls.localize("invalid file", "Error extracting {0}. Invalid file.", fileName));
|
||||||
|
|
||||||
return nextEntry();
|
return nextEntry();
|
||||||
}
|
}
|
||||||
|
|
||||||
mkdirp(targetDirName, void 0, token).then(() => {
|
return mkdirp(targetDirName, undefined, token).then(() => {
|
||||||
const fstream = fs.createWriteStream(targetFileName, { mode: header.mode });
|
const fstream = fs.createWriteStream(targetFileName, { mode: header.mode });
|
||||||
fstream.once('close', () => {
|
fstream.once("close", () => {
|
||||||
next();
|
next();
|
||||||
});
|
});
|
||||||
fstream.once('error', (err) => {
|
fstream.once("error", e);
|
||||||
e(err);
|
|
||||||
});
|
|
||||||
stream.pipe(fstream);
|
stream.pipe(fstream);
|
||||||
stream.resume();
|
stream.resume();
|
||||||
});
|
});
|
||||||
});
|
});
|
||||||
extractor.once('finish', () => {
|
extractor.once("finish", c);
|
||||||
c();
|
|
||||||
});
|
|
||||||
extractor.write(buffer);
|
extractor.write(buffer);
|
||||||
extractor.end();
|
extractor.end();
|
||||||
});
|
} catch (ex) {
|
||||||
}
|
|
||||||
|
|
||||||
export function buffer(tarPath: string, filePath: string): Promise<Buffer> {
|
|
||||||
return new Promise<Buffer>(async (c, e) => {
|
|
||||||
let done: boolean = false;
|
|
||||||
extractAssets(tarPath, new RegExp(filePath), (path: string, data: Buffer) => {
|
|
||||||
if (path === filePath) {
|
|
||||||
done = true;
|
|
||||||
c(data);
|
|
||||||
}
|
|
||||||
}).then(() => {
|
|
||||||
if (!done) {
|
|
||||||
e("couldnt find asset " + filePath);
|
|
||||||
}
|
|
||||||
}).catch((ex) => {
|
|
||||||
e(ex);
|
e(ex);
|
||||||
});
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
async function extractAssets(tarPath: string, match: RegExp, callback: (path: string, data: Buffer) => void): Promise<void> {
|
|
||||||
const buffer = await promisify(fs.readFile)(tarPath);
|
|
||||||
const extractor = tarStream.extract();
|
|
||||||
let callbackResolve: () => void;
|
|
||||||
let callbackReject: (ex?) => void;
|
|
||||||
const complete = new Promise<void>((r, rej) => {
|
|
||||||
callbackResolve = r;
|
|
||||||
callbackReject = rej;
|
|
||||||
});
|
|
||||||
extractor.once("error", (err) => {
|
|
||||||
callbackReject(err);
|
|
||||||
});
|
|
||||||
extractor.on("entry", (header, stream, next) => {
|
|
||||||
const name = header.name;
|
|
||||||
if (match.test(name)) {
|
|
||||||
extractData(stream).then((data) => {
|
|
||||||
callback(name, data);
|
|
||||||
next();
|
|
||||||
});
|
|
||||||
stream.resume();
|
|
||||||
} else {
|
|
||||||
stream.on("end", () => {
|
|
||||||
next();
|
|
||||||
});
|
|
||||||
stream.resume();
|
|
||||||
}
|
}
|
||||||
});
|
});
|
||||||
extractor.on("finish", () => {
|
};
|
||||||
callbackResolve();
|
|
||||||
});
|
|
||||||
extractor.write(buffer);
|
|
||||||
extractor.end();
|
|
||||||
return complete;
|
|
||||||
}
|
|
||||||
|
|
||||||
async function extractData(stream: NodeJS.ReadableStream): Promise<Buffer> {
|
|
||||||
return new Promise<Buffer>((res, rej) => {
|
|
||||||
const fileData: Buffer[] = [];
|
|
||||||
stream.on('data', (data) => fileData.push(data));
|
|
||||||
stream.on('end', () => {
|
|
||||||
const fd = Buffer.concat(fileData);
|
|
||||||
res(fd);
|
|
||||||
});
|
|
||||||
stream.on('error', (err) => {
|
|
||||||
rej(err);
|
|
||||||
});
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
Binary file not shown.
Binary file not shown.
|
@ -0,0 +1,59 @@
|
||||||
|
import * as zip from "../src/fill/zip";
|
||||||
|
import * as path from "path";
|
||||||
|
import * as fs from "fs";
|
||||||
|
import * as cp from "child_process";
|
||||||
|
import { CancellationToken } from "vs/base/common/cancellation";
|
||||||
|
|
||||||
|
// tslint:disable-next-line:no-any
|
||||||
|
jest.mock("vs/nls", () => ({ "localize": (...args: any): string => `${JSON.stringify(args)}` }));
|
||||||
|
|
||||||
|
describe("zip", () => {
|
||||||
|
const tarPath = path.resolve(__dirname, "./test-extension.tar");
|
||||||
|
const vsixPath = path.resolve(__dirname, "./test-extension.vsix");
|
||||||
|
const extractPath = path.resolve(__dirname, "./.test-extension");
|
||||||
|
|
||||||
|
beforeEach(() => {
|
||||||
|
if (!fs.existsSync(extractPath) || path.dirname(extractPath) !== __dirname) {
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
cp.execSync(`rm -rf '${extractPath}'`);
|
||||||
|
});
|
||||||
|
|
||||||
|
const resolveExtract = async (archivePath: string): Promise<void> => {
|
||||||
|
expect(fs.existsSync(archivePath)).toEqual(true);
|
||||||
|
await expect(zip.extract(
|
||||||
|
archivePath,
|
||||||
|
extractPath,
|
||||||
|
{ sourcePath: "extension", overwrite: true },
|
||||||
|
CancellationToken.None,
|
||||||
|
)).resolves.toBe(undefined);
|
||||||
|
expect(fs.existsSync(extractPath)).toEqual(true);
|
||||||
|
};
|
||||||
|
|
||||||
|
// tslint:disable-next-line:no-any
|
||||||
|
const extract = (archivePath: string): () => any => {
|
||||||
|
// tslint:disable-next-line:no-any
|
||||||
|
return async (): Promise<any> => {
|
||||||
|
await resolveExtract(archivePath);
|
||||||
|
expect(fs.existsSync(path.resolve(extractPath, ".vsixmanifest"))).toEqual(true);
|
||||||
|
expect(fs.existsSync(path.resolve(extractPath, "package.json"))).toEqual(true);
|
||||||
|
};
|
||||||
|
};
|
||||||
|
it("should extract from tarred VSIX", extract(tarPath), 2000);
|
||||||
|
it("should extract from zipped VSIX", extract(vsixPath), 2000);
|
||||||
|
|
||||||
|
// tslint:disable-next-line:no-any
|
||||||
|
const buffer = (archivePath: string): () => any => {
|
||||||
|
// tslint:disable-next-line:no-any
|
||||||
|
return async (): Promise<any> => {
|
||||||
|
await resolveExtract(archivePath);
|
||||||
|
const manifestPath = path.resolve(extractPath, ".vsixmanifest");
|
||||||
|
expect(fs.existsSync(manifestPath)).toEqual(true);
|
||||||
|
const manifestBuf = fs.readFileSync(manifestPath);
|
||||||
|
expect(manifestBuf.length).toBeGreaterThan(0);
|
||||||
|
await expect(zip.buffer(archivePath, "extension.vsixmanifest")).resolves.toEqual(manifestBuf);
|
||||||
|
};
|
||||||
|
};
|
||||||
|
it("should buffer tarred VSIX", buffer(tarPath), 2000);
|
||||||
|
it("should buffer zipped VSIX", buffer(vsixPath), 2000);
|
||||||
|
});
|
|
@ -62,6 +62,7 @@ module.exports = merge(
|
||||||
"vs/platform/product/node/package": path.resolve(vsFills, "package.ts"),
|
"vs/platform/product/node/package": path.resolve(vsFills, "package.ts"),
|
||||||
"vs/platform/product/node/product": path.resolve(vsFills, "product.ts"),
|
"vs/platform/product/node/product": path.resolve(vsFills, "product.ts"),
|
||||||
"vs/base/node/zip": path.resolve(vsFills, "zip.ts"),
|
"vs/base/node/zip": path.resolve(vsFills, "zip.ts"),
|
||||||
|
"vszip": path.resolve(root, "lib/vscode/src/vs/base/node/zip.ts"),
|
||||||
"vs": path.resolve(root, "lib/vscode/src/vs"),
|
"vs": path.resolve(root, "lib/vscode/src/vs"),
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
|
|
|
@ -75,6 +75,7 @@ module.exports = merge(
|
||||||
"vs/platform/product/node/package": path.resolve(vsFills, "package.ts"),
|
"vs/platform/product/node/package": path.resolve(vsFills, "package.ts"),
|
||||||
"vs/platform/product/node/product": path.resolve(vsFills, "product.ts"),
|
"vs/platform/product/node/product": path.resolve(vsFills, "product.ts"),
|
||||||
"vs/base/node/zip": path.resolve(vsFills, "zip.ts"),
|
"vs/base/node/zip": path.resolve(vsFills, "zip.ts"),
|
||||||
|
"vszip": path.resolve(root, "lib/vscode/src/vs/base/node/zip.ts"),
|
||||||
"vs": path.join(root, "lib", "vscode", "src", "vs"),
|
"vs": path.join(root, "lib", "vscode", "src", "vs"),
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
|
|
Loading…
Reference in New Issue