quiet zipping

pull/353/head
Mark D 2022-04-03 21:39:20 +01:00
parent b6133c86b3
commit ff51a4a617
4 changed files with 13 additions and 13 deletions

12
dist/index.js vendored
View File

@ -647,7 +647,7 @@ class Caching {
return typeof arguments_[number] != 'undefined' ? arguments_[number] : match; return typeof arguments_[number] != 'undefined' ? arguments_[number] : match;
}); });
}; };
yield cloud_runner_system_1.CloudRunnerSystem.Run(`zip ${cacheKey}.zip ${path_1.default.basename(sourceFolder)}`); yield cloud_runner_system_1.CloudRunnerSystem.Run(`zip -q ${cacheKey}.zip ${path_1.default.basename(sourceFolder)}`);
console_1.assert(fs_1.default.existsSync(`${cacheKey}.zip`), 'cache zip exists'); console_1.assert(fs_1.default.existsSync(`${cacheKey}.zip`), 'cache zip exists');
console_1.assert(fs_1.default.existsSync(path_1.default.basename(sourceFolder)), 'source folder exists'); console_1.assert(fs_1.default.existsSync(path_1.default.basename(sourceFolder)), 'source folder exists');
if (cloud_runner_1.default.buildParameters.cachePushOverrideCommand) { if (cloud_runner_1.default.buildParameters.cachePushOverrideCommand) {
@ -702,7 +702,7 @@ class Caching {
console_1.assert(`${fs_1.default.existsSync(destinationFolder)}`); console_1.assert(`${fs_1.default.existsSync(destinationFolder)}`);
console_1.assert(`${fs_1.default.existsSync(`${cacheSelection}.zip`)}`); console_1.assert(`${fs_1.default.existsSync(`${cacheSelection}.zip`)}`);
const fullResultsFolder = path_1.default.join(cacheFolder, resultsFolder); const fullResultsFolder = path_1.default.join(cacheFolder, resultsFolder);
yield cloud_runner_system_1.CloudRunnerSystem.Run(`unzip ${cacheSelection}.zip -d ${path_1.default.basename(resultsFolder)}`); yield cloud_runner_system_1.CloudRunnerSystem.Run(`unzip -q ${cacheSelection}.zip -d ${path_1.default.basename(resultsFolder)}`);
remote_client_logger_1.RemoteClientLogger.log(`cache item extracted to ${fullResultsFolder}`); remote_client_logger_1.RemoteClientLogger.log(`cache item extracted to ${fullResultsFolder}`);
console_1.assert(`${fs_1.default.existsSync(fullResultsFolder)}`); console_1.assert(`${fs_1.default.existsSync(fullResultsFolder)}`);
const destinationParentFolder = path_1.default.resolve(destinationFolder, '..'); const destinationParentFolder = path_1.default.resolve(destinationFolder, '..');
@ -3207,15 +3207,15 @@ class BuildAutomationWorkflow {
const setupHooks = cloud_runner_build_command_process_1.CloudRunnerBuildCommandProcessor.getHooks(cloud_runner_1.default.buildParameters.customJobHooks).filter((x) => x.step.includes(`setup`)); const setupHooks = cloud_runner_build_command_process_1.CloudRunnerBuildCommandProcessor.getHooks(cloud_runner_1.default.buildParameters.customJobHooks).filter((x) => x.step.includes(`setup`));
const buildHooks = cloud_runner_build_command_process_1.CloudRunnerBuildCommandProcessor.getHooks(cloud_runner_1.default.buildParameters.customJobHooks).filter((x) => x.step.includes(`build`)); const buildHooks = cloud_runner_build_command_process_1.CloudRunnerBuildCommandProcessor.getHooks(cloud_runner_1.default.buildParameters.customJobHooks).filter((x) => x.step.includes(`build`));
return `apt-get update return `apt-get update
apt-get install -y -q zip tree npm git-lfs jq unzip git apt-get install -y -q zip tree npm git-lfs jq unzip git
npm install -g n npm install -g n
n stable n stable
${setupHooks.filter((x) => x.hook.includes(`before`)).map((x) => x.commands) || ' '} ${setupHooks.filter((x) => x.hook.includes(`before`)).map((x) => x.commands) || ' '}
export GITHUB_WORKSPACE="${cloud_runner_folders_1.CloudRunnerFolders.repoPathFull}" export GITHUB_WORKSPACE="${cloud_runner_folders_1.CloudRunnerFolders.repoPathFull.replace(/\\/g, `/`)}"
${BuildAutomationWorkflow.SetupCommands} ${BuildAutomationWorkflow.SetupCommands}
${setupHooks.filter((x) => x.hook.includes(`after`)).map((x) => x.commands) || ' '} ${setupHooks.filter((x) => x.hook.includes(`after`)).map((x) => x.commands) || ' '}
${buildHooks.filter((x) => x.hook.includes(`before`)).map((x) => x.commands) || ' '} ${buildHooks.filter((x) => x.hook.includes(`before`)).map((x) => x.commands) || ' '}
cd ${cloud_runner_folders_1.CloudRunnerFolders.projectPathFull.replace(/\\/g, `/`)} cd ${cloud_runner_folders_1.CloudRunnerFolders.repoPathFull.replace(/\\/g, `/`)}
${buildHooks.filter((x) => x.hook.includes(`after`)).map((x) => x.commands) || ' '} ${buildHooks.filter((x) => x.hook.includes(`after`)).map((x) => x.commands) || ' '}
${BuildAutomationWorkflow.BuildCommands}`; ${BuildAutomationWorkflow.BuildCommands}`;
} }
@ -3246,7 +3246,7 @@ class BuildAutomationWorkflow {
/entrypoint.sh /entrypoint.sh
echo "post build scripts" echo "post build scripts"
cd "${cloud_runner_folders_1.CloudRunnerFolders.libraryFolderFull.replace(/\\/g, `/`)}/.." cd "${cloud_runner_folders_1.CloudRunnerFolders.libraryFolderFull.replace(/\\/g, `/`)}/.."
zip -r "lib-${cloud_runner_1.default.buildParameters.buildGuid}.zip" "Library" zip -q -r "lib-${cloud_runner_1.default.buildParameters.buildGuid}.zip" "Library"
mv "lib-${cloud_runner_1.default.buildParameters.buildGuid}.zip" "${cloud_runner_folders_1.CloudRunnerFolders.cacheFolderFull.replace(/\\/g, `/`)}/Library" mv "lib-${cloud_runner_1.default.buildParameters.buildGuid}.zip" "${cloud_runner_folders_1.CloudRunnerFolders.cacheFolderFull.replace(/\\/g, `/`)}/Library"
cd "${cloud_runner_folders_1.CloudRunnerFolders.repoPathFull.replace(/\\/g, `/`)}" cd "${cloud_runner_folders_1.CloudRunnerFolders.repoPathFull.replace(/\\/g, `/`)}"
zip -r "build-${cloud_runner_1.default.buildParameters.buildGuid}.zip" "build" zip -r "build-${cloud_runner_1.default.buildParameters.buildGuid}.zip" "build"

2
dist/index.js.map vendored

File diff suppressed because one or more lines are too long

View File

@ -36,7 +36,7 @@ export class Caching {
return typeof arguments_[number] != 'undefined' ? arguments_[number] : match; return typeof arguments_[number] != 'undefined' ? arguments_[number] : match;
}); });
}; };
await CloudRunnerSystem.Run(`zip ${cacheKey}.zip ${path.basename(sourceFolder)}`); await CloudRunnerSystem.Run(`zip -q ${cacheKey}.zip ${path.basename(sourceFolder)}`);
assert(fs.existsSync(`${cacheKey}.zip`), 'cache zip exists'); assert(fs.existsSync(`${cacheKey}.zip`), 'cache zip exists');
assert(fs.existsSync(path.basename(sourceFolder)), 'source folder exists'); assert(fs.existsSync(path.basename(sourceFolder)), 'source folder exists');
if (CloudRunner.buildParameters.cachePushOverrideCommand) { if (CloudRunner.buildParameters.cachePushOverrideCommand) {
@ -99,7 +99,7 @@ export class Caching {
assert(`${fs.existsSync(destinationFolder)}`); assert(`${fs.existsSync(destinationFolder)}`);
assert(`${fs.existsSync(`${cacheSelection}.zip`)}`); assert(`${fs.existsSync(`${cacheSelection}.zip`)}`);
const fullResultsFolder = path.join(cacheFolder, resultsFolder); const fullResultsFolder = path.join(cacheFolder, resultsFolder);
await CloudRunnerSystem.Run(`unzip ${cacheSelection}.zip -d ${path.basename(resultsFolder)}`); await CloudRunnerSystem.Run(`unzip -q ${cacheSelection}.zip -d ${path.basename(resultsFolder)}`);
RemoteClientLogger.log(`cache item extracted to ${fullResultsFolder}`); RemoteClientLogger.log(`cache item extracted to ${fullResultsFolder}`);
assert(`${fs.existsSync(fullResultsFolder)}`); assert(`${fs.existsSync(fullResultsFolder)}`);
const destinationParentFolder = path.resolve(destinationFolder, '..'); const destinationParentFolder = path.resolve(destinationFolder, '..');

View File

@ -81,15 +81,15 @@ export class BuildAutomationWorkflow implements WorkflowInterface {
(x) => x.step.includes(`build`), (x) => x.step.includes(`build`),
); );
return `apt-get update return `apt-get update
apt-get install -y -q zip tree npm git-lfs jq unzip git apt-get install -y -q zip tree npm git-lfs jq unzip git
npm install -g n npm install -g n
n stable n stable
${setupHooks.filter((x) => x.hook.includes(`before`)).map((x) => x.commands) || ' '} ${setupHooks.filter((x) => x.hook.includes(`before`)).map((x) => x.commands) || ' '}
export GITHUB_WORKSPACE="${CloudRunnerFolders.repoPathFull}" export GITHUB_WORKSPACE="${CloudRunnerFolders.repoPathFull.replace(/\\/g, `/`)}"
${BuildAutomationWorkflow.SetupCommands} ${BuildAutomationWorkflow.SetupCommands}
${setupHooks.filter((x) => x.hook.includes(`after`)).map((x) => x.commands) || ' '} ${setupHooks.filter((x) => x.hook.includes(`after`)).map((x) => x.commands) || ' '}
${buildHooks.filter((x) => x.hook.includes(`before`)).map((x) => x.commands) || ' '} ${buildHooks.filter((x) => x.hook.includes(`before`)).map((x) => x.commands) || ' '}
cd ${CloudRunnerFolders.projectPathFull.replace(/\\/g, `/`)} cd ${CloudRunnerFolders.repoPathFull.replace(/\\/g, `/`)}
${buildHooks.filter((x) => x.hook.includes(`after`)).map((x) => x.commands) || ' '} ${buildHooks.filter((x) => x.hook.includes(`after`)).map((x) => x.commands) || ' '}
${BuildAutomationWorkflow.BuildCommands}`; ${BuildAutomationWorkflow.BuildCommands}`;
} }
@ -126,7 +126,7 @@ export class BuildAutomationWorkflow implements WorkflowInterface {
/entrypoint.sh /entrypoint.sh
echo "post build scripts" echo "post build scripts"
cd "${CloudRunnerFolders.libraryFolderFull.replace(/\\/g, `/`)}/.." cd "${CloudRunnerFolders.libraryFolderFull.replace(/\\/g, `/`)}/.."
zip -r "lib-${CloudRunner.buildParameters.buildGuid}.zip" "Library" zip -q -r "lib-${CloudRunner.buildParameters.buildGuid}.zip" "Library"
mv "lib-${CloudRunner.buildParameters.buildGuid}.zip" "${CloudRunnerFolders.cacheFolderFull.replace( mv "lib-${CloudRunner.buildParameters.buildGuid}.zip" "${CloudRunnerFolders.cacheFolderFull.replace(
/\\/g, /\\/g,
`/`, `/`,