Chunked Cache Upload APIs (#128)
* Initial pass at chunked upload apis * Fix cacheEntry type * Linting * Fix download cache entry tests * Linting tests * Pull in fixes from testing branch * Fix typo in ReserveCacheResponse * Add test convering reserve cache failure * Add retries to upload chunk * PR feedback * Format default chunk size * Remove responses array
This commit is contained in:
		
							parent
							
								
									a631fadf14
								
							
						
					
					
						commit
						b45d91cc4b
					
				| @ -248,7 +248,10 @@ test("restore with cache found", async () => { | |||||||
|     expect(getCacheMock).toHaveBeenCalledWith([key]); |     expect(getCacheMock).toHaveBeenCalledWith([key]); | ||||||
|     expect(setCacheStateMock).toHaveBeenCalledWith(cacheEntry); |     expect(setCacheStateMock).toHaveBeenCalledWith(cacheEntry); | ||||||
|     expect(createTempDirectoryMock).toHaveBeenCalledTimes(1); |     expect(createTempDirectoryMock).toHaveBeenCalledTimes(1); | ||||||
|     expect(downloadCacheMock).toHaveBeenCalledWith(cacheEntry, archivePath); |     expect(downloadCacheMock).toHaveBeenCalledWith( | ||||||
|  |         cacheEntry.archiveLocation, | ||||||
|  |         archivePath | ||||||
|  |     ); | ||||||
|     expect(getArchiveFileSizeMock).toHaveBeenCalledWith(archivePath); |     expect(getArchiveFileSizeMock).toHaveBeenCalledWith(archivePath); | ||||||
| 
 | 
 | ||||||
|     expect(extractTarMock).toHaveBeenCalledTimes(1); |     expect(extractTarMock).toHaveBeenCalledTimes(1); | ||||||
| @ -312,7 +315,10 @@ test("restore with a pull request event and cache found", async () => { | |||||||
|     expect(getCacheMock).toHaveBeenCalledWith([key]); |     expect(getCacheMock).toHaveBeenCalledWith([key]); | ||||||
|     expect(setCacheStateMock).toHaveBeenCalledWith(cacheEntry); |     expect(setCacheStateMock).toHaveBeenCalledWith(cacheEntry); | ||||||
|     expect(createTempDirectoryMock).toHaveBeenCalledTimes(1); |     expect(createTempDirectoryMock).toHaveBeenCalledTimes(1); | ||||||
|     expect(downloadCacheMock).toHaveBeenCalledWith(cacheEntry, archivePath); |     expect(downloadCacheMock).toHaveBeenCalledWith( | ||||||
|  |         cacheEntry.archiveLocation, | ||||||
|  |         archivePath | ||||||
|  |     ); | ||||||
|     expect(getArchiveFileSizeMock).toHaveBeenCalledWith(archivePath); |     expect(getArchiveFileSizeMock).toHaveBeenCalledWith(archivePath); | ||||||
|     expect(infoMock).toHaveBeenCalledWith(`Cache Size: ~60 MB (62915000 B)`); |     expect(infoMock).toHaveBeenCalledWith(`Cache Size: ~60 MB (62915000 B)`); | ||||||
| 
 | 
 | ||||||
| @ -377,7 +383,10 @@ test("restore with cache found for restore key", async () => { | |||||||
|     expect(getCacheMock).toHaveBeenCalledWith([key, restoreKey]); |     expect(getCacheMock).toHaveBeenCalledWith([key, restoreKey]); | ||||||
|     expect(setCacheStateMock).toHaveBeenCalledWith(cacheEntry); |     expect(setCacheStateMock).toHaveBeenCalledWith(cacheEntry); | ||||||
|     expect(createTempDirectoryMock).toHaveBeenCalledTimes(1); |     expect(createTempDirectoryMock).toHaveBeenCalledTimes(1); | ||||||
|     expect(downloadCacheMock).toHaveBeenCalledWith(cacheEntry, archivePath); |     expect(downloadCacheMock).toHaveBeenCalledWith( | ||||||
|  |         cacheEntry.archiveLocation, | ||||||
|  |         archivePath | ||||||
|  |     ); | ||||||
|     expect(getArchiveFileSizeMock).toHaveBeenCalledWith(archivePath); |     expect(getArchiveFileSizeMock).toHaveBeenCalledWith(archivePath); | ||||||
|     expect(infoMock).toHaveBeenCalledWith(`Cache Size: ~0 MB (142 B)`); |     expect(infoMock).toHaveBeenCalledWith(`Cache Size: ~0 MB (142 B)`); | ||||||
| 
 | 
 | ||||||
|  | |||||||
| @ -194,7 +194,7 @@ test("save with large cache outputs warning", async () => { | |||||||
| 
 | 
 | ||||||
|     const createTarMock = jest.spyOn(tar, "createTar"); |     const createTarMock = jest.spyOn(tar, "createTar"); | ||||||
| 
 | 
 | ||||||
|     const cacheSize = 1024 * 1024 * 1024; //~1GB, over the 400MB limit
 |     const cacheSize = 4 * 1024 * 1024 * 1024; //~4GB, over the 2GB limit
 | ||||||
|     jest.spyOn(actionUtils, "getArchiveFileSize").mockImplementationOnce(() => { |     jest.spyOn(actionUtils, "getArchiveFileSize").mockImplementationOnce(() => { | ||||||
|         return cacheSize; |         return cacheSize; | ||||||
|     }); |     }); | ||||||
| @ -208,12 +208,63 @@ test("save with large cache outputs warning", async () => { | |||||||
| 
 | 
 | ||||||
|     expect(logWarningMock).toHaveBeenCalledTimes(1); |     expect(logWarningMock).toHaveBeenCalledTimes(1); | ||||||
|     expect(logWarningMock).toHaveBeenCalledWith( |     expect(logWarningMock).toHaveBeenCalledWith( | ||||||
|         "Cache size of ~1024 MB (1073741824 B) is over the 400MB limit, not saving cache." |         "Cache size of ~4096 MB (4294967296 B) is over the 2GB limit, not saving cache." | ||||||
|     ); |     ); | ||||||
| 
 | 
 | ||||||
|     expect(failedMock).toHaveBeenCalledTimes(0); |     expect(failedMock).toHaveBeenCalledTimes(0); | ||||||
| }); | }); | ||||||
| 
 | 
 | ||||||
|  | test("save with reserve cache failure outputs warning", async () => { | ||||||
|  |     const infoMock = jest.spyOn(core, "info"); | ||||||
|  |     const logWarningMock = jest.spyOn(actionUtils, "logWarning"); | ||||||
|  |     const failedMock = jest.spyOn(core, "setFailed"); | ||||||
|  | 
 | ||||||
|  |     const primaryKey = "Linux-node-bb828da54c148048dd17899ba9fda624811cfb43"; | ||||||
|  |     const cacheEntry: ArtifactCacheEntry = { | ||||||
|  |         cacheKey: "Linux-node-", | ||||||
|  |         scope: "refs/heads/master", | ||||||
|  |         creationTime: "2019-11-13T19:18:02+00:00", | ||||||
|  |         archiveLocation: "www.actionscache.test/download" | ||||||
|  |     }; | ||||||
|  | 
 | ||||||
|  |     jest.spyOn(core, "getState") | ||||||
|  |         // Cache Entry State
 | ||||||
|  |         .mockImplementationOnce(() => { | ||||||
|  |             return JSON.stringify(cacheEntry); | ||||||
|  |         }) | ||||||
|  |         // Cache Key State
 | ||||||
|  |         .mockImplementationOnce(() => { | ||||||
|  |             return primaryKey; | ||||||
|  |         }); | ||||||
|  | 
 | ||||||
|  |     const inputPath = "node_modules"; | ||||||
|  |     testUtils.setInput(Inputs.Path, inputPath); | ||||||
|  | 
 | ||||||
|  |     const reserveCacheMock = jest | ||||||
|  |         .spyOn(cacheHttpClient, "reserveCache") | ||||||
|  |         .mockImplementationOnce(() => { | ||||||
|  |             return Promise.resolve(-1); | ||||||
|  |         }); | ||||||
|  | 
 | ||||||
|  |     const createTarMock = jest.spyOn(tar, "createTar"); | ||||||
|  | 
 | ||||||
|  |     const saveCacheMock = jest.spyOn(cacheHttpClient, "saveCache"); | ||||||
|  | 
 | ||||||
|  |     await run(); | ||||||
|  | 
 | ||||||
|  |     expect(reserveCacheMock).toHaveBeenCalledTimes(1); | ||||||
|  |     expect(reserveCacheMock).toHaveBeenCalledWith(primaryKey); | ||||||
|  | 
 | ||||||
|  |     expect(infoMock).toHaveBeenCalledWith( | ||||||
|  |         `Unable to reserve cache with key ${primaryKey}, another job may be creating this cache.` | ||||||
|  |     ); | ||||||
|  | 
 | ||||||
|  |     expect(createTarMock).toHaveBeenCalledTimes(0); | ||||||
|  |     expect(saveCacheMock).toHaveBeenCalledTimes(0); | ||||||
|  |     expect(logWarningMock).toHaveBeenCalledTimes(0); | ||||||
|  |     expect(failedMock).toHaveBeenCalledTimes(0); | ||||||
|  | }); | ||||||
|  | 
 | ||||||
| test("save with server error outputs warning", async () => { | test("save with server error outputs warning", async () => { | ||||||
|     const logWarningMock = jest.spyOn(actionUtils, "logWarning"); |     const logWarningMock = jest.spyOn(actionUtils, "logWarning"); | ||||||
|     const failedMock = jest.spyOn(core, "setFailed"); |     const failedMock = jest.spyOn(core, "setFailed"); | ||||||
| @ -240,6 +291,13 @@ test("save with server error outputs warning", async () => { | |||||||
|     const cachePath = path.resolve(inputPath); |     const cachePath = path.resolve(inputPath); | ||||||
|     testUtils.setInput(Inputs.Path, inputPath); |     testUtils.setInput(Inputs.Path, inputPath); | ||||||
| 
 | 
 | ||||||
|  |     const cacheId = 4; | ||||||
|  |     const reserveCacheMock = jest | ||||||
|  |         .spyOn(cacheHttpClient, "reserveCache") | ||||||
|  |         .mockImplementationOnce(() => { | ||||||
|  |             return Promise.resolve(cacheId); | ||||||
|  |         }); | ||||||
|  | 
 | ||||||
|     const createTarMock = jest.spyOn(tar, "createTar"); |     const createTarMock = jest.spyOn(tar, "createTar"); | ||||||
| 
 | 
 | ||||||
|     const saveCacheMock = jest |     const saveCacheMock = jest | ||||||
| @ -250,13 +308,16 @@ test("save with server error outputs warning", async () => { | |||||||
| 
 | 
 | ||||||
|     await run(); |     await run(); | ||||||
| 
 | 
 | ||||||
|  |     expect(reserveCacheMock).toHaveBeenCalledTimes(1); | ||||||
|  |     expect(reserveCacheMock).toHaveBeenCalledWith(primaryKey); | ||||||
|  | 
 | ||||||
|     const archivePath = path.join("/foo/bar", "cache.tgz"); |     const archivePath = path.join("/foo/bar", "cache.tgz"); | ||||||
| 
 | 
 | ||||||
|     expect(createTarMock).toHaveBeenCalledTimes(1); |     expect(createTarMock).toHaveBeenCalledTimes(1); | ||||||
|     expect(createTarMock).toHaveBeenCalledWith(archivePath, cachePath); |     expect(createTarMock).toHaveBeenCalledWith(archivePath, cachePath); | ||||||
| 
 | 
 | ||||||
|     expect(saveCacheMock).toHaveBeenCalledTimes(1); |     expect(saveCacheMock).toHaveBeenCalledTimes(1); | ||||||
|     expect(saveCacheMock).toHaveBeenCalledWith(primaryKey, archivePath); |     expect(saveCacheMock).toHaveBeenCalledWith(cacheId, archivePath); | ||||||
| 
 | 
 | ||||||
|     expect(logWarningMock).toHaveBeenCalledTimes(1); |     expect(logWarningMock).toHaveBeenCalledTimes(1); | ||||||
|     expect(logWarningMock).toHaveBeenCalledWith("HTTP Error Occurred"); |     expect(logWarningMock).toHaveBeenCalledWith("HTTP Error Occurred"); | ||||||
| @ -289,18 +350,29 @@ test("save with valid inputs uploads a cache", async () => { | |||||||
|     const cachePath = path.resolve(inputPath); |     const cachePath = path.resolve(inputPath); | ||||||
|     testUtils.setInput(Inputs.Path, inputPath); |     testUtils.setInput(Inputs.Path, inputPath); | ||||||
| 
 | 
 | ||||||
|  |     const cacheId = 4; | ||||||
|  |     const reserveCacheMock = jest | ||||||
|  |         .spyOn(cacheHttpClient, "reserveCache") | ||||||
|  |         .mockImplementationOnce(() => { | ||||||
|  |             return Promise.resolve(cacheId); | ||||||
|  |         }); | ||||||
|  | 
 | ||||||
|     const createTarMock = jest.spyOn(tar, "createTar"); |     const createTarMock = jest.spyOn(tar, "createTar"); | ||||||
|  | 
 | ||||||
|     const saveCacheMock = jest.spyOn(cacheHttpClient, "saveCache"); |     const saveCacheMock = jest.spyOn(cacheHttpClient, "saveCache"); | ||||||
| 
 | 
 | ||||||
|     await run(); |     await run(); | ||||||
| 
 | 
 | ||||||
|  |     expect(reserveCacheMock).toHaveBeenCalledTimes(1); | ||||||
|  |     expect(reserveCacheMock).toHaveBeenCalledWith(primaryKey); | ||||||
|  | 
 | ||||||
|     const archivePath = path.join("/foo/bar", "cache.tgz"); |     const archivePath = path.join("/foo/bar", "cache.tgz"); | ||||||
| 
 | 
 | ||||||
|     expect(createTarMock).toHaveBeenCalledTimes(1); |     expect(createTarMock).toHaveBeenCalledTimes(1); | ||||||
|     expect(createTarMock).toHaveBeenCalledWith(archivePath, cachePath); |     expect(createTarMock).toHaveBeenCalledWith(archivePath, cachePath); | ||||||
| 
 | 
 | ||||||
|     expect(saveCacheMock).toHaveBeenCalledTimes(1); |     expect(saveCacheMock).toHaveBeenCalledTimes(1); | ||||||
|     expect(saveCacheMock).toHaveBeenCalledWith(primaryKey, archivePath); |     expect(saveCacheMock).toHaveBeenCalledWith(cacheId, archivePath); | ||||||
| 
 | 
 | ||||||
|     expect(failedMock).toHaveBeenCalledTimes(0); |     expect(failedMock).toHaveBeenCalledTimes(0); | ||||||
| }); | }); | ||||||
|  | |||||||
							
								
								
									
										2
									
								
								package-lock.json
									
									
									
										generated
									
									
									
								
							
							
						
						
									
										2
									
								
								package-lock.json
									
									
									
										generated
									
									
									
								
							| @ -1,6 +1,6 @@ | |||||||
| { | { | ||||||
|   "name": "cache", |   "name": "cache", | ||||||
|   "version": "1.0.3", |   "version": "1.1.0", | ||||||
|   "lockfileVersion": 1, |   "lockfileVersion": 1, | ||||||
|   "requires": true, |   "requires": true, | ||||||
|   "dependencies": { |   "dependencies": { | ||||||
|  | |||||||
| @ -1,6 +1,6 @@ | |||||||
| { | { | ||||||
|   "name": "cache", |   "name": "cache", | ||||||
|   "version": "1.0.3", |   "version": "1.1.0", | ||||||
|   "private": true, |   "private": true, | ||||||
|   "description": "Cache dependencies and build outputs", |   "description": "Cache dependencies and build outputs", | ||||||
|   "main": "dist/restore/index.js", |   "main": "dist/restore/index.js", | ||||||
|  | |||||||
| @ -1,26 +1,49 @@ | |||||||
| import * as core from "@actions/core"; | import * as core from "@actions/core"; | ||||||
| import * as fs from "fs"; | import * as fs from "fs"; | ||||||
| import { BearerCredentialHandler } from "typed-rest-client/Handlers"; | import { BearerCredentialHandler } from "typed-rest-client/Handlers"; | ||||||
| import { HttpClient } from "typed-rest-client/HttpClient"; | import { HttpClient, HttpCodes } from "typed-rest-client/HttpClient"; | ||||||
| import { IHttpClientResponse } from "typed-rest-client/Interfaces"; | import { IHttpClientResponse } from "typed-rest-client/Interfaces"; | ||||||
| import { IRequestOptions, RestClient } from "typed-rest-client/RestClient"; | import { | ||||||
| import { ArtifactCacheEntry } from "./contracts"; |     IRequestOptions, | ||||||
|  |     RestClient, | ||||||
|  |     IRestResponse | ||||||
|  | } from "typed-rest-client/RestClient"; | ||||||
|  | import { | ||||||
|  |     ArtifactCacheEntry, | ||||||
|  |     CommitCacheRequest, | ||||||
|  |     ReserveCacheRequest, | ||||||
|  |     ReserveCacheResponse | ||||||
|  | } from "./contracts"; | ||||||
|  | import * as utils from "./utils/actionUtils"; | ||||||
| 
 | 
 | ||||||
| function getCacheUrl(): string { | function isSuccessStatusCode(statusCode: number): boolean { | ||||||
|  |     return statusCode >= 200 && statusCode < 300; | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | function isRetryableStatusCode(statusCode: number): boolean { | ||||||
|  |     const retryableStatusCodes = [ | ||||||
|  |         HttpCodes.BadGateway, | ||||||
|  |         HttpCodes.ServiceUnavailable, | ||||||
|  |         HttpCodes.GatewayTimeout | ||||||
|  |     ]; | ||||||
|  |     return retryableStatusCodes.includes(statusCode); | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | function getCacheApiUrl(): string { | ||||||
|     // Ideally we just use ACTIONS_CACHE_URL
 |     // Ideally we just use ACTIONS_CACHE_URL
 | ||||||
|     const cacheUrl: string = ( |     const baseUrl: string = ( | ||||||
|         process.env["ACTIONS_CACHE_URL"] || |         process.env["ACTIONS_CACHE_URL"] || | ||||||
|         process.env["ACTIONS_RUNTIME_URL"] || |         process.env["ACTIONS_RUNTIME_URL"] || | ||||||
|         "" |         "" | ||||||
|     ).replace("pipelines", "artifactcache"); |     ).replace("pipelines", "artifactcache"); | ||||||
|     if (!cacheUrl) { |     if (!baseUrl) { | ||||||
|         throw new Error( |         throw new Error( | ||||||
|             "Cache Service Url not found, unable to restore cache." |             "Cache Service Url not found, unable to restore cache." | ||||||
|         ); |         ); | ||||||
|     } |     } | ||||||
| 
 | 
 | ||||||
|     core.debug(`Cache Url: ${cacheUrl}`); |     core.debug(`Cache Url: ${baseUrl}`); | ||||||
|     return cacheUrl; |     return `${baseUrl}_apis/artifactcache/`; | ||||||
| } | } | ||||||
| 
 | 
 | ||||||
| function createAcceptHeader(type: string, apiVersion: string): string { | function createAcceptHeader(type: string, apiVersion: string): string { | ||||||
| @ -29,26 +52,26 @@ function createAcceptHeader(type: string, apiVersion: string): string { | |||||||
| 
 | 
 | ||||||
| function getRequestOptions(): IRequestOptions { | function getRequestOptions(): IRequestOptions { | ||||||
|     const requestOptions: IRequestOptions = { |     const requestOptions: IRequestOptions = { | ||||||
|         acceptHeader: createAcceptHeader("application/json", "5.2-preview.1") |         acceptHeader: createAcceptHeader("application/json", "6.0-preview.1") | ||||||
|     }; |     }; | ||||||
| 
 | 
 | ||||||
|     return requestOptions; |     return requestOptions; | ||||||
| } | } | ||||||
| 
 | 
 | ||||||
| export async function getCacheEntry( | function createRestClient(): RestClient { | ||||||
|     keys: string[] |  | ||||||
| ): Promise<ArtifactCacheEntry | null> { |  | ||||||
|     const cacheUrl = getCacheUrl(); |  | ||||||
|     const token = process.env["ACTIONS_RUNTIME_TOKEN"] || ""; |     const token = process.env["ACTIONS_RUNTIME_TOKEN"] || ""; | ||||||
|     const bearerCredentialHandler = new BearerCredentialHandler(token); |     const bearerCredentialHandler = new BearerCredentialHandler(token); | ||||||
| 
 | 
 | ||||||
|     const resource = `_apis/artifactcache/cache?keys=${encodeURIComponent( |     return new RestClient("actions/cache", getCacheApiUrl(), [ | ||||||
|         keys.join(",") |  | ||||||
|     )}`;
 |  | ||||||
| 
 |  | ||||||
|     const restClient = new RestClient("actions/cache", cacheUrl, [ |  | ||||||
|         bearerCredentialHandler |         bearerCredentialHandler | ||||||
|     ]); |     ]); | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | export async function getCacheEntry( | ||||||
|  |     keys: string[] | ||||||
|  | ): Promise<ArtifactCacheEntry | null> { | ||||||
|  |     const restClient = createRestClient(); | ||||||
|  |     const resource = `cache?keys=${encodeURIComponent(keys.join(","))}`; | ||||||
| 
 | 
 | ||||||
|     const response = await restClient.get<ArtifactCacheEntry>( |     const response = await restClient.get<ArtifactCacheEntry>( | ||||||
|         resource, |         resource, | ||||||
| @ -57,14 +80,15 @@ export async function getCacheEntry( | |||||||
|     if (response.statusCode === 204) { |     if (response.statusCode === 204) { | ||||||
|         return null; |         return null; | ||||||
|     } |     } | ||||||
|     if (response.statusCode !== 200) { |     if (!isSuccessStatusCode(response.statusCode)) { | ||||||
|         throw new Error(`Cache service responded with ${response.statusCode}`); |         throw new Error(`Cache service responded with ${response.statusCode}`); | ||||||
|     } |     } | ||||||
|     const cacheResult = response.result; |     const cacheResult = response.result; | ||||||
|     if (!cacheResult || !cacheResult.archiveLocation) { |     const cacheDownloadUrl = cacheResult?.archiveLocation; | ||||||
|  |     if (!cacheDownloadUrl) { | ||||||
|         throw new Error("Cache not found."); |         throw new Error("Cache not found."); | ||||||
|     } |     } | ||||||
|     core.setSecret(cacheResult.archiveLocation); |     core.setSecret(cacheDownloadUrl); | ||||||
|     core.debug(`Cache Result:`); |     core.debug(`Cache Result:`); | ||||||
|     core.debug(JSON.stringify(cacheResult)); |     core.debug(JSON.stringify(cacheResult)); | ||||||
| 
 | 
 | ||||||
| @ -83,46 +107,178 @@ async function pipeResponseToStream( | |||||||
| } | } | ||||||
| 
 | 
 | ||||||
| export async function downloadCache( | export async function downloadCache( | ||||||
|     cacheEntry: ArtifactCacheEntry, |     archiveLocation: string, | ||||||
|     archivePath: string |     archivePath: string | ||||||
| ): Promise<void> { | ): Promise<void> { | ||||||
|     const stream = fs.createWriteStream(archivePath); |     const stream = fs.createWriteStream(archivePath); | ||||||
|     const httpClient = new HttpClient("actions/cache"); |     const httpClient = new HttpClient("actions/cache"); | ||||||
|     // eslint-disable-next-line @typescript-eslint/no-non-null-assertion
 |     const downloadResponse = await httpClient.get(archiveLocation); | ||||||
|     const downloadResponse = await httpClient.get(cacheEntry.archiveLocation!); |  | ||||||
|     await pipeResponseToStream(downloadResponse, stream); |     await pipeResponseToStream(downloadResponse, stream); | ||||||
| } | } | ||||||
| 
 | 
 | ||||||
| export async function saveCache( | // Reserve Cache
 | ||||||
|     key: string, | export async function reserveCache(key: string): Promise<number> { | ||||||
|     archivePath: string |     const restClient = createRestClient(); | ||||||
|  | 
 | ||||||
|  |     const reserveCacheRequest: ReserveCacheRequest = { | ||||||
|  |         key | ||||||
|  |     }; | ||||||
|  |     const response = await restClient.create<ReserveCacheResponse>( | ||||||
|  |         "caches", | ||||||
|  |         reserveCacheRequest, | ||||||
|  |         getRequestOptions() | ||||||
|  |     ); | ||||||
|  | 
 | ||||||
|  |     return response?.result?.cacheId ?? -1; | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | function getContentRange(start: number, end: number): string { | ||||||
|  |     // Format: `bytes start-end/filesize
 | ||||||
|  |     // start and end are inclusive
 | ||||||
|  |     // filesize can be *
 | ||||||
|  |     // For a 200 byte chunk starting at byte 0:
 | ||||||
|  |     // Content-Range: bytes 0-199/*
 | ||||||
|  |     return `bytes ${start}-${end}/*`; | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | async function uploadChunk( | ||||||
|  |     restClient: RestClient, | ||||||
|  |     resourceUrl: string, | ||||||
|  |     data: NodeJS.ReadableStream, | ||||||
|  |     start: number, | ||||||
|  |     end: number | ||||||
| ): Promise<void> { | ): Promise<void> { | ||||||
|     const stream = fs.createReadStream(archivePath); |     core.debug( | ||||||
| 
 |         `Uploading chunk of size ${end - | ||||||
|     const cacheUrl = getCacheUrl(); |             start + | ||||||
|     const token = process.env["ACTIONS_RUNTIME_TOKEN"] || ""; |             1} bytes at offset ${start} with content range: ${getContentRange( | ||||||
|     const bearerCredentialHandler = new BearerCredentialHandler(token); |             start, | ||||||
| 
 |             end | ||||||
|     const resource = `_apis/artifactcache/cache/${encodeURIComponent(key)}`; |         )}` | ||||||
|     const postUrl = cacheUrl + resource; |     ); | ||||||
| 
 |  | ||||||
|     const restClient = new RestClient("actions/cache", undefined, [ |  | ||||||
|         bearerCredentialHandler |  | ||||||
|     ]); |  | ||||||
| 
 |  | ||||||
|     const requestOptions = getRequestOptions(); |     const requestOptions = getRequestOptions(); | ||||||
|     requestOptions.additionalHeaders = { |     requestOptions.additionalHeaders = { | ||||||
|         "Content-Type": "application/octet-stream" |         "Content-Type": "application/octet-stream", | ||||||
|  |         "Content-Range": getContentRange(start, end) | ||||||
|     }; |     }; | ||||||
| 
 | 
 | ||||||
|     const response = await restClient.uploadStream<void>( |     const uploadChunkRequest = async (): Promise<IRestResponse<void>> => { | ||||||
|         "POST", |         return await restClient.uploadStream<void>( | ||||||
|         postUrl, |             "PATCH", | ||||||
|         stream, |             resourceUrl, | ||||||
|  |             data, | ||||||
|  |             requestOptions | ||||||
|  |         ); | ||||||
|  |     }; | ||||||
|  | 
 | ||||||
|  |     const response = await uploadChunkRequest(); | ||||||
|  |     if (isSuccessStatusCode(response.statusCode)) { | ||||||
|  |         return; | ||||||
|  |     } | ||||||
|  | 
 | ||||||
|  |     if (isRetryableStatusCode(response.statusCode)) { | ||||||
|  |         core.debug( | ||||||
|  |             `Received ${response.statusCode}, retrying chunk at offset ${start}.` | ||||||
|  |         ); | ||||||
|  |         const retryResponse = await uploadChunkRequest(); | ||||||
|  |         if (isSuccessStatusCode(retryResponse.statusCode)) { | ||||||
|  |             return; | ||||||
|  |         } | ||||||
|  |     } | ||||||
|  | 
 | ||||||
|  |     throw new Error( | ||||||
|  |         `Cache service responded with ${response.statusCode} during chunk upload.` | ||||||
|  |     ); | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | async function uploadFile( | ||||||
|  |     restClient: RestClient, | ||||||
|  |     cacheId: number, | ||||||
|  |     archivePath: string | ||||||
|  | ): Promise<void> { | ||||||
|  |     // Upload Chunks
 | ||||||
|  |     const fileSize = fs.statSync(archivePath).size; | ||||||
|  |     const resourceUrl = getCacheApiUrl() + "caches/" + cacheId.toString(); | ||||||
|  |     const fd = fs.openSync(archivePath, "r"); | ||||||
|  | 
 | ||||||
|  |     const concurrency = Number(process.env["CACHE_UPLOAD_CONCURRENCY"]) ?? 4; // # of HTTP requests in parallel
 | ||||||
|  |     const MAX_CHUNK_SIZE = | ||||||
|  |         Number(process.env["CACHE_UPLOAD_CHUNK_SIZE"]) ?? 32 * 1024 * 1024; // 32 MB Chunks
 | ||||||
|  |     core.debug(`Concurrency: ${concurrency} and Chunk Size: ${MAX_CHUNK_SIZE}`); | ||||||
|  | 
 | ||||||
|  |     const parallelUploads = [...new Array(concurrency).keys()]; | ||||||
|  |     core.debug("Awaiting all uploads"); | ||||||
|  |     let offset = 0; | ||||||
|  | 
 | ||||||
|  |     try { | ||||||
|  |         await Promise.all( | ||||||
|  |             parallelUploads.map(async () => { | ||||||
|  |                 while (offset < fileSize) { | ||||||
|  |                     const chunkSize = Math.min( | ||||||
|  |                         fileSize - offset, | ||||||
|  |                         MAX_CHUNK_SIZE | ||||||
|  |                     ); | ||||||
|  |                     const start = offset; | ||||||
|  |                     const end = offset + chunkSize - 1; | ||||||
|  |                     offset += MAX_CHUNK_SIZE; | ||||||
|  |                     const chunk = fs.createReadStream(archivePath, { | ||||||
|  |                         fd, | ||||||
|  |                         start, | ||||||
|  |                         end, | ||||||
|  |                         autoClose: false | ||||||
|  |                     }); | ||||||
|  | 
 | ||||||
|  |                     await uploadChunk( | ||||||
|  |                         restClient, | ||||||
|  |                         resourceUrl, | ||||||
|  |                         chunk, | ||||||
|  |                         start, | ||||||
|  |                         end | ||||||
|  |                     ); | ||||||
|  |                 } | ||||||
|  |             }) | ||||||
|  |         ); | ||||||
|  |     } finally { | ||||||
|  |         fs.closeSync(fd); | ||||||
|  |     } | ||||||
|  |     return; | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | async function commitCache( | ||||||
|  |     restClient: RestClient, | ||||||
|  |     cacheId: number, | ||||||
|  |     filesize: number | ||||||
|  | ): Promise<IRestResponse<void>> { | ||||||
|  |     const requestOptions = getRequestOptions(); | ||||||
|  |     const commitCacheRequest: CommitCacheRequest = { size: filesize }; | ||||||
|  |     return await restClient.create( | ||||||
|  |         `caches/${cacheId.toString()}`, | ||||||
|  |         commitCacheRequest, | ||||||
|         requestOptions |         requestOptions | ||||||
|     ); |     ); | ||||||
|     if (response.statusCode !== 200) { | } | ||||||
|         throw new Error(`Cache service responded with ${response.statusCode}`); | 
 | ||||||
|  | export async function saveCache( | ||||||
|  |     cacheId: number, | ||||||
|  |     archivePath: string | ||||||
|  | ): Promise<void> { | ||||||
|  |     const restClient = createRestClient(); | ||||||
|  | 
 | ||||||
|  |     core.debug("Upload cache"); | ||||||
|  |     await uploadFile(restClient, cacheId, archivePath); | ||||||
|  | 
 | ||||||
|  |     // Commit Cache
 | ||||||
|  |     core.debug("Commiting cache"); | ||||||
|  |     const cacheSize = utils.getArchiveFileSize(archivePath); | ||||||
|  |     const commitCacheResponse = await commitCache( | ||||||
|  |         restClient, | ||||||
|  |         cacheId, | ||||||
|  |         cacheSize | ||||||
|  |     ); | ||||||
|  |     if (!isSuccessStatusCode(commitCacheResponse.statusCode)) { | ||||||
|  |         throw new Error( | ||||||
|  |             `Cache service responded with ${commitCacheResponse.statusCode} during commit cache.` | ||||||
|  |         ); | ||||||
|     } |     } | ||||||
| 
 | 
 | ||||||
|     core.info("Cache saved successfully"); |     core.info("Cache saved successfully"); | ||||||
|  | |||||||
							
								
								
									
										13
									
								
								src/contracts.d.ts
									
									
									
									
										vendored
									
									
								
							
							
						
						
									
										13
									
								
								src/contracts.d.ts
									
									
									
									
										vendored
									
									
								
							| @ -4,3 +4,16 @@ export interface ArtifactCacheEntry { | |||||||
|     creationTime?: string; |     creationTime?: string; | ||||||
|     archiveLocation?: string; |     archiveLocation?: string; | ||||||
| } | } | ||||||
|  | 
 | ||||||
|  | export interface CommitCacheRequest { | ||||||
|  |     size: number; | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | export interface ReserveCacheRequest { | ||||||
|  |     key: string; | ||||||
|  |     version?: string; | ||||||
|  | } | ||||||
|  | 
 | ||||||
|  | export interface ReserveCacheResponse { | ||||||
|  |     cacheId: number; | ||||||
|  | } | ||||||
|  | |||||||
| @ -60,7 +60,7 @@ async function run(): Promise<void> { | |||||||
| 
 | 
 | ||||||
|         try { |         try { | ||||||
|             const cacheEntry = await cacheHttpClient.getCacheEntry(keys); |             const cacheEntry = await cacheHttpClient.getCacheEntry(keys); | ||||||
|             if (!cacheEntry) { |             if (!cacheEntry?.archiveLocation) { | ||||||
|                 core.info( |                 core.info( | ||||||
|                     `Cache not found for input keys: ${keys.join(", ")}.` |                     `Cache not found for input keys: ${keys.join(", ")}.` | ||||||
|                 ); |                 ); | ||||||
| @ -77,7 +77,10 @@ async function run(): Promise<void> { | |||||||
|             utils.setCacheState(cacheEntry); |             utils.setCacheState(cacheEntry); | ||||||
| 
 | 
 | ||||||
|             // Download the cache from the cache entry
 |             // Download the cache from the cache entry
 | ||||||
|             await cacheHttpClient.downloadCache(cacheEntry, archivePath); |             await cacheHttpClient.downloadCache( | ||||||
|  |                 cacheEntry.archiveLocation, | ||||||
|  |                 archivePath | ||||||
|  |             ); | ||||||
| 
 | 
 | ||||||
|             const archiveFileSize = utils.getArchiveFileSize(archivePath); |             const archiveFileSize = utils.getArchiveFileSize(archivePath); | ||||||
|             core.info( |             core.info( | ||||||
|  | |||||||
							
								
								
									
										16
									
								
								src/save.ts
									
									
									
									
									
								
							
							
						
						
									
										16
									
								
								src/save.ts
									
									
									
									
									
								
							| @ -34,6 +34,15 @@ async function run(): Promise<void> { | |||||||
|             return; |             return; | ||||||
|         } |         } | ||||||
| 
 | 
 | ||||||
|  |         core.debug("Reserving Cache"); | ||||||
|  |         const cacheId = await cacheHttpClient.reserveCache(primaryKey); | ||||||
|  |         if (cacheId == -1) { | ||||||
|  |             core.info( | ||||||
|  |                 `Unable to reserve cache with key ${primaryKey}, another job may be creating this cache.` | ||||||
|  |             ); | ||||||
|  |             return; | ||||||
|  |         } | ||||||
|  |         core.debug(`Cache ID: ${cacheId}`); | ||||||
|         const cachePath = utils.resolvePath( |         const cachePath = utils.resolvePath( | ||||||
|             core.getInput(Inputs.Path, { required: true }) |             core.getInput(Inputs.Path, { required: true }) | ||||||
|         ); |         ); | ||||||
| @ -47,19 +56,20 @@ async function run(): Promise<void> { | |||||||
| 
 | 
 | ||||||
|         await createTar(archivePath, cachePath); |         await createTar(archivePath, cachePath); | ||||||
| 
 | 
 | ||||||
|         const fileSizeLimit = 400 * 1024 * 1024; // 400MB
 |         const fileSizeLimit = 2 * 1024 * 1024 * 1024; // 2GB per repo limit
 | ||||||
|         const archiveFileSize = utils.getArchiveFileSize(archivePath); |         const archiveFileSize = utils.getArchiveFileSize(archivePath); | ||||||
|         core.debug(`File Size: ${archiveFileSize}`); |         core.debug(`File Size: ${archiveFileSize}`); | ||||||
|         if (archiveFileSize > fileSizeLimit) { |         if (archiveFileSize > fileSizeLimit) { | ||||||
|             utils.logWarning( |             utils.logWarning( | ||||||
|                 `Cache size of ~${Math.round( |                 `Cache size of ~${Math.round( | ||||||
|                     archiveFileSize / (1024 * 1024) |                     archiveFileSize / (1024 * 1024) | ||||||
|                 )} MB (${archiveFileSize} B) is over the 400MB limit, not saving cache.` |                 )} MB (${archiveFileSize} B) is over the 2GB limit, not saving cache.` | ||||||
|             ); |             ); | ||||||
|             return; |             return; | ||||||
|         } |         } | ||||||
| 
 | 
 | ||||||
|         await cacheHttpClient.saveCache(primaryKey, archivePath); |         core.debug(`Saving Cache (ID: ${cacheId})`); | ||||||
|  |         await cacheHttpClient.saveCache(cacheId, archivePath); | ||||||
|     } catch (error) { |     } catch (error) { | ||||||
|         utils.logWarning(error.message); |         utils.logWarning(error.message); | ||||||
|     } |     } | ||||||
|  | |||||||
		Loading…
	
	
			
			x
			
			
		
	
		Reference in New Issue
	
	Block a user