Compare commits

...

2 Commits

Author SHA1 Message Date
Bassem Dghaidi 0a39994fe4
Merge 2501f58c53 into 6849a64899 2024-11-20 22:55:41 +01:00
Bassem Dghaidi 2501f58c53 Update cache service APIs & cleanup 2024-11-20 13:55:06 -08:00
5 changed files with 980 additions and 894 deletions

View File

@ -6064,11 +6064,8 @@ function restoreCacheV2(paths, primaryKey, restoreKeys, options, enableCrossOsAr
let archivePath = ''; let archivePath = '';
try { try {
const twirpClient = cacheTwirpClient.internalCacheTwirpClient(); const twirpClient = cacheTwirpClient.internalCacheTwirpClient();
const backendIds = utils.getBackendIdsFromToken();
const compressionMethod = yield utils.getCompressionMethod(); const compressionMethod = yield utils.getCompressionMethod();
const request = { const request = {
workflowRunBackendId: backendIds.workflowRunBackendId,
workflowJobRunBackendId: backendIds.workflowJobRunBackendId,
key: primaryKey, key: primaryKey,
restoreKeys, restoreKeys,
version: utils.getCacheVersion(paths, compressionMethod, enableCrossOsArchive) version: utils.getCacheVersion(paths, compressionMethod, enableCrossOsArchive)
@ -6223,8 +6220,6 @@ function saveCacheV1(paths, key, options, enableCrossOsArchive = false) {
*/ */
function saveCacheV2(paths, key, options, enableCrossOsArchive = false) { function saveCacheV2(paths, key, options, enableCrossOsArchive = false) {
return __awaiter(this, void 0, void 0, function* () { return __awaiter(this, void 0, void 0, function* () {
// BackendIds are retrieved form the signed JWT
const backendIds = utils.getBackendIdsFromToken();
const compressionMethod = yield utils.getCompressionMethod(); const compressionMethod = yield utils.getCompressionMethod();
const twirpClient = cacheTwirpClient.internalCacheTwirpClient(); const twirpClient = cacheTwirpClient.internalCacheTwirpClient();
let cacheId = -1; let cacheId = -1;
@ -6251,8 +6246,6 @@ function saveCacheV2(paths, key, options, enableCrossOsArchive = false) {
core.debug('Reserving Cache'); core.debug('Reserving Cache');
const version = utils.getCacheVersion(paths, compressionMethod, enableCrossOsArchive); const version = utils.getCacheVersion(paths, compressionMethod, enableCrossOsArchive);
const request = { const request = {
workflowRunBackendId: backendIds.workflowRunBackendId,
workflowJobRunBackendId: backendIds.workflowJobRunBackendId,
key, key,
version version
}; };
@ -6263,8 +6256,6 @@ function saveCacheV2(paths, key, options, enableCrossOsArchive = false) {
core.debug(`Attempting to upload cache located at: ${archivePath}`); core.debug(`Attempting to upload cache located at: ${archivePath}`);
yield (0, upload_cache_1.UploadCacheFile)(response.signedUploadUrl, archivePath); yield (0, upload_cache_1.UploadCacheFile)(response.signedUploadUrl, archivePath);
const finalizeRequest = { const finalizeRequest = {
workflowRunBackendId: backendIds.workflowRunBackendId,
workflowJobRunBackendId: backendIds.workflowJobRunBackendId,
key, key,
version, version,
sizeBytes: `${archiveFileSize}` sizeBytes: `${archiveFileSize}`
@ -6456,18 +6447,18 @@ const runtime_3 = __nccwpck_require__(3503);
const runtime_4 = __nccwpck_require__(3503); const runtime_4 = __nccwpck_require__(3503);
const runtime_5 = __nccwpck_require__(3503); const runtime_5 = __nccwpck_require__(3503);
const timestamp_1 = __nccwpck_require__(8983); const timestamp_1 = __nccwpck_require__(8983);
const cachemetadata_1 = __nccwpck_require__(2773);
// @generated message type with reflection information, may provide speed optimized methods // @generated message type with reflection information, may provide speed optimized methods
class CreateCacheEntryRequest$Type extends runtime_5.MessageType { class CreateCacheEntryRequest$Type extends runtime_5.MessageType {
constructor() { constructor() {
super("github.actions.results.api.v1.CreateCacheEntryRequest", [ super("github.actions.results.api.v1.CreateCacheEntryRequest", [
{ no: 1, name: "workflow_run_backend_id", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 1, name: "metadata", kind: "message", T: () => cachemetadata_1.CacheMetadata },
{ no: 2, name: "workflow_job_run_backend_id", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 2, name: "key", kind: "scalar", T: 9 /*ScalarType.STRING*/ },
{ no: 3, name: "key", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 3, name: "version", kind: "scalar", T: 9 /*ScalarType.STRING*/ }
{ no: 4, name: "version", kind: "scalar", T: 9 /*ScalarType.STRING*/ }
]); ]);
} }
create(value) { create(value) {
const message = { workflowRunBackendId: "", workflowJobRunBackendId: "", key: "", version: "" }; const message = { key: "", version: "" };
globalThis.Object.defineProperty(message, runtime_4.MESSAGE_TYPE, { enumerable: false, value: this }); globalThis.Object.defineProperty(message, runtime_4.MESSAGE_TYPE, { enumerable: false, value: this });
if (value !== undefined) if (value !== undefined)
(0, runtime_3.reflectionMergePartial)(this, message, value); (0, runtime_3.reflectionMergePartial)(this, message, value);
@ -6478,16 +6469,13 @@ class CreateCacheEntryRequest$Type extends runtime_5.MessageType {
while (reader.pos < end) { while (reader.pos < end) {
let [fieldNo, wireType] = reader.tag(); let [fieldNo, wireType] = reader.tag();
switch (fieldNo) { switch (fieldNo) {
case /* string workflow_run_backend_id */ 1: case /* github.actions.results.entities.v1.CacheMetadata metadata */ 1:
message.workflowRunBackendId = reader.string(); message.metadata = cachemetadata_1.CacheMetadata.internalBinaryRead(reader, reader.uint32(), options, message.metadata);
break; break;
case /* string workflow_job_run_backend_id */ 2: case /* string key */ 2:
message.workflowJobRunBackendId = reader.string();
break;
case /* string key */ 3:
message.key = reader.string(); message.key = reader.string();
break; break;
case /* string version */ 4: case /* string version */ 3:
message.version = reader.string(); message.version = reader.string();
break; break;
default: default:
@ -6502,18 +6490,15 @@ class CreateCacheEntryRequest$Type extends runtime_5.MessageType {
return message; return message;
} }
internalBinaryWrite(message, writer, options) { internalBinaryWrite(message, writer, options) {
/* string workflow_run_backend_id = 1; */ /* github.actions.results.entities.v1.CacheMetadata metadata = 1; */
if (message.workflowRunBackendId !== "") if (message.metadata)
writer.tag(1, runtime_1.WireType.LengthDelimited).string(message.workflowRunBackendId); cachemetadata_1.CacheMetadata.internalBinaryWrite(message.metadata, writer.tag(1, runtime_1.WireType.LengthDelimited).fork(), options).join();
/* string workflow_job_run_backend_id = 2; */ /* string key = 2; */
if (message.workflowJobRunBackendId !== "")
writer.tag(2, runtime_1.WireType.LengthDelimited).string(message.workflowJobRunBackendId);
/* string key = 3; */
if (message.key !== "") if (message.key !== "")
writer.tag(3, runtime_1.WireType.LengthDelimited).string(message.key); writer.tag(2, runtime_1.WireType.LengthDelimited).string(message.key);
/* string version = 4; */ /* string version = 3; */
if (message.version !== "") if (message.version !== "")
writer.tag(4, runtime_1.WireType.LengthDelimited).string(message.version); writer.tag(3, runtime_1.WireType.LengthDelimited).string(message.version);
let u = options.writeUnknownFields; let u = options.writeUnknownFields;
if (u !== false) if (u !== false)
(u == true ? runtime_2.UnknownFieldHandler.onWrite : u)(this.typeName, message, writer); (u == true ? runtime_2.UnknownFieldHandler.onWrite : u)(this.typeName, message, writer);
@ -6582,15 +6567,14 @@ exports.CreateCacheEntryResponse = new CreateCacheEntryResponse$Type();
class FinalizeCacheEntryUploadRequest$Type extends runtime_5.MessageType { class FinalizeCacheEntryUploadRequest$Type extends runtime_5.MessageType {
constructor() { constructor() {
super("github.actions.results.api.v1.FinalizeCacheEntryUploadRequest", [ super("github.actions.results.api.v1.FinalizeCacheEntryUploadRequest", [
{ no: 1, name: "workflow_run_backend_id", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 1, name: "metadata", kind: "message", T: () => cachemetadata_1.CacheMetadata },
{ no: 2, name: "workflow_job_run_backend_id", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 2, name: "key", kind: "scalar", T: 9 /*ScalarType.STRING*/ },
{ no: 3, name: "key", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 3, name: "size_bytes", kind: "scalar", T: 3 /*ScalarType.INT64*/ },
{ no: 4, name: "size_bytes", kind: "scalar", T: 3 /*ScalarType.INT64*/ }, { no: 4, name: "version", kind: "scalar", T: 9 /*ScalarType.STRING*/ }
{ no: 5, name: "version", kind: "scalar", T: 9 /*ScalarType.STRING*/ }
]); ]);
} }
create(value) { create(value) {
const message = { workflowRunBackendId: "", workflowJobRunBackendId: "", key: "", sizeBytes: "0", version: "" }; const message = { key: "", sizeBytes: "0", version: "" };
globalThis.Object.defineProperty(message, runtime_4.MESSAGE_TYPE, { enumerable: false, value: this }); globalThis.Object.defineProperty(message, runtime_4.MESSAGE_TYPE, { enumerable: false, value: this });
if (value !== undefined) if (value !== undefined)
(0, runtime_3.reflectionMergePartial)(this, message, value); (0, runtime_3.reflectionMergePartial)(this, message, value);
@ -6601,19 +6585,16 @@ class FinalizeCacheEntryUploadRequest$Type extends runtime_5.MessageType {
while (reader.pos < end) { while (reader.pos < end) {
let [fieldNo, wireType] = reader.tag(); let [fieldNo, wireType] = reader.tag();
switch (fieldNo) { switch (fieldNo) {
case /* string workflow_run_backend_id */ 1: case /* github.actions.results.entities.v1.CacheMetadata metadata */ 1:
message.workflowRunBackendId = reader.string(); message.metadata = cachemetadata_1.CacheMetadata.internalBinaryRead(reader, reader.uint32(), options, message.metadata);
break; break;
case /* string workflow_job_run_backend_id */ 2: case /* string key */ 2:
message.workflowJobRunBackendId = reader.string();
break;
case /* string key */ 3:
message.key = reader.string(); message.key = reader.string();
break; break;
case /* int64 size_bytes */ 4: case /* int64 size_bytes */ 3:
message.sizeBytes = reader.int64().toString(); message.sizeBytes = reader.int64().toString();
break; break;
case /* string version */ 5: case /* string version */ 4:
message.version = reader.string(); message.version = reader.string();
break; break;
default: default:
@ -6628,21 +6609,18 @@ class FinalizeCacheEntryUploadRequest$Type extends runtime_5.MessageType {
return message; return message;
} }
internalBinaryWrite(message, writer, options) { internalBinaryWrite(message, writer, options) {
/* string workflow_run_backend_id = 1; */ /* github.actions.results.entities.v1.CacheMetadata metadata = 1; */
if (message.workflowRunBackendId !== "") if (message.metadata)
writer.tag(1, runtime_1.WireType.LengthDelimited).string(message.workflowRunBackendId); cachemetadata_1.CacheMetadata.internalBinaryWrite(message.metadata, writer.tag(1, runtime_1.WireType.LengthDelimited).fork(), options).join();
/* string workflow_job_run_backend_id = 2; */ /* string key = 2; */
if (message.workflowJobRunBackendId !== "")
writer.tag(2, runtime_1.WireType.LengthDelimited).string(message.workflowJobRunBackendId);
/* string key = 3; */
if (message.key !== "") if (message.key !== "")
writer.tag(3, runtime_1.WireType.LengthDelimited).string(message.key); writer.tag(2, runtime_1.WireType.LengthDelimited).string(message.key);
/* int64 size_bytes = 4; */ /* int64 size_bytes = 3; */
if (message.sizeBytes !== "0") if (message.sizeBytes !== "0")
writer.tag(4, runtime_1.WireType.Varint).int64(message.sizeBytes); writer.tag(3, runtime_1.WireType.Varint).int64(message.sizeBytes);
/* string version = 5; */ /* string version = 4; */
if (message.version !== "") if (message.version !== "")
writer.tag(5, runtime_1.WireType.LengthDelimited).string(message.version); writer.tag(4, runtime_1.WireType.LengthDelimited).string(message.version);
let u = options.writeUnknownFields; let u = options.writeUnknownFields;
if (u !== false) if (u !== false)
(u == true ? runtime_2.UnknownFieldHandler.onWrite : u)(this.typeName, message, writer); (u == true ? runtime_2.UnknownFieldHandler.onWrite : u)(this.typeName, message, writer);
@ -6711,15 +6689,14 @@ exports.FinalizeCacheEntryUploadResponse = new FinalizeCacheEntryUploadResponse$
class GetCacheEntryDownloadURLRequest$Type extends runtime_5.MessageType { class GetCacheEntryDownloadURLRequest$Type extends runtime_5.MessageType {
constructor() { constructor() {
super("github.actions.results.api.v1.GetCacheEntryDownloadURLRequest", [ super("github.actions.results.api.v1.GetCacheEntryDownloadURLRequest", [
{ no: 1, name: "workflow_run_backend_id", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 1, name: "metadata", kind: "message", T: () => cachemetadata_1.CacheMetadata },
{ no: 2, name: "workflow_job_run_backend_id", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 2, name: "key", kind: "scalar", T: 9 /*ScalarType.STRING*/ },
{ no: 3, name: "key", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 3, name: "restore_keys", kind: "scalar", repeat: 2 /*RepeatType.UNPACKED*/, T: 9 /*ScalarType.STRING*/ },
{ no: 4, name: "restore_keys", kind: "scalar", repeat: 2 /*RepeatType.UNPACKED*/, T: 9 /*ScalarType.STRING*/ }, { no: 4, name: "version", kind: "scalar", T: 9 /*ScalarType.STRING*/ }
{ no: 5, name: "version", kind: "scalar", T: 9 /*ScalarType.STRING*/ }
]); ]);
} }
create(value) { create(value) {
const message = { workflowRunBackendId: "", workflowJobRunBackendId: "", key: "", restoreKeys: [], version: "" }; const message = { key: "", restoreKeys: [], version: "" };
globalThis.Object.defineProperty(message, runtime_4.MESSAGE_TYPE, { enumerable: false, value: this }); globalThis.Object.defineProperty(message, runtime_4.MESSAGE_TYPE, { enumerable: false, value: this });
if (value !== undefined) if (value !== undefined)
(0, runtime_3.reflectionMergePartial)(this, message, value); (0, runtime_3.reflectionMergePartial)(this, message, value);
@ -6730,19 +6707,16 @@ class GetCacheEntryDownloadURLRequest$Type extends runtime_5.MessageType {
while (reader.pos < end) { while (reader.pos < end) {
let [fieldNo, wireType] = reader.tag(); let [fieldNo, wireType] = reader.tag();
switch (fieldNo) { switch (fieldNo) {
case /* string workflow_run_backend_id */ 1: case /* github.actions.results.entities.v1.CacheMetadata metadata */ 1:
message.workflowRunBackendId = reader.string(); message.metadata = cachemetadata_1.CacheMetadata.internalBinaryRead(reader, reader.uint32(), options, message.metadata);
break; break;
case /* string workflow_job_run_backend_id */ 2: case /* string key */ 2:
message.workflowJobRunBackendId = reader.string();
break;
case /* string key */ 3:
message.key = reader.string(); message.key = reader.string();
break; break;
case /* repeated string restore_keys */ 4: case /* repeated string restore_keys */ 3:
message.restoreKeys.push(reader.string()); message.restoreKeys.push(reader.string());
break; break;
case /* string version */ 5: case /* string version */ 4:
message.version = reader.string(); message.version = reader.string();
break; break;
default: default:
@ -6757,21 +6731,18 @@ class GetCacheEntryDownloadURLRequest$Type extends runtime_5.MessageType {
return message; return message;
} }
internalBinaryWrite(message, writer, options) { internalBinaryWrite(message, writer, options) {
/* string workflow_run_backend_id = 1; */ /* github.actions.results.entities.v1.CacheMetadata metadata = 1; */
if (message.workflowRunBackendId !== "") if (message.metadata)
writer.tag(1, runtime_1.WireType.LengthDelimited).string(message.workflowRunBackendId); cachemetadata_1.CacheMetadata.internalBinaryWrite(message.metadata, writer.tag(1, runtime_1.WireType.LengthDelimited).fork(), options).join();
/* string workflow_job_run_backend_id = 2; */ /* string key = 2; */
if (message.workflowJobRunBackendId !== "")
writer.tag(2, runtime_1.WireType.LengthDelimited).string(message.workflowJobRunBackendId);
/* string key = 3; */
if (message.key !== "") if (message.key !== "")
writer.tag(3, runtime_1.WireType.LengthDelimited).string(message.key); writer.tag(2, runtime_1.WireType.LengthDelimited).string(message.key);
/* repeated string restore_keys = 4; */ /* repeated string restore_keys = 3; */
for (let i = 0; i < message.restoreKeys.length; i++) for (let i = 0; i < message.restoreKeys.length; i++)
writer.tag(4, runtime_1.WireType.LengthDelimited).string(message.restoreKeys[i]); writer.tag(3, runtime_1.WireType.LengthDelimited).string(message.restoreKeys[i]);
/* string version = 5; */ /* string version = 4; */
if (message.version !== "") if (message.version !== "")
writer.tag(5, runtime_1.WireType.LengthDelimited).string(message.version); writer.tag(4, runtime_1.WireType.LengthDelimited).string(message.version);
let u = options.writeUnknownFields; let u = options.writeUnknownFields;
if (u !== false) if (u !== false)
(u == true ? runtime_2.UnknownFieldHandler.onWrite : u)(this.typeName, message, writer); (u == true ? runtime_2.UnknownFieldHandler.onWrite : u)(this.typeName, message, writer);
@ -6840,13 +6811,12 @@ exports.GetCacheEntryDownloadURLResponse = new GetCacheEntryDownloadURLResponse$
class DeleteCacheEntryRequest$Type extends runtime_5.MessageType { class DeleteCacheEntryRequest$Type extends runtime_5.MessageType {
constructor() { constructor() {
super("github.actions.results.api.v1.DeleteCacheEntryRequest", [ super("github.actions.results.api.v1.DeleteCacheEntryRequest", [
{ no: 1, name: "workflow_run_backend_id", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 1, name: "metadata", kind: "message", T: () => cachemetadata_1.CacheMetadata },
{ no: 2, name: "workflow_job_run_backend_id", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 2, name: "key", kind: "scalar", T: 9 /*ScalarType.STRING*/ }
{ no: 3, name: "key", kind: "scalar", T: 9 /*ScalarType.STRING*/ }
]); ]);
} }
create(value) { create(value) {
const message = { workflowRunBackendId: "", workflowJobRunBackendId: "", key: "" }; const message = { key: "" };
globalThis.Object.defineProperty(message, runtime_4.MESSAGE_TYPE, { enumerable: false, value: this }); globalThis.Object.defineProperty(message, runtime_4.MESSAGE_TYPE, { enumerable: false, value: this });
if (value !== undefined) if (value !== undefined)
(0, runtime_3.reflectionMergePartial)(this, message, value); (0, runtime_3.reflectionMergePartial)(this, message, value);
@ -6857,13 +6827,10 @@ class DeleteCacheEntryRequest$Type extends runtime_5.MessageType {
while (reader.pos < end) { while (reader.pos < end) {
let [fieldNo, wireType] = reader.tag(); let [fieldNo, wireType] = reader.tag();
switch (fieldNo) { switch (fieldNo) {
case /* string workflow_run_backend_id */ 1: case /* github.actions.results.entities.v1.CacheMetadata metadata */ 1:
message.workflowRunBackendId = reader.string(); message.metadata = cachemetadata_1.CacheMetadata.internalBinaryRead(reader, reader.uint32(), options, message.metadata);
break; break;
case /* string workflow_job_run_backend_id */ 2: case /* string key */ 2:
message.workflowJobRunBackendId = reader.string();
break;
case /* string key */ 3:
message.key = reader.string(); message.key = reader.string();
break; break;
default: default:
@ -6878,15 +6845,12 @@ class DeleteCacheEntryRequest$Type extends runtime_5.MessageType {
return message; return message;
} }
internalBinaryWrite(message, writer, options) { internalBinaryWrite(message, writer, options) {
/* string workflow_run_backend_id = 1; */ /* github.actions.results.entities.v1.CacheMetadata metadata = 1; */
if (message.workflowRunBackendId !== "") if (message.metadata)
writer.tag(1, runtime_1.WireType.LengthDelimited).string(message.workflowRunBackendId); cachemetadata_1.CacheMetadata.internalBinaryWrite(message.metadata, writer.tag(1, runtime_1.WireType.LengthDelimited).fork(), options).join();
/* string workflow_job_run_backend_id = 2; */ /* string key = 2; */
if (message.workflowJobRunBackendId !== "")
writer.tag(2, runtime_1.WireType.LengthDelimited).string(message.workflowJobRunBackendId);
/* string key = 3; */
if (message.key !== "") if (message.key !== "")
writer.tag(3, runtime_1.WireType.LengthDelimited).string(message.key); writer.tag(2, runtime_1.WireType.LengthDelimited).string(message.key);
let u = options.writeUnknownFields; let u = options.writeUnknownFields;
if (u !== false) if (u !== false)
(u == true ? runtime_2.UnknownFieldHandler.onWrite : u)(this.typeName, message, writer); (u == true ? runtime_2.UnknownFieldHandler.onWrite : u)(this.typeName, message, writer);
@ -6955,14 +6919,13 @@ exports.DeleteCacheEntryResponse = new DeleteCacheEntryResponse$Type();
class ListCacheEntriesRequest$Type extends runtime_5.MessageType { class ListCacheEntriesRequest$Type extends runtime_5.MessageType {
constructor() { constructor() {
super("github.actions.results.api.v1.ListCacheEntriesRequest", [ super("github.actions.results.api.v1.ListCacheEntriesRequest", [
{ no: 1, name: "workflow_run_backend_id", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 1, name: "metadata", kind: "message", T: () => cachemetadata_1.CacheMetadata },
{ no: 2, name: "workflow_job_run_backend_id", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 2, name: "key", kind: "scalar", T: 9 /*ScalarType.STRING*/ },
{ no: 3, name: "key", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 3, name: "restore_keys", kind: "scalar", repeat: 2 /*RepeatType.UNPACKED*/, T: 9 /*ScalarType.STRING*/ }
{ no: 4, name: "restore_keys", kind: "scalar", repeat: 2 /*RepeatType.UNPACKED*/, T: 9 /*ScalarType.STRING*/ }
]); ]);
} }
create(value) { create(value) {
const message = { workflowRunBackendId: "", workflowJobRunBackendId: "", key: "", restoreKeys: [] }; const message = { key: "", restoreKeys: [] };
globalThis.Object.defineProperty(message, runtime_4.MESSAGE_TYPE, { enumerable: false, value: this }); globalThis.Object.defineProperty(message, runtime_4.MESSAGE_TYPE, { enumerable: false, value: this });
if (value !== undefined) if (value !== undefined)
(0, runtime_3.reflectionMergePartial)(this, message, value); (0, runtime_3.reflectionMergePartial)(this, message, value);
@ -6973,16 +6936,13 @@ class ListCacheEntriesRequest$Type extends runtime_5.MessageType {
while (reader.pos < end) { while (reader.pos < end) {
let [fieldNo, wireType] = reader.tag(); let [fieldNo, wireType] = reader.tag();
switch (fieldNo) { switch (fieldNo) {
case /* string workflow_run_backend_id */ 1: case /* github.actions.results.entities.v1.CacheMetadata metadata */ 1:
message.workflowRunBackendId = reader.string(); message.metadata = cachemetadata_1.CacheMetadata.internalBinaryRead(reader, reader.uint32(), options, message.metadata);
break; break;
case /* string workflow_job_run_backend_id */ 2: case /* string key */ 2:
message.workflowJobRunBackendId = reader.string();
break;
case /* string key */ 3:
message.key = reader.string(); message.key = reader.string();
break; break;
case /* repeated string restore_keys */ 4: case /* repeated string restore_keys */ 3:
message.restoreKeys.push(reader.string()); message.restoreKeys.push(reader.string());
break; break;
default: default:
@ -6997,18 +6957,15 @@ class ListCacheEntriesRequest$Type extends runtime_5.MessageType {
return message; return message;
} }
internalBinaryWrite(message, writer, options) { internalBinaryWrite(message, writer, options) {
/* string workflow_run_backend_id = 1; */ /* github.actions.results.entities.v1.CacheMetadata metadata = 1; */
if (message.workflowRunBackendId !== "") if (message.metadata)
writer.tag(1, runtime_1.WireType.LengthDelimited).string(message.workflowRunBackendId); cachemetadata_1.CacheMetadata.internalBinaryWrite(message.metadata, writer.tag(1, runtime_1.WireType.LengthDelimited).fork(), options).join();
/* string workflow_job_run_backend_id = 2; */ /* string key = 2; */
if (message.workflowJobRunBackendId !== "")
writer.tag(2, runtime_1.WireType.LengthDelimited).string(message.workflowJobRunBackendId);
/* string key = 3; */
if (message.key !== "") if (message.key !== "")
writer.tag(3, runtime_1.WireType.LengthDelimited).string(message.key); writer.tag(2, runtime_1.WireType.LengthDelimited).string(message.key);
/* repeated string restore_keys = 4; */ /* repeated string restore_keys = 3; */
for (let i = 0; i < message.restoreKeys.length; i++) for (let i = 0; i < message.restoreKeys.length; i++)
writer.tag(4, runtime_1.WireType.LengthDelimited).string(message.restoreKeys[i]); writer.tag(3, runtime_1.WireType.LengthDelimited).string(message.restoreKeys[i]);
let u = options.writeUnknownFields; let u = options.writeUnknownFields;
if (u !== false) if (u !== false)
(u == true ? runtime_2.UnknownFieldHandler.onWrite : u)(this.typeName, message, writer); (u == true ? runtime_2.UnknownFieldHandler.onWrite : u)(this.typeName, message, writer);
@ -7166,15 +7123,14 @@ exports.ListCacheEntriesResponse_CacheEntry = new ListCacheEntriesResponse_Cache
class LookupCacheEntryRequest$Type extends runtime_5.MessageType { class LookupCacheEntryRequest$Type extends runtime_5.MessageType {
constructor() { constructor() {
super("github.actions.results.api.v1.LookupCacheEntryRequest", [ super("github.actions.results.api.v1.LookupCacheEntryRequest", [
{ no: 1, name: "workflow_run_backend_id", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 1, name: "metadata", kind: "message", T: () => cachemetadata_1.CacheMetadata },
{ no: 2, name: "workflow_job_run_backend_id", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 2, name: "key", kind: "scalar", T: 9 /*ScalarType.STRING*/ },
{ no: 3, name: "key", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 3, name: "restore_keys", kind: "scalar", repeat: 2 /*RepeatType.UNPACKED*/, T: 9 /*ScalarType.STRING*/ },
{ no: 4, name: "restore_keys", kind: "scalar", repeat: 2 /*RepeatType.UNPACKED*/, T: 9 /*ScalarType.STRING*/ }, { no: 4, name: "version", kind: "scalar", T: 9 /*ScalarType.STRING*/ }
{ no: 5, name: "version", kind: "scalar", T: 9 /*ScalarType.STRING*/ }
]); ]);
} }
create(value) { create(value) {
const message = { workflowRunBackendId: "", workflowJobRunBackendId: "", key: "", restoreKeys: [], version: "" }; const message = { key: "", restoreKeys: [], version: "" };
globalThis.Object.defineProperty(message, runtime_4.MESSAGE_TYPE, { enumerable: false, value: this }); globalThis.Object.defineProperty(message, runtime_4.MESSAGE_TYPE, { enumerable: false, value: this });
if (value !== undefined) if (value !== undefined)
(0, runtime_3.reflectionMergePartial)(this, message, value); (0, runtime_3.reflectionMergePartial)(this, message, value);
@ -7185,19 +7141,16 @@ class LookupCacheEntryRequest$Type extends runtime_5.MessageType {
while (reader.pos < end) { while (reader.pos < end) {
let [fieldNo, wireType] = reader.tag(); let [fieldNo, wireType] = reader.tag();
switch (fieldNo) { switch (fieldNo) {
case /* string workflow_run_backend_id */ 1: case /* github.actions.results.entities.v1.CacheMetadata metadata */ 1:
message.workflowRunBackendId = reader.string(); message.metadata = cachemetadata_1.CacheMetadata.internalBinaryRead(reader, reader.uint32(), options, message.metadata);
break; break;
case /* string workflow_job_run_backend_id */ 2: case /* string key */ 2:
message.workflowJobRunBackendId = reader.string();
break;
case /* string key */ 3:
message.key = reader.string(); message.key = reader.string();
break; break;
case /* repeated string restore_keys */ 4: case /* repeated string restore_keys */ 3:
message.restoreKeys.push(reader.string()); message.restoreKeys.push(reader.string());
break; break;
case /* string version */ 5: case /* string version */ 4:
message.version = reader.string(); message.version = reader.string();
break; break;
default: default:
@ -7212,21 +7165,18 @@ class LookupCacheEntryRequest$Type extends runtime_5.MessageType {
return message; return message;
} }
internalBinaryWrite(message, writer, options) { internalBinaryWrite(message, writer, options) {
/* string workflow_run_backend_id = 1; */ /* github.actions.results.entities.v1.CacheMetadata metadata = 1; */
if (message.workflowRunBackendId !== "") if (message.metadata)
writer.tag(1, runtime_1.WireType.LengthDelimited).string(message.workflowRunBackendId); cachemetadata_1.CacheMetadata.internalBinaryWrite(message.metadata, writer.tag(1, runtime_1.WireType.LengthDelimited).fork(), options).join();
/* string workflow_job_run_backend_id = 2; */ /* string key = 2; */
if (message.workflowJobRunBackendId !== "")
writer.tag(2, runtime_1.WireType.LengthDelimited).string(message.workflowJobRunBackendId);
/* string key = 3; */
if (message.key !== "") if (message.key !== "")
writer.tag(3, runtime_1.WireType.LengthDelimited).string(message.key); writer.tag(2, runtime_1.WireType.LengthDelimited).string(message.key);
/* repeated string restore_keys = 4; */ /* repeated string restore_keys = 3; */
for (let i = 0; i < message.restoreKeys.length; i++) for (let i = 0; i < message.restoreKeys.length; i++)
writer.tag(4, runtime_1.WireType.LengthDelimited).string(message.restoreKeys[i]); writer.tag(3, runtime_1.WireType.LengthDelimited).string(message.restoreKeys[i]);
/* string version = 5; */ /* string version = 4; */
if (message.version !== "") if (message.version !== "")
writer.tag(5, runtime_1.WireType.LengthDelimited).string(message.version); writer.tag(4, runtime_1.WireType.LengthDelimited).string(message.version);
let u = options.writeUnknownFields; let u = options.writeUnknownFields;
if (u !== false) if (u !== false)
(u == true ? runtime_2.UnknownFieldHandler.onWrite : u)(this.typeName, message, writer); (u == true ? runtime_2.UnknownFieldHandler.onWrite : u)(this.typeName, message, writer);
@ -8004,6 +7954,147 @@ function handleCacheServiceLookupCacheEntryProtobuf(ctx, service, data, intercep
/***/ }), /***/ }),
/***/ 2773:
/***/ ((__unused_webpack_module, exports, __nccwpck_require__) => {
"use strict";
Object.defineProperty(exports, "__esModule", ({ value: true }));
exports.CacheMetadata = void 0;
const runtime_1 = __nccwpck_require__(3503);
const runtime_2 = __nccwpck_require__(3503);
const runtime_3 = __nccwpck_require__(3503);
const runtime_4 = __nccwpck_require__(3503);
const runtime_5 = __nccwpck_require__(3503);
const cachescope_1 = __nccwpck_require__(7685);
// @generated message type with reflection information, may provide speed optimized methods
class CacheMetadata$Type extends runtime_5.MessageType {
constructor() {
super("github.actions.results.entities.v1.CacheMetadata", [
{ no: 1, name: "repository_id", kind: "scalar", T: 3 /*ScalarType.INT64*/ },
{ no: 2, name: "scope", kind: "message", repeat: 1 /*RepeatType.PACKED*/, T: () => cachescope_1.CacheScope }
]);
}
create(value) {
const message = { repositoryId: "0", scope: [] };
globalThis.Object.defineProperty(message, runtime_4.MESSAGE_TYPE, { enumerable: false, value: this });
if (value !== undefined)
(0, runtime_3.reflectionMergePartial)(this, message, value);
return message;
}
internalBinaryRead(reader, length, options, target) {
let message = target !== null && target !== void 0 ? target : this.create(), end = reader.pos + length;
while (reader.pos < end) {
let [fieldNo, wireType] = reader.tag();
switch (fieldNo) {
case /* int64 repository_id */ 1:
message.repositoryId = reader.int64().toString();
break;
case /* repeated github.actions.results.entities.v1.CacheScope scope */ 2:
message.scope.push(cachescope_1.CacheScope.internalBinaryRead(reader, reader.uint32(), options));
break;
default:
let u = options.readUnknownField;
if (u === "throw")
throw new globalThis.Error(`Unknown field ${fieldNo} (wire type ${wireType}) for ${this.typeName}`);
let d = reader.skip(wireType);
if (u !== false)
(u === true ? runtime_2.UnknownFieldHandler.onRead : u)(this.typeName, message, fieldNo, wireType, d);
}
}
return message;
}
internalBinaryWrite(message, writer, options) {
/* int64 repository_id = 1; */
if (message.repositoryId !== "0")
writer.tag(1, runtime_1.WireType.Varint).int64(message.repositoryId);
/* repeated github.actions.results.entities.v1.CacheScope scope = 2; */
for (let i = 0; i < message.scope.length; i++)
cachescope_1.CacheScope.internalBinaryWrite(message.scope[i], writer.tag(2, runtime_1.WireType.LengthDelimited).fork(), options).join();
let u = options.writeUnknownFields;
if (u !== false)
(u == true ? runtime_2.UnknownFieldHandler.onWrite : u)(this.typeName, message, writer);
return writer;
}
}
/**
* @generated MessageType for protobuf message github.actions.results.entities.v1.CacheMetadata
*/
exports.CacheMetadata = new CacheMetadata$Type();
//# sourceMappingURL=cachemetadata.js.map
/***/ }),
/***/ 7685:
/***/ ((__unused_webpack_module, exports, __nccwpck_require__) => {
"use strict";
Object.defineProperty(exports, "__esModule", ({ value: true }));
exports.CacheScope = void 0;
const runtime_1 = __nccwpck_require__(3503);
const runtime_2 = __nccwpck_require__(3503);
const runtime_3 = __nccwpck_require__(3503);
const runtime_4 = __nccwpck_require__(3503);
const runtime_5 = __nccwpck_require__(3503);
// @generated message type with reflection information, may provide speed optimized methods
class CacheScope$Type extends runtime_5.MessageType {
constructor() {
super("github.actions.results.entities.v1.CacheScope", [
{ no: 1, name: "scope", kind: "scalar", T: 9 /*ScalarType.STRING*/ },
{ no: 2, name: "permission", kind: "scalar", T: 3 /*ScalarType.INT64*/ }
]);
}
create(value) {
const message = { scope: "", permission: "0" };
globalThis.Object.defineProperty(message, runtime_4.MESSAGE_TYPE, { enumerable: false, value: this });
if (value !== undefined)
(0, runtime_3.reflectionMergePartial)(this, message, value);
return message;
}
internalBinaryRead(reader, length, options, target) {
let message = target !== null && target !== void 0 ? target : this.create(), end = reader.pos + length;
while (reader.pos < end) {
let [fieldNo, wireType] = reader.tag();
switch (fieldNo) {
case /* string scope */ 1:
message.scope = reader.string();
break;
case /* int64 permission */ 2:
message.permission = reader.int64().toString();
break;
default:
let u = options.readUnknownField;
if (u === "throw")
throw new globalThis.Error(`Unknown field ${fieldNo} (wire type ${wireType}) for ${this.typeName}`);
let d = reader.skip(wireType);
if (u !== false)
(u === true ? runtime_2.UnknownFieldHandler.onRead : u)(this.typeName, message, fieldNo, wireType, d);
}
}
return message;
}
internalBinaryWrite(message, writer, options) {
/* string scope = 1; */
if (message.scope !== "")
writer.tag(1, runtime_1.WireType.LengthDelimited).string(message.scope);
/* int64 permission = 2; */
if (message.permission !== "0")
writer.tag(2, runtime_1.WireType.Varint).int64(message.permission);
let u = options.writeUnknownFields;
if (u !== false)
(u == true ? runtime_2.UnknownFieldHandler.onWrite : u)(this.typeName, message, writer);
return writer;
}
}
/**
* @generated MessageType for protobuf message github.actions.results.entities.v1.CacheScope
*/
exports.CacheScope = new CacheScope$Type();
//# sourceMappingURL=cachescope.js.map
/***/ }),
/***/ 4076: /***/ 4076:
/***/ (function(__unused_webpack_module, exports, __nccwpck_require__) { /***/ (function(__unused_webpack_module, exports, __nccwpck_require__) {
@ -8416,11 +8507,8 @@ var __asyncValues = (this && this.__asyncValues) || function (o) {
function verb(n) { i[n] = o[n] && function (v) { return new Promise(function (resolve, reject) { v = o[n](v), settle(resolve, reject, v.done, v.value); }); }; } function verb(n) { i[n] = o[n] && function (v) { return new Promise(function (resolve, reject) { v = o[n](v), settle(resolve, reject, v.done, v.value); }); }; }
function settle(resolve, reject, d, v) { Promise.resolve(v).then(function(v) { resolve({ value: v, done: d }); }, reject); } function settle(resolve, reject, d, v) { Promise.resolve(v).then(function(v) { resolve({ value: v, done: d }); }, reject); }
}; };
var __importDefault = (this && this.__importDefault) || function (mod) {
return (mod && mod.__esModule) ? mod : { "default": mod };
};
Object.defineProperty(exports, "__esModule", ({ value: true })); Object.defineProperty(exports, "__esModule", ({ value: true }));
exports.getBackendIdsFromToken = exports.getRuntimeToken = exports.getCacheVersion = exports.isGhes = exports.assertDefined = exports.getGnuTarPathOnWindows = exports.getCacheFileName = exports.getCompressionMethod = exports.unlinkFile = exports.resolvePaths = exports.getArchiveFileSizeInBytes = exports.createTempDirectory = void 0; exports.getRuntimeToken = exports.getCacheVersion = exports.isGhes = exports.assertDefined = exports.getGnuTarPathOnWindows = exports.getCacheFileName = exports.getCompressionMethod = exports.unlinkFile = exports.resolvePaths = exports.getArchiveFileSizeInBytes = exports.createTempDirectory = void 0;
const core = __importStar(__nccwpck_require__(4850)); const core = __importStar(__nccwpck_require__(4850));
const exec = __importStar(__nccwpck_require__(309)); const exec = __importStar(__nccwpck_require__(309));
const glob = __importStar(__nccwpck_require__(9590)); const glob = __importStar(__nccwpck_require__(9590));
@ -8430,7 +8518,6 @@ const fs = __importStar(__nccwpck_require__(7147));
const path = __importStar(__nccwpck_require__(1017)); const path = __importStar(__nccwpck_require__(1017));
const semver = __importStar(__nccwpck_require__(3910)); const semver = __importStar(__nccwpck_require__(3910));
const util = __importStar(__nccwpck_require__(3837)); const util = __importStar(__nccwpck_require__(3837));
const jwt_decode_1 = __importDefault(__nccwpck_require__(1300));
const constants_1 = __nccwpck_require__(4010); const constants_1 = __nccwpck_require__(4010);
const versionSalt = '1.0'; const versionSalt = '1.0';
// From https://github.com/actions/toolkit/blob/main/packages/tool-cache/src/tool-cache.ts#L23 // From https://github.com/actions/toolkit/blob/main/packages/tool-cache/src/tool-cache.ts#L23
@ -8603,54 +8690,6 @@ function getRuntimeToken() {
return token; return token;
} }
exports.getRuntimeToken = getRuntimeToken; exports.getRuntimeToken = getRuntimeToken;
const InvalidJwtError = new Error('Failed to get backend IDs: The provided JWT token is invalid and/or missing claims');
// uses the JWT token claims to get the
// workflow run and workflow job run backend ids
function getBackendIdsFromToken() {
const token = getRuntimeToken();
const decoded = (0, jwt_decode_1.default)(token);
if (!decoded.scp) {
throw InvalidJwtError;
}
/*
* example decoded:
* {
* scp: "Actions.ExampleScope Actions.Results:ce7f54c7-61c7-4aae-887f-30da475f5f1a:ca395085-040a-526b-2ce8-bdc85f692774"
* }
*/
const scpParts = decoded.scp.split(' ');
if (scpParts.length === 0) {
throw InvalidJwtError;
}
/*
* example scpParts:
* ["Actions.ExampleScope", "Actions.Results:ce7f54c7-61c7-4aae-887f-30da475f5f1a:ca395085-040a-526b-2ce8-bdc85f692774"]
*/
for (const scopes of scpParts) {
const scopeParts = scopes.split(':');
if ((scopeParts === null || scopeParts === void 0 ? void 0 : scopeParts[0]) !== 'Actions.Results') {
// not the Actions.Results scope
continue;
}
/*
* example scopeParts:
* ["Actions.Results", "ce7f54c7-61c7-4aae-887f-30da475f5f1a", "ca395085-040a-526b-2ce8-bdc85f692774"]
*/
if (scopeParts.length !== 3) {
// missing expected number of claims
throw InvalidJwtError;
}
const ids = {
workflowRunBackendId: scopeParts[1],
workflowJobRunBackendId: scopeParts[2]
};
core.debug(`Workflow Run Backend ID: ${ids.workflowRunBackendId}`);
core.debug(`Workflow Job Run Backend ID: ${ids.workflowJobRunBackendId}`);
return ids;
}
throw InvalidJwtError;
}
exports.getBackendIdsFromToken = getBackendIdsFromToken;
//# sourceMappingURL=cacheUtils.js.map //# sourceMappingURL=cacheUtils.js.map
/***/ }), /***/ }),
@ -8661,15 +8700,7 @@ exports.getBackendIdsFromToken = getBackendIdsFromToken;
"use strict"; "use strict";
Object.defineProperty(exports, "__esModule", ({ value: true })); Object.defineProperty(exports, "__esModule", ({ value: true }));
exports.getCacheServiceURL = exports.getCacheServiceVersion = exports.getRuntimeToken = void 0; exports.getCacheServiceURL = exports.getCacheServiceVersion = void 0;
function getRuntimeToken() {
const token = process.env['ACTIONS_RUNTIME_TOKEN'];
if (!token) {
throw new Error('Unable to get the ACTIONS_RUNTIME_TOKEN env variable');
}
return token;
}
exports.getRuntimeToken = getRuntimeToken;
function getCacheServiceVersion() { function getCacheServiceVersion() {
return process.env['ACTIONS_CACHE_SERVICE_V2'] ? 'v2' : 'v1'; return process.env['ACTIONS_CACHE_SERVICE_V2'] ? 'v2' : 'v1';
} }
@ -9285,6 +9316,7 @@ const core_1 = __nccwpck_require__(4850);
const user_agent_1 = __nccwpck_require__(5736); const user_agent_1 = __nccwpck_require__(5736);
const errors_1 = __nccwpck_require__(6333); const errors_1 = __nccwpck_require__(6333);
const config_1 = __nccwpck_require__(6490); const config_1 = __nccwpck_require__(6490);
const cacheUtils_1 = __nccwpck_require__(3310);
const auth_1 = __nccwpck_require__(7231); const auth_1 = __nccwpck_require__(7231);
const http_client_1 = __nccwpck_require__(1283); const http_client_1 = __nccwpck_require__(1283);
const cache_twirp_1 = __nccwpck_require__(894); const cache_twirp_1 = __nccwpck_require__(894);
@ -9300,7 +9332,7 @@ class CacheServiceClient {
this.maxAttempts = 5; this.maxAttempts = 5;
this.baseRetryIntervalMilliseconds = 3000; this.baseRetryIntervalMilliseconds = 3000;
this.retryMultiplier = 1.5; this.retryMultiplier = 1.5;
const token = (0, config_1.getRuntimeToken)(); const token = (0, cacheUtils_1.getRuntimeToken)();
this.baseUrl = (0, config_1.getCacheServiceURL)(); this.baseUrl = (0, config_1.getCacheServiceURL)();
if (maxAttempts) { if (maxAttempts) {
this.maxAttempts = maxAttempts; this.maxAttempts = maxAttempts;
@ -57677,16 +57709,6 @@ DotObject._process = _process
module.exports = DotObject module.exports = DotObject
/***/ }),
/***/ 1300:
/***/ ((module) => {
"use strict";
function e(e){this.message=e}e.prototype=new Error,e.prototype.name="InvalidCharacterError";var r="undefined"!=typeof window&&window.atob&&window.atob.bind(window)||function(r){var t=String(r).replace(/=+$/,"");if(t.length%4==1)throw new e("'atob' failed: The string to be decoded is not correctly encoded.");for(var n,o,a=0,i=0,c="";o=t.charAt(i++);~o&&(n=a%4?64*n+o:o,a++%4)?c+=String.fromCharCode(255&n>>(-2*a&6)):0)o="ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/=".indexOf(o);return c};function t(e){var t=e.replace(/-/g,"+").replace(/_/g,"/");switch(t.length%4){case 0:break;case 2:t+="==";break;case 3:t+="=";break;default:throw"Illegal base64url string!"}try{return function(e){return decodeURIComponent(r(e).replace(/(.)/g,(function(e,r){var t=r.charCodeAt(0).toString(16).toUpperCase();return t.length<2&&(t="0"+t),"%"+t})))}(t)}catch(e){return r(t)}}function n(e){this.message=e}function o(e,r){if("string"!=typeof e)throw new n("Invalid token specified");var o=!0===(r=r||{}).header?0:1;try{return JSON.parse(t(e.split(".")[o]))}catch(e){throw new n("Invalid token specified: "+e.message)}}n.prototype=new Error,n.prototype.name="InvalidTokenError";const a=o;a.default=o,a.InvalidTokenError=n,module.exports=a;
//# sourceMappingURL=jwt-decode.cjs.js.map
/***/ }), /***/ }),
/***/ 2020: /***/ 2020:
@ -73485,7 +73507,7 @@ module.exports = JSON.parse('[[[0,44],"disallowed_STD3_valid"],[[45,46],"valid"]
/***/ ((module) => { /***/ ((module) => {
"use strict"; "use strict";
module.exports = JSON.parse('{"name":"@actions/cache","version":"3.3.0","preview":true,"description":"Actions cache lib","keywords":["github","actions","cache"],"homepage":"https://github.com/actions/toolkit/tree/main/packages/cache","license":"MIT","main":"lib/cache.js","types":"lib/cache.d.ts","directories":{"lib":"lib","test":"__tests__"},"files":["lib","!.DS_Store"],"publishConfig":{"access":"public"},"repository":{"type":"git","url":"git+https://github.com/actions/toolkit.git","directory":"packages/cache"},"scripts":{"audit-moderate":"npm install && npm audit --json --audit-level=moderate > audit.json","test":"echo \\"Error: run tests from root\\" && exit 1","tsc":"tsc"},"bugs":{"url":"https://github.com/actions/toolkit/issues"},"dependencies":{"@actions/core":"^1.11.1","@actions/exec":"^1.0.1","@actions/glob":"^0.1.0","@actions/http-client":"^2.1.1","@actions/io":"^1.0.1","@azure/abort-controller":"^1.1.0","@azure/ms-rest-js":"^2.6.0","@azure/storage-blob":"^12.13.0","@protobuf-ts/plugin":"^2.9.4","semver":"^6.3.1","jwt-decode":"^3.1.2","twirp-ts":"^2.5.0"},"devDependencies":{"@types/semver":"^6.0.0","typescript":"^5.2.2"}}'); module.exports = JSON.parse('{"name":"@actions/cache","version":"3.3.0","preview":true,"description":"Actions cache lib","keywords":["github","actions","cache"],"homepage":"https://github.com/actions/toolkit/tree/main/packages/cache","license":"MIT","main":"lib/cache.js","types":"lib/cache.d.ts","directories":{"lib":"lib","test":"__tests__"},"files":["lib","!.DS_Store"],"publishConfig":{"access":"public"},"repository":{"type":"git","url":"git+https://github.com/actions/toolkit.git","directory":"packages/cache"},"scripts":{"audit-moderate":"npm install && npm audit --json --audit-level=moderate > audit.json","test":"echo \\"Error: run tests from root\\" && exit 1","tsc":"tsc"},"bugs":{"url":"https://github.com/actions/toolkit/issues"},"dependencies":{"@actions/core":"^1.11.1","@actions/exec":"^1.0.1","@actions/glob":"^0.1.0","@actions/http-client":"^2.1.1","@actions/io":"^1.0.1","@azure/abort-controller":"^1.1.0","@azure/ms-rest-js":"^2.6.0","@azure/storage-blob":"^12.13.0","@protobuf-ts/plugin":"^2.9.4","semver":"^6.3.1","twirp-ts":"^2.5.0"},"devDependencies":{"@types/semver":"^6.0.0","typescript":"^5.2.2"}}');
/***/ }) /***/ })

468
dist/restore/index.js vendored
View File

@ -6064,11 +6064,8 @@ function restoreCacheV2(paths, primaryKey, restoreKeys, options, enableCrossOsAr
let archivePath = ''; let archivePath = '';
try { try {
const twirpClient = cacheTwirpClient.internalCacheTwirpClient(); const twirpClient = cacheTwirpClient.internalCacheTwirpClient();
const backendIds = utils.getBackendIdsFromToken();
const compressionMethod = yield utils.getCompressionMethod(); const compressionMethod = yield utils.getCompressionMethod();
const request = { const request = {
workflowRunBackendId: backendIds.workflowRunBackendId,
workflowJobRunBackendId: backendIds.workflowJobRunBackendId,
key: primaryKey, key: primaryKey,
restoreKeys, restoreKeys,
version: utils.getCacheVersion(paths, compressionMethod, enableCrossOsArchive) version: utils.getCacheVersion(paths, compressionMethod, enableCrossOsArchive)
@ -6223,8 +6220,6 @@ function saveCacheV1(paths, key, options, enableCrossOsArchive = false) {
*/ */
function saveCacheV2(paths, key, options, enableCrossOsArchive = false) { function saveCacheV2(paths, key, options, enableCrossOsArchive = false) {
return __awaiter(this, void 0, void 0, function* () { return __awaiter(this, void 0, void 0, function* () {
// BackendIds are retrieved form the signed JWT
const backendIds = utils.getBackendIdsFromToken();
const compressionMethod = yield utils.getCompressionMethod(); const compressionMethod = yield utils.getCompressionMethod();
const twirpClient = cacheTwirpClient.internalCacheTwirpClient(); const twirpClient = cacheTwirpClient.internalCacheTwirpClient();
let cacheId = -1; let cacheId = -1;
@ -6251,8 +6246,6 @@ function saveCacheV2(paths, key, options, enableCrossOsArchive = false) {
core.debug('Reserving Cache'); core.debug('Reserving Cache');
const version = utils.getCacheVersion(paths, compressionMethod, enableCrossOsArchive); const version = utils.getCacheVersion(paths, compressionMethod, enableCrossOsArchive);
const request = { const request = {
workflowRunBackendId: backendIds.workflowRunBackendId,
workflowJobRunBackendId: backendIds.workflowJobRunBackendId,
key, key,
version version
}; };
@ -6263,8 +6256,6 @@ function saveCacheV2(paths, key, options, enableCrossOsArchive = false) {
core.debug(`Attempting to upload cache located at: ${archivePath}`); core.debug(`Attempting to upload cache located at: ${archivePath}`);
yield (0, upload_cache_1.UploadCacheFile)(response.signedUploadUrl, archivePath); yield (0, upload_cache_1.UploadCacheFile)(response.signedUploadUrl, archivePath);
const finalizeRequest = { const finalizeRequest = {
workflowRunBackendId: backendIds.workflowRunBackendId,
workflowJobRunBackendId: backendIds.workflowJobRunBackendId,
key, key,
version, version,
sizeBytes: `${archiveFileSize}` sizeBytes: `${archiveFileSize}`
@ -6456,18 +6447,18 @@ const runtime_3 = __nccwpck_require__(3503);
const runtime_4 = __nccwpck_require__(3503); const runtime_4 = __nccwpck_require__(3503);
const runtime_5 = __nccwpck_require__(3503); const runtime_5 = __nccwpck_require__(3503);
const timestamp_1 = __nccwpck_require__(8983); const timestamp_1 = __nccwpck_require__(8983);
const cachemetadata_1 = __nccwpck_require__(2773);
// @generated message type with reflection information, may provide speed optimized methods // @generated message type with reflection information, may provide speed optimized methods
class CreateCacheEntryRequest$Type extends runtime_5.MessageType { class CreateCacheEntryRequest$Type extends runtime_5.MessageType {
constructor() { constructor() {
super("github.actions.results.api.v1.CreateCacheEntryRequest", [ super("github.actions.results.api.v1.CreateCacheEntryRequest", [
{ no: 1, name: "workflow_run_backend_id", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 1, name: "metadata", kind: "message", T: () => cachemetadata_1.CacheMetadata },
{ no: 2, name: "workflow_job_run_backend_id", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 2, name: "key", kind: "scalar", T: 9 /*ScalarType.STRING*/ },
{ no: 3, name: "key", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 3, name: "version", kind: "scalar", T: 9 /*ScalarType.STRING*/ }
{ no: 4, name: "version", kind: "scalar", T: 9 /*ScalarType.STRING*/ }
]); ]);
} }
create(value) { create(value) {
const message = { workflowRunBackendId: "", workflowJobRunBackendId: "", key: "", version: "" }; const message = { key: "", version: "" };
globalThis.Object.defineProperty(message, runtime_4.MESSAGE_TYPE, { enumerable: false, value: this }); globalThis.Object.defineProperty(message, runtime_4.MESSAGE_TYPE, { enumerable: false, value: this });
if (value !== undefined) if (value !== undefined)
(0, runtime_3.reflectionMergePartial)(this, message, value); (0, runtime_3.reflectionMergePartial)(this, message, value);
@ -6478,16 +6469,13 @@ class CreateCacheEntryRequest$Type extends runtime_5.MessageType {
while (reader.pos < end) { while (reader.pos < end) {
let [fieldNo, wireType] = reader.tag(); let [fieldNo, wireType] = reader.tag();
switch (fieldNo) { switch (fieldNo) {
case /* string workflow_run_backend_id */ 1: case /* github.actions.results.entities.v1.CacheMetadata metadata */ 1:
message.workflowRunBackendId = reader.string(); message.metadata = cachemetadata_1.CacheMetadata.internalBinaryRead(reader, reader.uint32(), options, message.metadata);
break; break;
case /* string workflow_job_run_backend_id */ 2: case /* string key */ 2:
message.workflowJobRunBackendId = reader.string();
break;
case /* string key */ 3:
message.key = reader.string(); message.key = reader.string();
break; break;
case /* string version */ 4: case /* string version */ 3:
message.version = reader.string(); message.version = reader.string();
break; break;
default: default:
@ -6502,18 +6490,15 @@ class CreateCacheEntryRequest$Type extends runtime_5.MessageType {
return message; return message;
} }
internalBinaryWrite(message, writer, options) { internalBinaryWrite(message, writer, options) {
/* string workflow_run_backend_id = 1; */ /* github.actions.results.entities.v1.CacheMetadata metadata = 1; */
if (message.workflowRunBackendId !== "") if (message.metadata)
writer.tag(1, runtime_1.WireType.LengthDelimited).string(message.workflowRunBackendId); cachemetadata_1.CacheMetadata.internalBinaryWrite(message.metadata, writer.tag(1, runtime_1.WireType.LengthDelimited).fork(), options).join();
/* string workflow_job_run_backend_id = 2; */ /* string key = 2; */
if (message.workflowJobRunBackendId !== "")
writer.tag(2, runtime_1.WireType.LengthDelimited).string(message.workflowJobRunBackendId);
/* string key = 3; */
if (message.key !== "") if (message.key !== "")
writer.tag(3, runtime_1.WireType.LengthDelimited).string(message.key); writer.tag(2, runtime_1.WireType.LengthDelimited).string(message.key);
/* string version = 4; */ /* string version = 3; */
if (message.version !== "") if (message.version !== "")
writer.tag(4, runtime_1.WireType.LengthDelimited).string(message.version); writer.tag(3, runtime_1.WireType.LengthDelimited).string(message.version);
let u = options.writeUnknownFields; let u = options.writeUnknownFields;
if (u !== false) if (u !== false)
(u == true ? runtime_2.UnknownFieldHandler.onWrite : u)(this.typeName, message, writer); (u == true ? runtime_2.UnknownFieldHandler.onWrite : u)(this.typeName, message, writer);
@ -6582,15 +6567,14 @@ exports.CreateCacheEntryResponse = new CreateCacheEntryResponse$Type();
class FinalizeCacheEntryUploadRequest$Type extends runtime_5.MessageType { class FinalizeCacheEntryUploadRequest$Type extends runtime_5.MessageType {
constructor() { constructor() {
super("github.actions.results.api.v1.FinalizeCacheEntryUploadRequest", [ super("github.actions.results.api.v1.FinalizeCacheEntryUploadRequest", [
{ no: 1, name: "workflow_run_backend_id", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 1, name: "metadata", kind: "message", T: () => cachemetadata_1.CacheMetadata },
{ no: 2, name: "workflow_job_run_backend_id", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 2, name: "key", kind: "scalar", T: 9 /*ScalarType.STRING*/ },
{ no: 3, name: "key", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 3, name: "size_bytes", kind: "scalar", T: 3 /*ScalarType.INT64*/ },
{ no: 4, name: "size_bytes", kind: "scalar", T: 3 /*ScalarType.INT64*/ }, { no: 4, name: "version", kind: "scalar", T: 9 /*ScalarType.STRING*/ }
{ no: 5, name: "version", kind: "scalar", T: 9 /*ScalarType.STRING*/ }
]); ]);
} }
create(value) { create(value) {
const message = { workflowRunBackendId: "", workflowJobRunBackendId: "", key: "", sizeBytes: "0", version: "" }; const message = { key: "", sizeBytes: "0", version: "" };
globalThis.Object.defineProperty(message, runtime_4.MESSAGE_TYPE, { enumerable: false, value: this }); globalThis.Object.defineProperty(message, runtime_4.MESSAGE_TYPE, { enumerable: false, value: this });
if (value !== undefined) if (value !== undefined)
(0, runtime_3.reflectionMergePartial)(this, message, value); (0, runtime_3.reflectionMergePartial)(this, message, value);
@ -6601,19 +6585,16 @@ class FinalizeCacheEntryUploadRequest$Type extends runtime_5.MessageType {
while (reader.pos < end) { while (reader.pos < end) {
let [fieldNo, wireType] = reader.tag(); let [fieldNo, wireType] = reader.tag();
switch (fieldNo) { switch (fieldNo) {
case /* string workflow_run_backend_id */ 1: case /* github.actions.results.entities.v1.CacheMetadata metadata */ 1:
message.workflowRunBackendId = reader.string(); message.metadata = cachemetadata_1.CacheMetadata.internalBinaryRead(reader, reader.uint32(), options, message.metadata);
break; break;
case /* string workflow_job_run_backend_id */ 2: case /* string key */ 2:
message.workflowJobRunBackendId = reader.string();
break;
case /* string key */ 3:
message.key = reader.string(); message.key = reader.string();
break; break;
case /* int64 size_bytes */ 4: case /* int64 size_bytes */ 3:
message.sizeBytes = reader.int64().toString(); message.sizeBytes = reader.int64().toString();
break; break;
case /* string version */ 5: case /* string version */ 4:
message.version = reader.string(); message.version = reader.string();
break; break;
default: default:
@ -6628,21 +6609,18 @@ class FinalizeCacheEntryUploadRequest$Type extends runtime_5.MessageType {
return message; return message;
} }
internalBinaryWrite(message, writer, options) { internalBinaryWrite(message, writer, options) {
/* string workflow_run_backend_id = 1; */ /* github.actions.results.entities.v1.CacheMetadata metadata = 1; */
if (message.workflowRunBackendId !== "") if (message.metadata)
writer.tag(1, runtime_1.WireType.LengthDelimited).string(message.workflowRunBackendId); cachemetadata_1.CacheMetadata.internalBinaryWrite(message.metadata, writer.tag(1, runtime_1.WireType.LengthDelimited).fork(), options).join();
/* string workflow_job_run_backend_id = 2; */ /* string key = 2; */
if (message.workflowJobRunBackendId !== "")
writer.tag(2, runtime_1.WireType.LengthDelimited).string(message.workflowJobRunBackendId);
/* string key = 3; */
if (message.key !== "") if (message.key !== "")
writer.tag(3, runtime_1.WireType.LengthDelimited).string(message.key); writer.tag(2, runtime_1.WireType.LengthDelimited).string(message.key);
/* int64 size_bytes = 4; */ /* int64 size_bytes = 3; */
if (message.sizeBytes !== "0") if (message.sizeBytes !== "0")
writer.tag(4, runtime_1.WireType.Varint).int64(message.sizeBytes); writer.tag(3, runtime_1.WireType.Varint).int64(message.sizeBytes);
/* string version = 5; */ /* string version = 4; */
if (message.version !== "") if (message.version !== "")
writer.tag(5, runtime_1.WireType.LengthDelimited).string(message.version); writer.tag(4, runtime_1.WireType.LengthDelimited).string(message.version);
let u = options.writeUnknownFields; let u = options.writeUnknownFields;
if (u !== false) if (u !== false)
(u == true ? runtime_2.UnknownFieldHandler.onWrite : u)(this.typeName, message, writer); (u == true ? runtime_2.UnknownFieldHandler.onWrite : u)(this.typeName, message, writer);
@ -6711,15 +6689,14 @@ exports.FinalizeCacheEntryUploadResponse = new FinalizeCacheEntryUploadResponse$
class GetCacheEntryDownloadURLRequest$Type extends runtime_5.MessageType { class GetCacheEntryDownloadURLRequest$Type extends runtime_5.MessageType {
constructor() { constructor() {
super("github.actions.results.api.v1.GetCacheEntryDownloadURLRequest", [ super("github.actions.results.api.v1.GetCacheEntryDownloadURLRequest", [
{ no: 1, name: "workflow_run_backend_id", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 1, name: "metadata", kind: "message", T: () => cachemetadata_1.CacheMetadata },
{ no: 2, name: "workflow_job_run_backend_id", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 2, name: "key", kind: "scalar", T: 9 /*ScalarType.STRING*/ },
{ no: 3, name: "key", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 3, name: "restore_keys", kind: "scalar", repeat: 2 /*RepeatType.UNPACKED*/, T: 9 /*ScalarType.STRING*/ },
{ no: 4, name: "restore_keys", kind: "scalar", repeat: 2 /*RepeatType.UNPACKED*/, T: 9 /*ScalarType.STRING*/ }, { no: 4, name: "version", kind: "scalar", T: 9 /*ScalarType.STRING*/ }
{ no: 5, name: "version", kind: "scalar", T: 9 /*ScalarType.STRING*/ }
]); ]);
} }
create(value) { create(value) {
const message = { workflowRunBackendId: "", workflowJobRunBackendId: "", key: "", restoreKeys: [], version: "" }; const message = { key: "", restoreKeys: [], version: "" };
globalThis.Object.defineProperty(message, runtime_4.MESSAGE_TYPE, { enumerable: false, value: this }); globalThis.Object.defineProperty(message, runtime_4.MESSAGE_TYPE, { enumerable: false, value: this });
if (value !== undefined) if (value !== undefined)
(0, runtime_3.reflectionMergePartial)(this, message, value); (0, runtime_3.reflectionMergePartial)(this, message, value);
@ -6730,19 +6707,16 @@ class GetCacheEntryDownloadURLRequest$Type extends runtime_5.MessageType {
while (reader.pos < end) { while (reader.pos < end) {
let [fieldNo, wireType] = reader.tag(); let [fieldNo, wireType] = reader.tag();
switch (fieldNo) { switch (fieldNo) {
case /* string workflow_run_backend_id */ 1: case /* github.actions.results.entities.v1.CacheMetadata metadata */ 1:
message.workflowRunBackendId = reader.string(); message.metadata = cachemetadata_1.CacheMetadata.internalBinaryRead(reader, reader.uint32(), options, message.metadata);
break; break;
case /* string workflow_job_run_backend_id */ 2: case /* string key */ 2:
message.workflowJobRunBackendId = reader.string();
break;
case /* string key */ 3:
message.key = reader.string(); message.key = reader.string();
break; break;
case /* repeated string restore_keys */ 4: case /* repeated string restore_keys */ 3:
message.restoreKeys.push(reader.string()); message.restoreKeys.push(reader.string());
break; break;
case /* string version */ 5: case /* string version */ 4:
message.version = reader.string(); message.version = reader.string();
break; break;
default: default:
@ -6757,21 +6731,18 @@ class GetCacheEntryDownloadURLRequest$Type extends runtime_5.MessageType {
return message; return message;
} }
internalBinaryWrite(message, writer, options) { internalBinaryWrite(message, writer, options) {
/* string workflow_run_backend_id = 1; */ /* github.actions.results.entities.v1.CacheMetadata metadata = 1; */
if (message.workflowRunBackendId !== "") if (message.metadata)
writer.tag(1, runtime_1.WireType.LengthDelimited).string(message.workflowRunBackendId); cachemetadata_1.CacheMetadata.internalBinaryWrite(message.metadata, writer.tag(1, runtime_1.WireType.LengthDelimited).fork(), options).join();
/* string workflow_job_run_backend_id = 2; */ /* string key = 2; */
if (message.workflowJobRunBackendId !== "")
writer.tag(2, runtime_1.WireType.LengthDelimited).string(message.workflowJobRunBackendId);
/* string key = 3; */
if (message.key !== "") if (message.key !== "")
writer.tag(3, runtime_1.WireType.LengthDelimited).string(message.key); writer.tag(2, runtime_1.WireType.LengthDelimited).string(message.key);
/* repeated string restore_keys = 4; */ /* repeated string restore_keys = 3; */
for (let i = 0; i < message.restoreKeys.length; i++) for (let i = 0; i < message.restoreKeys.length; i++)
writer.tag(4, runtime_1.WireType.LengthDelimited).string(message.restoreKeys[i]); writer.tag(3, runtime_1.WireType.LengthDelimited).string(message.restoreKeys[i]);
/* string version = 5; */ /* string version = 4; */
if (message.version !== "") if (message.version !== "")
writer.tag(5, runtime_1.WireType.LengthDelimited).string(message.version); writer.tag(4, runtime_1.WireType.LengthDelimited).string(message.version);
let u = options.writeUnknownFields; let u = options.writeUnknownFields;
if (u !== false) if (u !== false)
(u == true ? runtime_2.UnknownFieldHandler.onWrite : u)(this.typeName, message, writer); (u == true ? runtime_2.UnknownFieldHandler.onWrite : u)(this.typeName, message, writer);
@ -6840,13 +6811,12 @@ exports.GetCacheEntryDownloadURLResponse = new GetCacheEntryDownloadURLResponse$
class DeleteCacheEntryRequest$Type extends runtime_5.MessageType { class DeleteCacheEntryRequest$Type extends runtime_5.MessageType {
constructor() { constructor() {
super("github.actions.results.api.v1.DeleteCacheEntryRequest", [ super("github.actions.results.api.v1.DeleteCacheEntryRequest", [
{ no: 1, name: "workflow_run_backend_id", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 1, name: "metadata", kind: "message", T: () => cachemetadata_1.CacheMetadata },
{ no: 2, name: "workflow_job_run_backend_id", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 2, name: "key", kind: "scalar", T: 9 /*ScalarType.STRING*/ }
{ no: 3, name: "key", kind: "scalar", T: 9 /*ScalarType.STRING*/ }
]); ]);
} }
create(value) { create(value) {
const message = { workflowRunBackendId: "", workflowJobRunBackendId: "", key: "" }; const message = { key: "" };
globalThis.Object.defineProperty(message, runtime_4.MESSAGE_TYPE, { enumerable: false, value: this }); globalThis.Object.defineProperty(message, runtime_4.MESSAGE_TYPE, { enumerable: false, value: this });
if (value !== undefined) if (value !== undefined)
(0, runtime_3.reflectionMergePartial)(this, message, value); (0, runtime_3.reflectionMergePartial)(this, message, value);
@ -6857,13 +6827,10 @@ class DeleteCacheEntryRequest$Type extends runtime_5.MessageType {
while (reader.pos < end) { while (reader.pos < end) {
let [fieldNo, wireType] = reader.tag(); let [fieldNo, wireType] = reader.tag();
switch (fieldNo) { switch (fieldNo) {
case /* string workflow_run_backend_id */ 1: case /* github.actions.results.entities.v1.CacheMetadata metadata */ 1:
message.workflowRunBackendId = reader.string(); message.metadata = cachemetadata_1.CacheMetadata.internalBinaryRead(reader, reader.uint32(), options, message.metadata);
break; break;
case /* string workflow_job_run_backend_id */ 2: case /* string key */ 2:
message.workflowJobRunBackendId = reader.string();
break;
case /* string key */ 3:
message.key = reader.string(); message.key = reader.string();
break; break;
default: default:
@ -6878,15 +6845,12 @@ class DeleteCacheEntryRequest$Type extends runtime_5.MessageType {
return message; return message;
} }
internalBinaryWrite(message, writer, options) { internalBinaryWrite(message, writer, options) {
/* string workflow_run_backend_id = 1; */ /* github.actions.results.entities.v1.CacheMetadata metadata = 1; */
if (message.workflowRunBackendId !== "") if (message.metadata)
writer.tag(1, runtime_1.WireType.LengthDelimited).string(message.workflowRunBackendId); cachemetadata_1.CacheMetadata.internalBinaryWrite(message.metadata, writer.tag(1, runtime_1.WireType.LengthDelimited).fork(), options).join();
/* string workflow_job_run_backend_id = 2; */ /* string key = 2; */
if (message.workflowJobRunBackendId !== "")
writer.tag(2, runtime_1.WireType.LengthDelimited).string(message.workflowJobRunBackendId);
/* string key = 3; */
if (message.key !== "") if (message.key !== "")
writer.tag(3, runtime_1.WireType.LengthDelimited).string(message.key); writer.tag(2, runtime_1.WireType.LengthDelimited).string(message.key);
let u = options.writeUnknownFields; let u = options.writeUnknownFields;
if (u !== false) if (u !== false)
(u == true ? runtime_2.UnknownFieldHandler.onWrite : u)(this.typeName, message, writer); (u == true ? runtime_2.UnknownFieldHandler.onWrite : u)(this.typeName, message, writer);
@ -6955,14 +6919,13 @@ exports.DeleteCacheEntryResponse = new DeleteCacheEntryResponse$Type();
class ListCacheEntriesRequest$Type extends runtime_5.MessageType { class ListCacheEntriesRequest$Type extends runtime_5.MessageType {
constructor() { constructor() {
super("github.actions.results.api.v1.ListCacheEntriesRequest", [ super("github.actions.results.api.v1.ListCacheEntriesRequest", [
{ no: 1, name: "workflow_run_backend_id", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 1, name: "metadata", kind: "message", T: () => cachemetadata_1.CacheMetadata },
{ no: 2, name: "workflow_job_run_backend_id", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 2, name: "key", kind: "scalar", T: 9 /*ScalarType.STRING*/ },
{ no: 3, name: "key", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 3, name: "restore_keys", kind: "scalar", repeat: 2 /*RepeatType.UNPACKED*/, T: 9 /*ScalarType.STRING*/ }
{ no: 4, name: "restore_keys", kind: "scalar", repeat: 2 /*RepeatType.UNPACKED*/, T: 9 /*ScalarType.STRING*/ }
]); ]);
} }
create(value) { create(value) {
const message = { workflowRunBackendId: "", workflowJobRunBackendId: "", key: "", restoreKeys: [] }; const message = { key: "", restoreKeys: [] };
globalThis.Object.defineProperty(message, runtime_4.MESSAGE_TYPE, { enumerable: false, value: this }); globalThis.Object.defineProperty(message, runtime_4.MESSAGE_TYPE, { enumerable: false, value: this });
if (value !== undefined) if (value !== undefined)
(0, runtime_3.reflectionMergePartial)(this, message, value); (0, runtime_3.reflectionMergePartial)(this, message, value);
@ -6973,16 +6936,13 @@ class ListCacheEntriesRequest$Type extends runtime_5.MessageType {
while (reader.pos < end) { while (reader.pos < end) {
let [fieldNo, wireType] = reader.tag(); let [fieldNo, wireType] = reader.tag();
switch (fieldNo) { switch (fieldNo) {
case /* string workflow_run_backend_id */ 1: case /* github.actions.results.entities.v1.CacheMetadata metadata */ 1:
message.workflowRunBackendId = reader.string(); message.metadata = cachemetadata_1.CacheMetadata.internalBinaryRead(reader, reader.uint32(), options, message.metadata);
break; break;
case /* string workflow_job_run_backend_id */ 2: case /* string key */ 2:
message.workflowJobRunBackendId = reader.string();
break;
case /* string key */ 3:
message.key = reader.string(); message.key = reader.string();
break; break;
case /* repeated string restore_keys */ 4: case /* repeated string restore_keys */ 3:
message.restoreKeys.push(reader.string()); message.restoreKeys.push(reader.string());
break; break;
default: default:
@ -6997,18 +6957,15 @@ class ListCacheEntriesRequest$Type extends runtime_5.MessageType {
return message; return message;
} }
internalBinaryWrite(message, writer, options) { internalBinaryWrite(message, writer, options) {
/* string workflow_run_backend_id = 1; */ /* github.actions.results.entities.v1.CacheMetadata metadata = 1; */
if (message.workflowRunBackendId !== "") if (message.metadata)
writer.tag(1, runtime_1.WireType.LengthDelimited).string(message.workflowRunBackendId); cachemetadata_1.CacheMetadata.internalBinaryWrite(message.metadata, writer.tag(1, runtime_1.WireType.LengthDelimited).fork(), options).join();
/* string workflow_job_run_backend_id = 2; */ /* string key = 2; */
if (message.workflowJobRunBackendId !== "")
writer.tag(2, runtime_1.WireType.LengthDelimited).string(message.workflowJobRunBackendId);
/* string key = 3; */
if (message.key !== "") if (message.key !== "")
writer.tag(3, runtime_1.WireType.LengthDelimited).string(message.key); writer.tag(2, runtime_1.WireType.LengthDelimited).string(message.key);
/* repeated string restore_keys = 4; */ /* repeated string restore_keys = 3; */
for (let i = 0; i < message.restoreKeys.length; i++) for (let i = 0; i < message.restoreKeys.length; i++)
writer.tag(4, runtime_1.WireType.LengthDelimited).string(message.restoreKeys[i]); writer.tag(3, runtime_1.WireType.LengthDelimited).string(message.restoreKeys[i]);
let u = options.writeUnknownFields; let u = options.writeUnknownFields;
if (u !== false) if (u !== false)
(u == true ? runtime_2.UnknownFieldHandler.onWrite : u)(this.typeName, message, writer); (u == true ? runtime_2.UnknownFieldHandler.onWrite : u)(this.typeName, message, writer);
@ -7166,15 +7123,14 @@ exports.ListCacheEntriesResponse_CacheEntry = new ListCacheEntriesResponse_Cache
class LookupCacheEntryRequest$Type extends runtime_5.MessageType { class LookupCacheEntryRequest$Type extends runtime_5.MessageType {
constructor() { constructor() {
super("github.actions.results.api.v1.LookupCacheEntryRequest", [ super("github.actions.results.api.v1.LookupCacheEntryRequest", [
{ no: 1, name: "workflow_run_backend_id", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 1, name: "metadata", kind: "message", T: () => cachemetadata_1.CacheMetadata },
{ no: 2, name: "workflow_job_run_backend_id", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 2, name: "key", kind: "scalar", T: 9 /*ScalarType.STRING*/ },
{ no: 3, name: "key", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 3, name: "restore_keys", kind: "scalar", repeat: 2 /*RepeatType.UNPACKED*/, T: 9 /*ScalarType.STRING*/ },
{ no: 4, name: "restore_keys", kind: "scalar", repeat: 2 /*RepeatType.UNPACKED*/, T: 9 /*ScalarType.STRING*/ }, { no: 4, name: "version", kind: "scalar", T: 9 /*ScalarType.STRING*/ }
{ no: 5, name: "version", kind: "scalar", T: 9 /*ScalarType.STRING*/ }
]); ]);
} }
create(value) { create(value) {
const message = { workflowRunBackendId: "", workflowJobRunBackendId: "", key: "", restoreKeys: [], version: "" }; const message = { key: "", restoreKeys: [], version: "" };
globalThis.Object.defineProperty(message, runtime_4.MESSAGE_TYPE, { enumerable: false, value: this }); globalThis.Object.defineProperty(message, runtime_4.MESSAGE_TYPE, { enumerable: false, value: this });
if (value !== undefined) if (value !== undefined)
(0, runtime_3.reflectionMergePartial)(this, message, value); (0, runtime_3.reflectionMergePartial)(this, message, value);
@ -7185,19 +7141,16 @@ class LookupCacheEntryRequest$Type extends runtime_5.MessageType {
while (reader.pos < end) { while (reader.pos < end) {
let [fieldNo, wireType] = reader.tag(); let [fieldNo, wireType] = reader.tag();
switch (fieldNo) { switch (fieldNo) {
case /* string workflow_run_backend_id */ 1: case /* github.actions.results.entities.v1.CacheMetadata metadata */ 1:
message.workflowRunBackendId = reader.string(); message.metadata = cachemetadata_1.CacheMetadata.internalBinaryRead(reader, reader.uint32(), options, message.metadata);
break; break;
case /* string workflow_job_run_backend_id */ 2: case /* string key */ 2:
message.workflowJobRunBackendId = reader.string();
break;
case /* string key */ 3:
message.key = reader.string(); message.key = reader.string();
break; break;
case /* repeated string restore_keys */ 4: case /* repeated string restore_keys */ 3:
message.restoreKeys.push(reader.string()); message.restoreKeys.push(reader.string());
break; break;
case /* string version */ 5: case /* string version */ 4:
message.version = reader.string(); message.version = reader.string();
break; break;
default: default:
@ -7212,21 +7165,18 @@ class LookupCacheEntryRequest$Type extends runtime_5.MessageType {
return message; return message;
} }
internalBinaryWrite(message, writer, options) { internalBinaryWrite(message, writer, options) {
/* string workflow_run_backend_id = 1; */ /* github.actions.results.entities.v1.CacheMetadata metadata = 1; */
if (message.workflowRunBackendId !== "") if (message.metadata)
writer.tag(1, runtime_1.WireType.LengthDelimited).string(message.workflowRunBackendId); cachemetadata_1.CacheMetadata.internalBinaryWrite(message.metadata, writer.tag(1, runtime_1.WireType.LengthDelimited).fork(), options).join();
/* string workflow_job_run_backend_id = 2; */ /* string key = 2; */
if (message.workflowJobRunBackendId !== "")
writer.tag(2, runtime_1.WireType.LengthDelimited).string(message.workflowJobRunBackendId);
/* string key = 3; */
if (message.key !== "") if (message.key !== "")
writer.tag(3, runtime_1.WireType.LengthDelimited).string(message.key); writer.tag(2, runtime_1.WireType.LengthDelimited).string(message.key);
/* repeated string restore_keys = 4; */ /* repeated string restore_keys = 3; */
for (let i = 0; i < message.restoreKeys.length; i++) for (let i = 0; i < message.restoreKeys.length; i++)
writer.tag(4, runtime_1.WireType.LengthDelimited).string(message.restoreKeys[i]); writer.tag(3, runtime_1.WireType.LengthDelimited).string(message.restoreKeys[i]);
/* string version = 5; */ /* string version = 4; */
if (message.version !== "") if (message.version !== "")
writer.tag(5, runtime_1.WireType.LengthDelimited).string(message.version); writer.tag(4, runtime_1.WireType.LengthDelimited).string(message.version);
let u = options.writeUnknownFields; let u = options.writeUnknownFields;
if (u !== false) if (u !== false)
(u == true ? runtime_2.UnknownFieldHandler.onWrite : u)(this.typeName, message, writer); (u == true ? runtime_2.UnknownFieldHandler.onWrite : u)(this.typeName, message, writer);
@ -8004,6 +7954,147 @@ function handleCacheServiceLookupCacheEntryProtobuf(ctx, service, data, intercep
/***/ }), /***/ }),
/***/ 2773:
/***/ ((__unused_webpack_module, exports, __nccwpck_require__) => {
"use strict";
Object.defineProperty(exports, "__esModule", ({ value: true }));
exports.CacheMetadata = void 0;
const runtime_1 = __nccwpck_require__(3503);
const runtime_2 = __nccwpck_require__(3503);
const runtime_3 = __nccwpck_require__(3503);
const runtime_4 = __nccwpck_require__(3503);
const runtime_5 = __nccwpck_require__(3503);
const cachescope_1 = __nccwpck_require__(7685);
// @generated message type with reflection information, may provide speed optimized methods
class CacheMetadata$Type extends runtime_5.MessageType {
constructor() {
super("github.actions.results.entities.v1.CacheMetadata", [
{ no: 1, name: "repository_id", kind: "scalar", T: 3 /*ScalarType.INT64*/ },
{ no: 2, name: "scope", kind: "message", repeat: 1 /*RepeatType.PACKED*/, T: () => cachescope_1.CacheScope }
]);
}
create(value) {
const message = { repositoryId: "0", scope: [] };
globalThis.Object.defineProperty(message, runtime_4.MESSAGE_TYPE, { enumerable: false, value: this });
if (value !== undefined)
(0, runtime_3.reflectionMergePartial)(this, message, value);
return message;
}
internalBinaryRead(reader, length, options, target) {
let message = target !== null && target !== void 0 ? target : this.create(), end = reader.pos + length;
while (reader.pos < end) {
let [fieldNo, wireType] = reader.tag();
switch (fieldNo) {
case /* int64 repository_id */ 1:
message.repositoryId = reader.int64().toString();
break;
case /* repeated github.actions.results.entities.v1.CacheScope scope */ 2:
message.scope.push(cachescope_1.CacheScope.internalBinaryRead(reader, reader.uint32(), options));
break;
default:
let u = options.readUnknownField;
if (u === "throw")
throw new globalThis.Error(`Unknown field ${fieldNo} (wire type ${wireType}) for ${this.typeName}`);
let d = reader.skip(wireType);
if (u !== false)
(u === true ? runtime_2.UnknownFieldHandler.onRead : u)(this.typeName, message, fieldNo, wireType, d);
}
}
return message;
}
internalBinaryWrite(message, writer, options) {
/* int64 repository_id = 1; */
if (message.repositoryId !== "0")
writer.tag(1, runtime_1.WireType.Varint).int64(message.repositoryId);
/* repeated github.actions.results.entities.v1.CacheScope scope = 2; */
for (let i = 0; i < message.scope.length; i++)
cachescope_1.CacheScope.internalBinaryWrite(message.scope[i], writer.tag(2, runtime_1.WireType.LengthDelimited).fork(), options).join();
let u = options.writeUnknownFields;
if (u !== false)
(u == true ? runtime_2.UnknownFieldHandler.onWrite : u)(this.typeName, message, writer);
return writer;
}
}
/**
* @generated MessageType for protobuf message github.actions.results.entities.v1.CacheMetadata
*/
exports.CacheMetadata = new CacheMetadata$Type();
//# sourceMappingURL=cachemetadata.js.map
/***/ }),
/***/ 7685:
/***/ ((__unused_webpack_module, exports, __nccwpck_require__) => {
"use strict";
Object.defineProperty(exports, "__esModule", ({ value: true }));
exports.CacheScope = void 0;
const runtime_1 = __nccwpck_require__(3503);
const runtime_2 = __nccwpck_require__(3503);
const runtime_3 = __nccwpck_require__(3503);
const runtime_4 = __nccwpck_require__(3503);
const runtime_5 = __nccwpck_require__(3503);
// @generated message type with reflection information, may provide speed optimized methods
class CacheScope$Type extends runtime_5.MessageType {
constructor() {
super("github.actions.results.entities.v1.CacheScope", [
{ no: 1, name: "scope", kind: "scalar", T: 9 /*ScalarType.STRING*/ },
{ no: 2, name: "permission", kind: "scalar", T: 3 /*ScalarType.INT64*/ }
]);
}
create(value) {
const message = { scope: "", permission: "0" };
globalThis.Object.defineProperty(message, runtime_4.MESSAGE_TYPE, { enumerable: false, value: this });
if (value !== undefined)
(0, runtime_3.reflectionMergePartial)(this, message, value);
return message;
}
internalBinaryRead(reader, length, options, target) {
let message = target !== null && target !== void 0 ? target : this.create(), end = reader.pos + length;
while (reader.pos < end) {
let [fieldNo, wireType] = reader.tag();
switch (fieldNo) {
case /* string scope */ 1:
message.scope = reader.string();
break;
case /* int64 permission */ 2:
message.permission = reader.int64().toString();
break;
default:
let u = options.readUnknownField;
if (u === "throw")
throw new globalThis.Error(`Unknown field ${fieldNo} (wire type ${wireType}) for ${this.typeName}`);
let d = reader.skip(wireType);
if (u !== false)
(u === true ? runtime_2.UnknownFieldHandler.onRead : u)(this.typeName, message, fieldNo, wireType, d);
}
}
return message;
}
internalBinaryWrite(message, writer, options) {
/* string scope = 1; */
if (message.scope !== "")
writer.tag(1, runtime_1.WireType.LengthDelimited).string(message.scope);
/* int64 permission = 2; */
if (message.permission !== "0")
writer.tag(2, runtime_1.WireType.Varint).int64(message.permission);
let u = options.writeUnknownFields;
if (u !== false)
(u == true ? runtime_2.UnknownFieldHandler.onWrite : u)(this.typeName, message, writer);
return writer;
}
}
/**
* @generated MessageType for protobuf message github.actions.results.entities.v1.CacheScope
*/
exports.CacheScope = new CacheScope$Type();
//# sourceMappingURL=cachescope.js.map
/***/ }),
/***/ 4076: /***/ 4076:
/***/ (function(__unused_webpack_module, exports, __nccwpck_require__) { /***/ (function(__unused_webpack_module, exports, __nccwpck_require__) {
@ -8416,11 +8507,8 @@ var __asyncValues = (this && this.__asyncValues) || function (o) {
function verb(n) { i[n] = o[n] && function (v) { return new Promise(function (resolve, reject) { v = o[n](v), settle(resolve, reject, v.done, v.value); }); }; } function verb(n) { i[n] = o[n] && function (v) { return new Promise(function (resolve, reject) { v = o[n](v), settle(resolve, reject, v.done, v.value); }); }; }
function settle(resolve, reject, d, v) { Promise.resolve(v).then(function(v) { resolve({ value: v, done: d }); }, reject); } function settle(resolve, reject, d, v) { Promise.resolve(v).then(function(v) { resolve({ value: v, done: d }); }, reject); }
}; };
var __importDefault = (this && this.__importDefault) || function (mod) {
return (mod && mod.__esModule) ? mod : { "default": mod };
};
Object.defineProperty(exports, "__esModule", ({ value: true })); Object.defineProperty(exports, "__esModule", ({ value: true }));
exports.getBackendIdsFromToken = exports.getRuntimeToken = exports.getCacheVersion = exports.isGhes = exports.assertDefined = exports.getGnuTarPathOnWindows = exports.getCacheFileName = exports.getCompressionMethod = exports.unlinkFile = exports.resolvePaths = exports.getArchiveFileSizeInBytes = exports.createTempDirectory = void 0; exports.getRuntimeToken = exports.getCacheVersion = exports.isGhes = exports.assertDefined = exports.getGnuTarPathOnWindows = exports.getCacheFileName = exports.getCompressionMethod = exports.unlinkFile = exports.resolvePaths = exports.getArchiveFileSizeInBytes = exports.createTempDirectory = void 0;
const core = __importStar(__nccwpck_require__(4850)); const core = __importStar(__nccwpck_require__(4850));
const exec = __importStar(__nccwpck_require__(309)); const exec = __importStar(__nccwpck_require__(309));
const glob = __importStar(__nccwpck_require__(9590)); const glob = __importStar(__nccwpck_require__(9590));
@ -8430,7 +8518,6 @@ const fs = __importStar(__nccwpck_require__(7147));
const path = __importStar(__nccwpck_require__(1017)); const path = __importStar(__nccwpck_require__(1017));
const semver = __importStar(__nccwpck_require__(3910)); const semver = __importStar(__nccwpck_require__(3910));
const util = __importStar(__nccwpck_require__(3837)); const util = __importStar(__nccwpck_require__(3837));
const jwt_decode_1 = __importDefault(__nccwpck_require__(1300));
const constants_1 = __nccwpck_require__(4010); const constants_1 = __nccwpck_require__(4010);
const versionSalt = '1.0'; const versionSalt = '1.0';
// From https://github.com/actions/toolkit/blob/main/packages/tool-cache/src/tool-cache.ts#L23 // From https://github.com/actions/toolkit/blob/main/packages/tool-cache/src/tool-cache.ts#L23
@ -8603,54 +8690,6 @@ function getRuntimeToken() {
return token; return token;
} }
exports.getRuntimeToken = getRuntimeToken; exports.getRuntimeToken = getRuntimeToken;
const InvalidJwtError = new Error('Failed to get backend IDs: The provided JWT token is invalid and/or missing claims');
// uses the JWT token claims to get the
// workflow run and workflow job run backend ids
function getBackendIdsFromToken() {
const token = getRuntimeToken();
const decoded = (0, jwt_decode_1.default)(token);
if (!decoded.scp) {
throw InvalidJwtError;
}
/*
* example decoded:
* {
* scp: "Actions.ExampleScope Actions.Results:ce7f54c7-61c7-4aae-887f-30da475f5f1a:ca395085-040a-526b-2ce8-bdc85f692774"
* }
*/
const scpParts = decoded.scp.split(' ');
if (scpParts.length === 0) {
throw InvalidJwtError;
}
/*
* example scpParts:
* ["Actions.ExampleScope", "Actions.Results:ce7f54c7-61c7-4aae-887f-30da475f5f1a:ca395085-040a-526b-2ce8-bdc85f692774"]
*/
for (const scopes of scpParts) {
const scopeParts = scopes.split(':');
if ((scopeParts === null || scopeParts === void 0 ? void 0 : scopeParts[0]) !== 'Actions.Results') {
// not the Actions.Results scope
continue;
}
/*
* example scopeParts:
* ["Actions.Results", "ce7f54c7-61c7-4aae-887f-30da475f5f1a", "ca395085-040a-526b-2ce8-bdc85f692774"]
*/
if (scopeParts.length !== 3) {
// missing expected number of claims
throw InvalidJwtError;
}
const ids = {
workflowRunBackendId: scopeParts[1],
workflowJobRunBackendId: scopeParts[2]
};
core.debug(`Workflow Run Backend ID: ${ids.workflowRunBackendId}`);
core.debug(`Workflow Job Run Backend ID: ${ids.workflowJobRunBackendId}`);
return ids;
}
throw InvalidJwtError;
}
exports.getBackendIdsFromToken = getBackendIdsFromToken;
//# sourceMappingURL=cacheUtils.js.map //# sourceMappingURL=cacheUtils.js.map
/***/ }), /***/ }),
@ -8661,15 +8700,7 @@ exports.getBackendIdsFromToken = getBackendIdsFromToken;
"use strict"; "use strict";
Object.defineProperty(exports, "__esModule", ({ value: true })); Object.defineProperty(exports, "__esModule", ({ value: true }));
exports.getCacheServiceURL = exports.getCacheServiceVersion = exports.getRuntimeToken = void 0; exports.getCacheServiceURL = exports.getCacheServiceVersion = void 0;
function getRuntimeToken() {
const token = process.env['ACTIONS_RUNTIME_TOKEN'];
if (!token) {
throw new Error('Unable to get the ACTIONS_RUNTIME_TOKEN env variable');
}
return token;
}
exports.getRuntimeToken = getRuntimeToken;
function getCacheServiceVersion() { function getCacheServiceVersion() {
return process.env['ACTIONS_CACHE_SERVICE_V2'] ? 'v2' : 'v1'; return process.env['ACTIONS_CACHE_SERVICE_V2'] ? 'v2' : 'v1';
} }
@ -9285,6 +9316,7 @@ const core_1 = __nccwpck_require__(4850);
const user_agent_1 = __nccwpck_require__(5736); const user_agent_1 = __nccwpck_require__(5736);
const errors_1 = __nccwpck_require__(6333); const errors_1 = __nccwpck_require__(6333);
const config_1 = __nccwpck_require__(6490); const config_1 = __nccwpck_require__(6490);
const cacheUtils_1 = __nccwpck_require__(3310);
const auth_1 = __nccwpck_require__(7231); const auth_1 = __nccwpck_require__(7231);
const http_client_1 = __nccwpck_require__(1283); const http_client_1 = __nccwpck_require__(1283);
const cache_twirp_1 = __nccwpck_require__(894); const cache_twirp_1 = __nccwpck_require__(894);
@ -9300,7 +9332,7 @@ class CacheServiceClient {
this.maxAttempts = 5; this.maxAttempts = 5;
this.baseRetryIntervalMilliseconds = 3000; this.baseRetryIntervalMilliseconds = 3000;
this.retryMultiplier = 1.5; this.retryMultiplier = 1.5;
const token = (0, config_1.getRuntimeToken)(); const token = (0, cacheUtils_1.getRuntimeToken)();
this.baseUrl = (0, config_1.getCacheServiceURL)(); this.baseUrl = (0, config_1.getCacheServiceURL)();
if (maxAttempts) { if (maxAttempts) {
this.maxAttempts = maxAttempts; this.maxAttempts = maxAttempts;
@ -57677,16 +57709,6 @@ DotObject._process = _process
module.exports = DotObject module.exports = DotObject
/***/ }),
/***/ 1300:
/***/ ((module) => {
"use strict";
function e(e){this.message=e}e.prototype=new Error,e.prototype.name="InvalidCharacterError";var r="undefined"!=typeof window&&window.atob&&window.atob.bind(window)||function(r){var t=String(r).replace(/=+$/,"");if(t.length%4==1)throw new e("'atob' failed: The string to be decoded is not correctly encoded.");for(var n,o,a=0,i=0,c="";o=t.charAt(i++);~o&&(n=a%4?64*n+o:o,a++%4)?c+=String.fromCharCode(255&n>>(-2*a&6)):0)o="ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/=".indexOf(o);return c};function t(e){var t=e.replace(/-/g,"+").replace(/_/g,"/");switch(t.length%4){case 0:break;case 2:t+="==";break;case 3:t+="=";break;default:throw"Illegal base64url string!"}try{return function(e){return decodeURIComponent(r(e).replace(/(.)/g,(function(e,r){var t=r.charCodeAt(0).toString(16).toUpperCase();return t.length<2&&(t="0"+t),"%"+t})))}(t)}catch(e){return r(t)}}function n(e){this.message=e}function o(e,r){if("string"!=typeof e)throw new n("Invalid token specified");var o=!0===(r=r||{}).header?0:1;try{return JSON.parse(t(e.split(".")[o]))}catch(e){throw new n("Invalid token specified: "+e.message)}}n.prototype=new Error,n.prototype.name="InvalidTokenError";const a=o;a.default=o,a.InvalidTokenError=n,module.exports=a;
//# sourceMappingURL=jwt-decode.cjs.js.map
/***/ }), /***/ }),
/***/ 2020: /***/ 2020:
@ -73485,7 +73507,7 @@ module.exports = JSON.parse('[[[0,44],"disallowed_STD3_valid"],[[45,46],"valid"]
/***/ ((module) => { /***/ ((module) => {
"use strict"; "use strict";
module.exports = JSON.parse('{"name":"@actions/cache","version":"3.3.0","preview":true,"description":"Actions cache lib","keywords":["github","actions","cache"],"homepage":"https://github.com/actions/toolkit/tree/main/packages/cache","license":"MIT","main":"lib/cache.js","types":"lib/cache.d.ts","directories":{"lib":"lib","test":"__tests__"},"files":["lib","!.DS_Store"],"publishConfig":{"access":"public"},"repository":{"type":"git","url":"git+https://github.com/actions/toolkit.git","directory":"packages/cache"},"scripts":{"audit-moderate":"npm install && npm audit --json --audit-level=moderate > audit.json","test":"echo \\"Error: run tests from root\\" && exit 1","tsc":"tsc"},"bugs":{"url":"https://github.com/actions/toolkit/issues"},"dependencies":{"@actions/core":"^1.11.1","@actions/exec":"^1.0.1","@actions/glob":"^0.1.0","@actions/http-client":"^2.1.1","@actions/io":"^1.0.1","@azure/abort-controller":"^1.1.0","@azure/ms-rest-js":"^2.6.0","@azure/storage-blob":"^12.13.0","@protobuf-ts/plugin":"^2.9.4","semver":"^6.3.1","jwt-decode":"^3.1.2","twirp-ts":"^2.5.0"},"devDependencies":{"@types/semver":"^6.0.0","typescript":"^5.2.2"}}'); module.exports = JSON.parse('{"name":"@actions/cache","version":"3.3.0","preview":true,"description":"Actions cache lib","keywords":["github","actions","cache"],"homepage":"https://github.com/actions/toolkit/tree/main/packages/cache","license":"MIT","main":"lib/cache.js","types":"lib/cache.d.ts","directories":{"lib":"lib","test":"__tests__"},"files":["lib","!.DS_Store"],"publishConfig":{"access":"public"},"repository":{"type":"git","url":"git+https://github.com/actions/toolkit.git","directory":"packages/cache"},"scripts":{"audit-moderate":"npm install && npm audit --json --audit-level=moderate > audit.json","test":"echo \\"Error: run tests from root\\" && exit 1","tsc":"tsc"},"bugs":{"url":"https://github.com/actions/toolkit/issues"},"dependencies":{"@actions/core":"^1.11.1","@actions/exec":"^1.0.1","@actions/glob":"^0.1.0","@actions/http-client":"^2.1.1","@actions/io":"^1.0.1","@azure/abort-controller":"^1.1.0","@azure/ms-rest-js":"^2.6.0","@azure/storage-blob":"^12.13.0","@protobuf-ts/plugin":"^2.9.4","semver":"^6.3.1","twirp-ts":"^2.5.0"},"devDependencies":{"@types/semver":"^6.0.0","typescript":"^5.2.2"}}');
/***/ }) /***/ })

View File

@ -6064,11 +6064,8 @@ function restoreCacheV2(paths, primaryKey, restoreKeys, options, enableCrossOsAr
let archivePath = ''; let archivePath = '';
try { try {
const twirpClient = cacheTwirpClient.internalCacheTwirpClient(); const twirpClient = cacheTwirpClient.internalCacheTwirpClient();
const backendIds = utils.getBackendIdsFromToken();
const compressionMethod = yield utils.getCompressionMethod(); const compressionMethod = yield utils.getCompressionMethod();
const request = { const request = {
workflowRunBackendId: backendIds.workflowRunBackendId,
workflowJobRunBackendId: backendIds.workflowJobRunBackendId,
key: primaryKey, key: primaryKey,
restoreKeys, restoreKeys,
version: utils.getCacheVersion(paths, compressionMethod, enableCrossOsArchive) version: utils.getCacheVersion(paths, compressionMethod, enableCrossOsArchive)
@ -6223,8 +6220,6 @@ function saveCacheV1(paths, key, options, enableCrossOsArchive = false) {
*/ */
function saveCacheV2(paths, key, options, enableCrossOsArchive = false) { function saveCacheV2(paths, key, options, enableCrossOsArchive = false) {
return __awaiter(this, void 0, void 0, function* () { return __awaiter(this, void 0, void 0, function* () {
// BackendIds are retrieved form the signed JWT
const backendIds = utils.getBackendIdsFromToken();
const compressionMethod = yield utils.getCompressionMethod(); const compressionMethod = yield utils.getCompressionMethod();
const twirpClient = cacheTwirpClient.internalCacheTwirpClient(); const twirpClient = cacheTwirpClient.internalCacheTwirpClient();
let cacheId = -1; let cacheId = -1;
@ -6251,8 +6246,6 @@ function saveCacheV2(paths, key, options, enableCrossOsArchive = false) {
core.debug('Reserving Cache'); core.debug('Reserving Cache');
const version = utils.getCacheVersion(paths, compressionMethod, enableCrossOsArchive); const version = utils.getCacheVersion(paths, compressionMethod, enableCrossOsArchive);
const request = { const request = {
workflowRunBackendId: backendIds.workflowRunBackendId,
workflowJobRunBackendId: backendIds.workflowJobRunBackendId,
key, key,
version version
}; };
@ -6263,8 +6256,6 @@ function saveCacheV2(paths, key, options, enableCrossOsArchive = false) {
core.debug(`Attempting to upload cache located at: ${archivePath}`); core.debug(`Attempting to upload cache located at: ${archivePath}`);
yield (0, upload_cache_1.UploadCacheFile)(response.signedUploadUrl, archivePath); yield (0, upload_cache_1.UploadCacheFile)(response.signedUploadUrl, archivePath);
const finalizeRequest = { const finalizeRequest = {
workflowRunBackendId: backendIds.workflowRunBackendId,
workflowJobRunBackendId: backendIds.workflowJobRunBackendId,
key, key,
version, version,
sizeBytes: `${archiveFileSize}` sizeBytes: `${archiveFileSize}`
@ -6456,18 +6447,18 @@ const runtime_3 = __nccwpck_require__(3503);
const runtime_4 = __nccwpck_require__(3503); const runtime_4 = __nccwpck_require__(3503);
const runtime_5 = __nccwpck_require__(3503); const runtime_5 = __nccwpck_require__(3503);
const timestamp_1 = __nccwpck_require__(8983); const timestamp_1 = __nccwpck_require__(8983);
const cachemetadata_1 = __nccwpck_require__(2773);
// @generated message type with reflection information, may provide speed optimized methods // @generated message type with reflection information, may provide speed optimized methods
class CreateCacheEntryRequest$Type extends runtime_5.MessageType { class CreateCacheEntryRequest$Type extends runtime_5.MessageType {
constructor() { constructor() {
super("github.actions.results.api.v1.CreateCacheEntryRequest", [ super("github.actions.results.api.v1.CreateCacheEntryRequest", [
{ no: 1, name: "workflow_run_backend_id", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 1, name: "metadata", kind: "message", T: () => cachemetadata_1.CacheMetadata },
{ no: 2, name: "workflow_job_run_backend_id", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 2, name: "key", kind: "scalar", T: 9 /*ScalarType.STRING*/ },
{ no: 3, name: "key", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 3, name: "version", kind: "scalar", T: 9 /*ScalarType.STRING*/ }
{ no: 4, name: "version", kind: "scalar", T: 9 /*ScalarType.STRING*/ }
]); ]);
} }
create(value) { create(value) {
const message = { workflowRunBackendId: "", workflowJobRunBackendId: "", key: "", version: "" }; const message = { key: "", version: "" };
globalThis.Object.defineProperty(message, runtime_4.MESSAGE_TYPE, { enumerable: false, value: this }); globalThis.Object.defineProperty(message, runtime_4.MESSAGE_TYPE, { enumerable: false, value: this });
if (value !== undefined) if (value !== undefined)
(0, runtime_3.reflectionMergePartial)(this, message, value); (0, runtime_3.reflectionMergePartial)(this, message, value);
@ -6478,16 +6469,13 @@ class CreateCacheEntryRequest$Type extends runtime_5.MessageType {
while (reader.pos < end) { while (reader.pos < end) {
let [fieldNo, wireType] = reader.tag(); let [fieldNo, wireType] = reader.tag();
switch (fieldNo) { switch (fieldNo) {
case /* string workflow_run_backend_id */ 1: case /* github.actions.results.entities.v1.CacheMetadata metadata */ 1:
message.workflowRunBackendId = reader.string(); message.metadata = cachemetadata_1.CacheMetadata.internalBinaryRead(reader, reader.uint32(), options, message.metadata);
break; break;
case /* string workflow_job_run_backend_id */ 2: case /* string key */ 2:
message.workflowJobRunBackendId = reader.string();
break;
case /* string key */ 3:
message.key = reader.string(); message.key = reader.string();
break; break;
case /* string version */ 4: case /* string version */ 3:
message.version = reader.string(); message.version = reader.string();
break; break;
default: default:
@ -6502,18 +6490,15 @@ class CreateCacheEntryRequest$Type extends runtime_5.MessageType {
return message; return message;
} }
internalBinaryWrite(message, writer, options) { internalBinaryWrite(message, writer, options) {
/* string workflow_run_backend_id = 1; */ /* github.actions.results.entities.v1.CacheMetadata metadata = 1; */
if (message.workflowRunBackendId !== "") if (message.metadata)
writer.tag(1, runtime_1.WireType.LengthDelimited).string(message.workflowRunBackendId); cachemetadata_1.CacheMetadata.internalBinaryWrite(message.metadata, writer.tag(1, runtime_1.WireType.LengthDelimited).fork(), options).join();
/* string workflow_job_run_backend_id = 2; */ /* string key = 2; */
if (message.workflowJobRunBackendId !== "")
writer.tag(2, runtime_1.WireType.LengthDelimited).string(message.workflowJobRunBackendId);
/* string key = 3; */
if (message.key !== "") if (message.key !== "")
writer.tag(3, runtime_1.WireType.LengthDelimited).string(message.key); writer.tag(2, runtime_1.WireType.LengthDelimited).string(message.key);
/* string version = 4; */ /* string version = 3; */
if (message.version !== "") if (message.version !== "")
writer.tag(4, runtime_1.WireType.LengthDelimited).string(message.version); writer.tag(3, runtime_1.WireType.LengthDelimited).string(message.version);
let u = options.writeUnknownFields; let u = options.writeUnknownFields;
if (u !== false) if (u !== false)
(u == true ? runtime_2.UnknownFieldHandler.onWrite : u)(this.typeName, message, writer); (u == true ? runtime_2.UnknownFieldHandler.onWrite : u)(this.typeName, message, writer);
@ -6582,15 +6567,14 @@ exports.CreateCacheEntryResponse = new CreateCacheEntryResponse$Type();
class FinalizeCacheEntryUploadRequest$Type extends runtime_5.MessageType { class FinalizeCacheEntryUploadRequest$Type extends runtime_5.MessageType {
constructor() { constructor() {
super("github.actions.results.api.v1.FinalizeCacheEntryUploadRequest", [ super("github.actions.results.api.v1.FinalizeCacheEntryUploadRequest", [
{ no: 1, name: "workflow_run_backend_id", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 1, name: "metadata", kind: "message", T: () => cachemetadata_1.CacheMetadata },
{ no: 2, name: "workflow_job_run_backend_id", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 2, name: "key", kind: "scalar", T: 9 /*ScalarType.STRING*/ },
{ no: 3, name: "key", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 3, name: "size_bytes", kind: "scalar", T: 3 /*ScalarType.INT64*/ },
{ no: 4, name: "size_bytes", kind: "scalar", T: 3 /*ScalarType.INT64*/ }, { no: 4, name: "version", kind: "scalar", T: 9 /*ScalarType.STRING*/ }
{ no: 5, name: "version", kind: "scalar", T: 9 /*ScalarType.STRING*/ }
]); ]);
} }
create(value) { create(value) {
const message = { workflowRunBackendId: "", workflowJobRunBackendId: "", key: "", sizeBytes: "0", version: "" }; const message = { key: "", sizeBytes: "0", version: "" };
globalThis.Object.defineProperty(message, runtime_4.MESSAGE_TYPE, { enumerable: false, value: this }); globalThis.Object.defineProperty(message, runtime_4.MESSAGE_TYPE, { enumerable: false, value: this });
if (value !== undefined) if (value !== undefined)
(0, runtime_3.reflectionMergePartial)(this, message, value); (0, runtime_3.reflectionMergePartial)(this, message, value);
@ -6601,19 +6585,16 @@ class FinalizeCacheEntryUploadRequest$Type extends runtime_5.MessageType {
while (reader.pos < end) { while (reader.pos < end) {
let [fieldNo, wireType] = reader.tag(); let [fieldNo, wireType] = reader.tag();
switch (fieldNo) { switch (fieldNo) {
case /* string workflow_run_backend_id */ 1: case /* github.actions.results.entities.v1.CacheMetadata metadata */ 1:
message.workflowRunBackendId = reader.string(); message.metadata = cachemetadata_1.CacheMetadata.internalBinaryRead(reader, reader.uint32(), options, message.metadata);
break; break;
case /* string workflow_job_run_backend_id */ 2: case /* string key */ 2:
message.workflowJobRunBackendId = reader.string();
break;
case /* string key */ 3:
message.key = reader.string(); message.key = reader.string();
break; break;
case /* int64 size_bytes */ 4: case /* int64 size_bytes */ 3:
message.sizeBytes = reader.int64().toString(); message.sizeBytes = reader.int64().toString();
break; break;
case /* string version */ 5: case /* string version */ 4:
message.version = reader.string(); message.version = reader.string();
break; break;
default: default:
@ -6628,21 +6609,18 @@ class FinalizeCacheEntryUploadRequest$Type extends runtime_5.MessageType {
return message; return message;
} }
internalBinaryWrite(message, writer, options) { internalBinaryWrite(message, writer, options) {
/* string workflow_run_backend_id = 1; */ /* github.actions.results.entities.v1.CacheMetadata metadata = 1; */
if (message.workflowRunBackendId !== "") if (message.metadata)
writer.tag(1, runtime_1.WireType.LengthDelimited).string(message.workflowRunBackendId); cachemetadata_1.CacheMetadata.internalBinaryWrite(message.metadata, writer.tag(1, runtime_1.WireType.LengthDelimited).fork(), options).join();
/* string workflow_job_run_backend_id = 2; */ /* string key = 2; */
if (message.workflowJobRunBackendId !== "")
writer.tag(2, runtime_1.WireType.LengthDelimited).string(message.workflowJobRunBackendId);
/* string key = 3; */
if (message.key !== "") if (message.key !== "")
writer.tag(3, runtime_1.WireType.LengthDelimited).string(message.key); writer.tag(2, runtime_1.WireType.LengthDelimited).string(message.key);
/* int64 size_bytes = 4; */ /* int64 size_bytes = 3; */
if (message.sizeBytes !== "0") if (message.sizeBytes !== "0")
writer.tag(4, runtime_1.WireType.Varint).int64(message.sizeBytes); writer.tag(3, runtime_1.WireType.Varint).int64(message.sizeBytes);
/* string version = 5; */ /* string version = 4; */
if (message.version !== "") if (message.version !== "")
writer.tag(5, runtime_1.WireType.LengthDelimited).string(message.version); writer.tag(4, runtime_1.WireType.LengthDelimited).string(message.version);
let u = options.writeUnknownFields; let u = options.writeUnknownFields;
if (u !== false) if (u !== false)
(u == true ? runtime_2.UnknownFieldHandler.onWrite : u)(this.typeName, message, writer); (u == true ? runtime_2.UnknownFieldHandler.onWrite : u)(this.typeName, message, writer);
@ -6711,15 +6689,14 @@ exports.FinalizeCacheEntryUploadResponse = new FinalizeCacheEntryUploadResponse$
class GetCacheEntryDownloadURLRequest$Type extends runtime_5.MessageType { class GetCacheEntryDownloadURLRequest$Type extends runtime_5.MessageType {
constructor() { constructor() {
super("github.actions.results.api.v1.GetCacheEntryDownloadURLRequest", [ super("github.actions.results.api.v1.GetCacheEntryDownloadURLRequest", [
{ no: 1, name: "workflow_run_backend_id", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 1, name: "metadata", kind: "message", T: () => cachemetadata_1.CacheMetadata },
{ no: 2, name: "workflow_job_run_backend_id", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 2, name: "key", kind: "scalar", T: 9 /*ScalarType.STRING*/ },
{ no: 3, name: "key", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 3, name: "restore_keys", kind: "scalar", repeat: 2 /*RepeatType.UNPACKED*/, T: 9 /*ScalarType.STRING*/ },
{ no: 4, name: "restore_keys", kind: "scalar", repeat: 2 /*RepeatType.UNPACKED*/, T: 9 /*ScalarType.STRING*/ }, { no: 4, name: "version", kind: "scalar", T: 9 /*ScalarType.STRING*/ }
{ no: 5, name: "version", kind: "scalar", T: 9 /*ScalarType.STRING*/ }
]); ]);
} }
create(value) { create(value) {
const message = { workflowRunBackendId: "", workflowJobRunBackendId: "", key: "", restoreKeys: [], version: "" }; const message = { key: "", restoreKeys: [], version: "" };
globalThis.Object.defineProperty(message, runtime_4.MESSAGE_TYPE, { enumerable: false, value: this }); globalThis.Object.defineProperty(message, runtime_4.MESSAGE_TYPE, { enumerable: false, value: this });
if (value !== undefined) if (value !== undefined)
(0, runtime_3.reflectionMergePartial)(this, message, value); (0, runtime_3.reflectionMergePartial)(this, message, value);
@ -6730,19 +6707,16 @@ class GetCacheEntryDownloadURLRequest$Type extends runtime_5.MessageType {
while (reader.pos < end) { while (reader.pos < end) {
let [fieldNo, wireType] = reader.tag(); let [fieldNo, wireType] = reader.tag();
switch (fieldNo) { switch (fieldNo) {
case /* string workflow_run_backend_id */ 1: case /* github.actions.results.entities.v1.CacheMetadata metadata */ 1:
message.workflowRunBackendId = reader.string(); message.metadata = cachemetadata_1.CacheMetadata.internalBinaryRead(reader, reader.uint32(), options, message.metadata);
break; break;
case /* string workflow_job_run_backend_id */ 2: case /* string key */ 2:
message.workflowJobRunBackendId = reader.string();
break;
case /* string key */ 3:
message.key = reader.string(); message.key = reader.string();
break; break;
case /* repeated string restore_keys */ 4: case /* repeated string restore_keys */ 3:
message.restoreKeys.push(reader.string()); message.restoreKeys.push(reader.string());
break; break;
case /* string version */ 5: case /* string version */ 4:
message.version = reader.string(); message.version = reader.string();
break; break;
default: default:
@ -6757,21 +6731,18 @@ class GetCacheEntryDownloadURLRequest$Type extends runtime_5.MessageType {
return message; return message;
} }
internalBinaryWrite(message, writer, options) { internalBinaryWrite(message, writer, options) {
/* string workflow_run_backend_id = 1; */ /* github.actions.results.entities.v1.CacheMetadata metadata = 1; */
if (message.workflowRunBackendId !== "") if (message.metadata)
writer.tag(1, runtime_1.WireType.LengthDelimited).string(message.workflowRunBackendId); cachemetadata_1.CacheMetadata.internalBinaryWrite(message.metadata, writer.tag(1, runtime_1.WireType.LengthDelimited).fork(), options).join();
/* string workflow_job_run_backend_id = 2; */ /* string key = 2; */
if (message.workflowJobRunBackendId !== "")
writer.tag(2, runtime_1.WireType.LengthDelimited).string(message.workflowJobRunBackendId);
/* string key = 3; */
if (message.key !== "") if (message.key !== "")
writer.tag(3, runtime_1.WireType.LengthDelimited).string(message.key); writer.tag(2, runtime_1.WireType.LengthDelimited).string(message.key);
/* repeated string restore_keys = 4; */ /* repeated string restore_keys = 3; */
for (let i = 0; i < message.restoreKeys.length; i++) for (let i = 0; i < message.restoreKeys.length; i++)
writer.tag(4, runtime_1.WireType.LengthDelimited).string(message.restoreKeys[i]); writer.tag(3, runtime_1.WireType.LengthDelimited).string(message.restoreKeys[i]);
/* string version = 5; */ /* string version = 4; */
if (message.version !== "") if (message.version !== "")
writer.tag(5, runtime_1.WireType.LengthDelimited).string(message.version); writer.tag(4, runtime_1.WireType.LengthDelimited).string(message.version);
let u = options.writeUnknownFields; let u = options.writeUnknownFields;
if (u !== false) if (u !== false)
(u == true ? runtime_2.UnknownFieldHandler.onWrite : u)(this.typeName, message, writer); (u == true ? runtime_2.UnknownFieldHandler.onWrite : u)(this.typeName, message, writer);
@ -6840,13 +6811,12 @@ exports.GetCacheEntryDownloadURLResponse = new GetCacheEntryDownloadURLResponse$
class DeleteCacheEntryRequest$Type extends runtime_5.MessageType { class DeleteCacheEntryRequest$Type extends runtime_5.MessageType {
constructor() { constructor() {
super("github.actions.results.api.v1.DeleteCacheEntryRequest", [ super("github.actions.results.api.v1.DeleteCacheEntryRequest", [
{ no: 1, name: "workflow_run_backend_id", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 1, name: "metadata", kind: "message", T: () => cachemetadata_1.CacheMetadata },
{ no: 2, name: "workflow_job_run_backend_id", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 2, name: "key", kind: "scalar", T: 9 /*ScalarType.STRING*/ }
{ no: 3, name: "key", kind: "scalar", T: 9 /*ScalarType.STRING*/ }
]); ]);
} }
create(value) { create(value) {
const message = { workflowRunBackendId: "", workflowJobRunBackendId: "", key: "" }; const message = { key: "" };
globalThis.Object.defineProperty(message, runtime_4.MESSAGE_TYPE, { enumerable: false, value: this }); globalThis.Object.defineProperty(message, runtime_4.MESSAGE_TYPE, { enumerable: false, value: this });
if (value !== undefined) if (value !== undefined)
(0, runtime_3.reflectionMergePartial)(this, message, value); (0, runtime_3.reflectionMergePartial)(this, message, value);
@ -6857,13 +6827,10 @@ class DeleteCacheEntryRequest$Type extends runtime_5.MessageType {
while (reader.pos < end) { while (reader.pos < end) {
let [fieldNo, wireType] = reader.tag(); let [fieldNo, wireType] = reader.tag();
switch (fieldNo) { switch (fieldNo) {
case /* string workflow_run_backend_id */ 1: case /* github.actions.results.entities.v1.CacheMetadata metadata */ 1:
message.workflowRunBackendId = reader.string(); message.metadata = cachemetadata_1.CacheMetadata.internalBinaryRead(reader, reader.uint32(), options, message.metadata);
break; break;
case /* string workflow_job_run_backend_id */ 2: case /* string key */ 2:
message.workflowJobRunBackendId = reader.string();
break;
case /* string key */ 3:
message.key = reader.string(); message.key = reader.string();
break; break;
default: default:
@ -6878,15 +6845,12 @@ class DeleteCacheEntryRequest$Type extends runtime_5.MessageType {
return message; return message;
} }
internalBinaryWrite(message, writer, options) { internalBinaryWrite(message, writer, options) {
/* string workflow_run_backend_id = 1; */ /* github.actions.results.entities.v1.CacheMetadata metadata = 1; */
if (message.workflowRunBackendId !== "") if (message.metadata)
writer.tag(1, runtime_1.WireType.LengthDelimited).string(message.workflowRunBackendId); cachemetadata_1.CacheMetadata.internalBinaryWrite(message.metadata, writer.tag(1, runtime_1.WireType.LengthDelimited).fork(), options).join();
/* string workflow_job_run_backend_id = 2; */ /* string key = 2; */
if (message.workflowJobRunBackendId !== "")
writer.tag(2, runtime_1.WireType.LengthDelimited).string(message.workflowJobRunBackendId);
/* string key = 3; */
if (message.key !== "") if (message.key !== "")
writer.tag(3, runtime_1.WireType.LengthDelimited).string(message.key); writer.tag(2, runtime_1.WireType.LengthDelimited).string(message.key);
let u = options.writeUnknownFields; let u = options.writeUnknownFields;
if (u !== false) if (u !== false)
(u == true ? runtime_2.UnknownFieldHandler.onWrite : u)(this.typeName, message, writer); (u == true ? runtime_2.UnknownFieldHandler.onWrite : u)(this.typeName, message, writer);
@ -6955,14 +6919,13 @@ exports.DeleteCacheEntryResponse = new DeleteCacheEntryResponse$Type();
class ListCacheEntriesRequest$Type extends runtime_5.MessageType { class ListCacheEntriesRequest$Type extends runtime_5.MessageType {
constructor() { constructor() {
super("github.actions.results.api.v1.ListCacheEntriesRequest", [ super("github.actions.results.api.v1.ListCacheEntriesRequest", [
{ no: 1, name: "workflow_run_backend_id", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 1, name: "metadata", kind: "message", T: () => cachemetadata_1.CacheMetadata },
{ no: 2, name: "workflow_job_run_backend_id", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 2, name: "key", kind: "scalar", T: 9 /*ScalarType.STRING*/ },
{ no: 3, name: "key", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 3, name: "restore_keys", kind: "scalar", repeat: 2 /*RepeatType.UNPACKED*/, T: 9 /*ScalarType.STRING*/ }
{ no: 4, name: "restore_keys", kind: "scalar", repeat: 2 /*RepeatType.UNPACKED*/, T: 9 /*ScalarType.STRING*/ }
]); ]);
} }
create(value) { create(value) {
const message = { workflowRunBackendId: "", workflowJobRunBackendId: "", key: "", restoreKeys: [] }; const message = { key: "", restoreKeys: [] };
globalThis.Object.defineProperty(message, runtime_4.MESSAGE_TYPE, { enumerable: false, value: this }); globalThis.Object.defineProperty(message, runtime_4.MESSAGE_TYPE, { enumerable: false, value: this });
if (value !== undefined) if (value !== undefined)
(0, runtime_3.reflectionMergePartial)(this, message, value); (0, runtime_3.reflectionMergePartial)(this, message, value);
@ -6973,16 +6936,13 @@ class ListCacheEntriesRequest$Type extends runtime_5.MessageType {
while (reader.pos < end) { while (reader.pos < end) {
let [fieldNo, wireType] = reader.tag(); let [fieldNo, wireType] = reader.tag();
switch (fieldNo) { switch (fieldNo) {
case /* string workflow_run_backend_id */ 1: case /* github.actions.results.entities.v1.CacheMetadata metadata */ 1:
message.workflowRunBackendId = reader.string(); message.metadata = cachemetadata_1.CacheMetadata.internalBinaryRead(reader, reader.uint32(), options, message.metadata);
break; break;
case /* string workflow_job_run_backend_id */ 2: case /* string key */ 2:
message.workflowJobRunBackendId = reader.string();
break;
case /* string key */ 3:
message.key = reader.string(); message.key = reader.string();
break; break;
case /* repeated string restore_keys */ 4: case /* repeated string restore_keys */ 3:
message.restoreKeys.push(reader.string()); message.restoreKeys.push(reader.string());
break; break;
default: default:
@ -6997,18 +6957,15 @@ class ListCacheEntriesRequest$Type extends runtime_5.MessageType {
return message; return message;
} }
internalBinaryWrite(message, writer, options) { internalBinaryWrite(message, writer, options) {
/* string workflow_run_backend_id = 1; */ /* github.actions.results.entities.v1.CacheMetadata metadata = 1; */
if (message.workflowRunBackendId !== "") if (message.metadata)
writer.tag(1, runtime_1.WireType.LengthDelimited).string(message.workflowRunBackendId); cachemetadata_1.CacheMetadata.internalBinaryWrite(message.metadata, writer.tag(1, runtime_1.WireType.LengthDelimited).fork(), options).join();
/* string workflow_job_run_backend_id = 2; */ /* string key = 2; */
if (message.workflowJobRunBackendId !== "")
writer.tag(2, runtime_1.WireType.LengthDelimited).string(message.workflowJobRunBackendId);
/* string key = 3; */
if (message.key !== "") if (message.key !== "")
writer.tag(3, runtime_1.WireType.LengthDelimited).string(message.key); writer.tag(2, runtime_1.WireType.LengthDelimited).string(message.key);
/* repeated string restore_keys = 4; */ /* repeated string restore_keys = 3; */
for (let i = 0; i < message.restoreKeys.length; i++) for (let i = 0; i < message.restoreKeys.length; i++)
writer.tag(4, runtime_1.WireType.LengthDelimited).string(message.restoreKeys[i]); writer.tag(3, runtime_1.WireType.LengthDelimited).string(message.restoreKeys[i]);
let u = options.writeUnknownFields; let u = options.writeUnknownFields;
if (u !== false) if (u !== false)
(u == true ? runtime_2.UnknownFieldHandler.onWrite : u)(this.typeName, message, writer); (u == true ? runtime_2.UnknownFieldHandler.onWrite : u)(this.typeName, message, writer);
@ -7166,15 +7123,14 @@ exports.ListCacheEntriesResponse_CacheEntry = new ListCacheEntriesResponse_Cache
class LookupCacheEntryRequest$Type extends runtime_5.MessageType { class LookupCacheEntryRequest$Type extends runtime_5.MessageType {
constructor() { constructor() {
super("github.actions.results.api.v1.LookupCacheEntryRequest", [ super("github.actions.results.api.v1.LookupCacheEntryRequest", [
{ no: 1, name: "workflow_run_backend_id", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 1, name: "metadata", kind: "message", T: () => cachemetadata_1.CacheMetadata },
{ no: 2, name: "workflow_job_run_backend_id", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 2, name: "key", kind: "scalar", T: 9 /*ScalarType.STRING*/ },
{ no: 3, name: "key", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 3, name: "restore_keys", kind: "scalar", repeat: 2 /*RepeatType.UNPACKED*/, T: 9 /*ScalarType.STRING*/ },
{ no: 4, name: "restore_keys", kind: "scalar", repeat: 2 /*RepeatType.UNPACKED*/, T: 9 /*ScalarType.STRING*/ }, { no: 4, name: "version", kind: "scalar", T: 9 /*ScalarType.STRING*/ }
{ no: 5, name: "version", kind: "scalar", T: 9 /*ScalarType.STRING*/ }
]); ]);
} }
create(value) { create(value) {
const message = { workflowRunBackendId: "", workflowJobRunBackendId: "", key: "", restoreKeys: [], version: "" }; const message = { key: "", restoreKeys: [], version: "" };
globalThis.Object.defineProperty(message, runtime_4.MESSAGE_TYPE, { enumerable: false, value: this }); globalThis.Object.defineProperty(message, runtime_4.MESSAGE_TYPE, { enumerable: false, value: this });
if (value !== undefined) if (value !== undefined)
(0, runtime_3.reflectionMergePartial)(this, message, value); (0, runtime_3.reflectionMergePartial)(this, message, value);
@ -7185,19 +7141,16 @@ class LookupCacheEntryRequest$Type extends runtime_5.MessageType {
while (reader.pos < end) { while (reader.pos < end) {
let [fieldNo, wireType] = reader.tag(); let [fieldNo, wireType] = reader.tag();
switch (fieldNo) { switch (fieldNo) {
case /* string workflow_run_backend_id */ 1: case /* github.actions.results.entities.v1.CacheMetadata metadata */ 1:
message.workflowRunBackendId = reader.string(); message.metadata = cachemetadata_1.CacheMetadata.internalBinaryRead(reader, reader.uint32(), options, message.metadata);
break; break;
case /* string workflow_job_run_backend_id */ 2: case /* string key */ 2:
message.workflowJobRunBackendId = reader.string();
break;
case /* string key */ 3:
message.key = reader.string(); message.key = reader.string();
break; break;
case /* repeated string restore_keys */ 4: case /* repeated string restore_keys */ 3:
message.restoreKeys.push(reader.string()); message.restoreKeys.push(reader.string());
break; break;
case /* string version */ 5: case /* string version */ 4:
message.version = reader.string(); message.version = reader.string();
break; break;
default: default:
@ -7212,21 +7165,18 @@ class LookupCacheEntryRequest$Type extends runtime_5.MessageType {
return message; return message;
} }
internalBinaryWrite(message, writer, options) { internalBinaryWrite(message, writer, options) {
/* string workflow_run_backend_id = 1; */ /* github.actions.results.entities.v1.CacheMetadata metadata = 1; */
if (message.workflowRunBackendId !== "") if (message.metadata)
writer.tag(1, runtime_1.WireType.LengthDelimited).string(message.workflowRunBackendId); cachemetadata_1.CacheMetadata.internalBinaryWrite(message.metadata, writer.tag(1, runtime_1.WireType.LengthDelimited).fork(), options).join();
/* string workflow_job_run_backend_id = 2; */ /* string key = 2; */
if (message.workflowJobRunBackendId !== "")
writer.tag(2, runtime_1.WireType.LengthDelimited).string(message.workflowJobRunBackendId);
/* string key = 3; */
if (message.key !== "") if (message.key !== "")
writer.tag(3, runtime_1.WireType.LengthDelimited).string(message.key); writer.tag(2, runtime_1.WireType.LengthDelimited).string(message.key);
/* repeated string restore_keys = 4; */ /* repeated string restore_keys = 3; */
for (let i = 0; i < message.restoreKeys.length; i++) for (let i = 0; i < message.restoreKeys.length; i++)
writer.tag(4, runtime_1.WireType.LengthDelimited).string(message.restoreKeys[i]); writer.tag(3, runtime_1.WireType.LengthDelimited).string(message.restoreKeys[i]);
/* string version = 5; */ /* string version = 4; */
if (message.version !== "") if (message.version !== "")
writer.tag(5, runtime_1.WireType.LengthDelimited).string(message.version); writer.tag(4, runtime_1.WireType.LengthDelimited).string(message.version);
let u = options.writeUnknownFields; let u = options.writeUnknownFields;
if (u !== false) if (u !== false)
(u == true ? runtime_2.UnknownFieldHandler.onWrite : u)(this.typeName, message, writer); (u == true ? runtime_2.UnknownFieldHandler.onWrite : u)(this.typeName, message, writer);
@ -8004,6 +7954,147 @@ function handleCacheServiceLookupCacheEntryProtobuf(ctx, service, data, intercep
/***/ }), /***/ }),
/***/ 2773:
/***/ ((__unused_webpack_module, exports, __nccwpck_require__) => {
"use strict";
Object.defineProperty(exports, "__esModule", ({ value: true }));
exports.CacheMetadata = void 0;
const runtime_1 = __nccwpck_require__(3503);
const runtime_2 = __nccwpck_require__(3503);
const runtime_3 = __nccwpck_require__(3503);
const runtime_4 = __nccwpck_require__(3503);
const runtime_5 = __nccwpck_require__(3503);
const cachescope_1 = __nccwpck_require__(7685);
// @generated message type with reflection information, may provide speed optimized methods
class CacheMetadata$Type extends runtime_5.MessageType {
constructor() {
super("github.actions.results.entities.v1.CacheMetadata", [
{ no: 1, name: "repository_id", kind: "scalar", T: 3 /*ScalarType.INT64*/ },
{ no: 2, name: "scope", kind: "message", repeat: 1 /*RepeatType.PACKED*/, T: () => cachescope_1.CacheScope }
]);
}
create(value) {
const message = { repositoryId: "0", scope: [] };
globalThis.Object.defineProperty(message, runtime_4.MESSAGE_TYPE, { enumerable: false, value: this });
if (value !== undefined)
(0, runtime_3.reflectionMergePartial)(this, message, value);
return message;
}
internalBinaryRead(reader, length, options, target) {
let message = target !== null && target !== void 0 ? target : this.create(), end = reader.pos + length;
while (reader.pos < end) {
let [fieldNo, wireType] = reader.tag();
switch (fieldNo) {
case /* int64 repository_id */ 1:
message.repositoryId = reader.int64().toString();
break;
case /* repeated github.actions.results.entities.v1.CacheScope scope */ 2:
message.scope.push(cachescope_1.CacheScope.internalBinaryRead(reader, reader.uint32(), options));
break;
default:
let u = options.readUnknownField;
if (u === "throw")
throw new globalThis.Error(`Unknown field ${fieldNo} (wire type ${wireType}) for ${this.typeName}`);
let d = reader.skip(wireType);
if (u !== false)
(u === true ? runtime_2.UnknownFieldHandler.onRead : u)(this.typeName, message, fieldNo, wireType, d);
}
}
return message;
}
internalBinaryWrite(message, writer, options) {
/* int64 repository_id = 1; */
if (message.repositoryId !== "0")
writer.tag(1, runtime_1.WireType.Varint).int64(message.repositoryId);
/* repeated github.actions.results.entities.v1.CacheScope scope = 2; */
for (let i = 0; i < message.scope.length; i++)
cachescope_1.CacheScope.internalBinaryWrite(message.scope[i], writer.tag(2, runtime_1.WireType.LengthDelimited).fork(), options).join();
let u = options.writeUnknownFields;
if (u !== false)
(u == true ? runtime_2.UnknownFieldHandler.onWrite : u)(this.typeName, message, writer);
return writer;
}
}
/**
* @generated MessageType for protobuf message github.actions.results.entities.v1.CacheMetadata
*/
exports.CacheMetadata = new CacheMetadata$Type();
//# sourceMappingURL=cachemetadata.js.map
/***/ }),
/***/ 7685:
/***/ ((__unused_webpack_module, exports, __nccwpck_require__) => {
"use strict";
Object.defineProperty(exports, "__esModule", ({ value: true }));
exports.CacheScope = void 0;
const runtime_1 = __nccwpck_require__(3503);
const runtime_2 = __nccwpck_require__(3503);
const runtime_3 = __nccwpck_require__(3503);
const runtime_4 = __nccwpck_require__(3503);
const runtime_5 = __nccwpck_require__(3503);
// @generated message type with reflection information, may provide speed optimized methods
class CacheScope$Type extends runtime_5.MessageType {
constructor() {
super("github.actions.results.entities.v1.CacheScope", [
{ no: 1, name: "scope", kind: "scalar", T: 9 /*ScalarType.STRING*/ },
{ no: 2, name: "permission", kind: "scalar", T: 3 /*ScalarType.INT64*/ }
]);
}
create(value) {
const message = { scope: "", permission: "0" };
globalThis.Object.defineProperty(message, runtime_4.MESSAGE_TYPE, { enumerable: false, value: this });
if (value !== undefined)
(0, runtime_3.reflectionMergePartial)(this, message, value);
return message;
}
internalBinaryRead(reader, length, options, target) {
let message = target !== null && target !== void 0 ? target : this.create(), end = reader.pos + length;
while (reader.pos < end) {
let [fieldNo, wireType] = reader.tag();
switch (fieldNo) {
case /* string scope */ 1:
message.scope = reader.string();
break;
case /* int64 permission */ 2:
message.permission = reader.int64().toString();
break;
default:
let u = options.readUnknownField;
if (u === "throw")
throw new globalThis.Error(`Unknown field ${fieldNo} (wire type ${wireType}) for ${this.typeName}`);
let d = reader.skip(wireType);
if (u !== false)
(u === true ? runtime_2.UnknownFieldHandler.onRead : u)(this.typeName, message, fieldNo, wireType, d);
}
}
return message;
}
internalBinaryWrite(message, writer, options) {
/* string scope = 1; */
if (message.scope !== "")
writer.tag(1, runtime_1.WireType.LengthDelimited).string(message.scope);
/* int64 permission = 2; */
if (message.permission !== "0")
writer.tag(2, runtime_1.WireType.Varint).int64(message.permission);
let u = options.writeUnknownFields;
if (u !== false)
(u == true ? runtime_2.UnknownFieldHandler.onWrite : u)(this.typeName, message, writer);
return writer;
}
}
/**
* @generated MessageType for protobuf message github.actions.results.entities.v1.CacheScope
*/
exports.CacheScope = new CacheScope$Type();
//# sourceMappingURL=cachescope.js.map
/***/ }),
/***/ 4076: /***/ 4076:
/***/ (function(__unused_webpack_module, exports, __nccwpck_require__) { /***/ (function(__unused_webpack_module, exports, __nccwpck_require__) {
@ -8416,11 +8507,8 @@ var __asyncValues = (this && this.__asyncValues) || function (o) {
function verb(n) { i[n] = o[n] && function (v) { return new Promise(function (resolve, reject) { v = o[n](v), settle(resolve, reject, v.done, v.value); }); }; } function verb(n) { i[n] = o[n] && function (v) { return new Promise(function (resolve, reject) { v = o[n](v), settle(resolve, reject, v.done, v.value); }); }; }
function settle(resolve, reject, d, v) { Promise.resolve(v).then(function(v) { resolve({ value: v, done: d }); }, reject); } function settle(resolve, reject, d, v) { Promise.resolve(v).then(function(v) { resolve({ value: v, done: d }); }, reject); }
}; };
var __importDefault = (this && this.__importDefault) || function (mod) {
return (mod && mod.__esModule) ? mod : { "default": mod };
};
Object.defineProperty(exports, "__esModule", ({ value: true })); Object.defineProperty(exports, "__esModule", ({ value: true }));
exports.getBackendIdsFromToken = exports.getRuntimeToken = exports.getCacheVersion = exports.isGhes = exports.assertDefined = exports.getGnuTarPathOnWindows = exports.getCacheFileName = exports.getCompressionMethod = exports.unlinkFile = exports.resolvePaths = exports.getArchiveFileSizeInBytes = exports.createTempDirectory = void 0; exports.getRuntimeToken = exports.getCacheVersion = exports.isGhes = exports.assertDefined = exports.getGnuTarPathOnWindows = exports.getCacheFileName = exports.getCompressionMethod = exports.unlinkFile = exports.resolvePaths = exports.getArchiveFileSizeInBytes = exports.createTempDirectory = void 0;
const core = __importStar(__nccwpck_require__(4850)); const core = __importStar(__nccwpck_require__(4850));
const exec = __importStar(__nccwpck_require__(309)); const exec = __importStar(__nccwpck_require__(309));
const glob = __importStar(__nccwpck_require__(9590)); const glob = __importStar(__nccwpck_require__(9590));
@ -8430,7 +8518,6 @@ const fs = __importStar(__nccwpck_require__(7147));
const path = __importStar(__nccwpck_require__(1017)); const path = __importStar(__nccwpck_require__(1017));
const semver = __importStar(__nccwpck_require__(3910)); const semver = __importStar(__nccwpck_require__(3910));
const util = __importStar(__nccwpck_require__(3837)); const util = __importStar(__nccwpck_require__(3837));
const jwt_decode_1 = __importDefault(__nccwpck_require__(1300));
const constants_1 = __nccwpck_require__(4010); const constants_1 = __nccwpck_require__(4010);
const versionSalt = '1.0'; const versionSalt = '1.0';
// From https://github.com/actions/toolkit/blob/main/packages/tool-cache/src/tool-cache.ts#L23 // From https://github.com/actions/toolkit/blob/main/packages/tool-cache/src/tool-cache.ts#L23
@ -8603,54 +8690,6 @@ function getRuntimeToken() {
return token; return token;
} }
exports.getRuntimeToken = getRuntimeToken; exports.getRuntimeToken = getRuntimeToken;
const InvalidJwtError = new Error('Failed to get backend IDs: The provided JWT token is invalid and/or missing claims');
// uses the JWT token claims to get the
// workflow run and workflow job run backend ids
function getBackendIdsFromToken() {
const token = getRuntimeToken();
const decoded = (0, jwt_decode_1.default)(token);
if (!decoded.scp) {
throw InvalidJwtError;
}
/*
* example decoded:
* {
* scp: "Actions.ExampleScope Actions.Results:ce7f54c7-61c7-4aae-887f-30da475f5f1a:ca395085-040a-526b-2ce8-bdc85f692774"
* }
*/
const scpParts = decoded.scp.split(' ');
if (scpParts.length === 0) {
throw InvalidJwtError;
}
/*
* example scpParts:
* ["Actions.ExampleScope", "Actions.Results:ce7f54c7-61c7-4aae-887f-30da475f5f1a:ca395085-040a-526b-2ce8-bdc85f692774"]
*/
for (const scopes of scpParts) {
const scopeParts = scopes.split(':');
if ((scopeParts === null || scopeParts === void 0 ? void 0 : scopeParts[0]) !== 'Actions.Results') {
// not the Actions.Results scope
continue;
}
/*
* example scopeParts:
* ["Actions.Results", "ce7f54c7-61c7-4aae-887f-30da475f5f1a", "ca395085-040a-526b-2ce8-bdc85f692774"]
*/
if (scopeParts.length !== 3) {
// missing expected number of claims
throw InvalidJwtError;
}
const ids = {
workflowRunBackendId: scopeParts[1],
workflowJobRunBackendId: scopeParts[2]
};
core.debug(`Workflow Run Backend ID: ${ids.workflowRunBackendId}`);
core.debug(`Workflow Job Run Backend ID: ${ids.workflowJobRunBackendId}`);
return ids;
}
throw InvalidJwtError;
}
exports.getBackendIdsFromToken = getBackendIdsFromToken;
//# sourceMappingURL=cacheUtils.js.map //# sourceMappingURL=cacheUtils.js.map
/***/ }), /***/ }),
@ -8661,15 +8700,7 @@ exports.getBackendIdsFromToken = getBackendIdsFromToken;
"use strict"; "use strict";
Object.defineProperty(exports, "__esModule", ({ value: true })); Object.defineProperty(exports, "__esModule", ({ value: true }));
exports.getCacheServiceURL = exports.getCacheServiceVersion = exports.getRuntimeToken = void 0; exports.getCacheServiceURL = exports.getCacheServiceVersion = void 0;
function getRuntimeToken() {
const token = process.env['ACTIONS_RUNTIME_TOKEN'];
if (!token) {
throw new Error('Unable to get the ACTIONS_RUNTIME_TOKEN env variable');
}
return token;
}
exports.getRuntimeToken = getRuntimeToken;
function getCacheServiceVersion() { function getCacheServiceVersion() {
return process.env['ACTIONS_CACHE_SERVICE_V2'] ? 'v2' : 'v1'; return process.env['ACTIONS_CACHE_SERVICE_V2'] ? 'v2' : 'v1';
} }
@ -9285,6 +9316,7 @@ const core_1 = __nccwpck_require__(4850);
const user_agent_1 = __nccwpck_require__(5736); const user_agent_1 = __nccwpck_require__(5736);
const errors_1 = __nccwpck_require__(6333); const errors_1 = __nccwpck_require__(6333);
const config_1 = __nccwpck_require__(6490); const config_1 = __nccwpck_require__(6490);
const cacheUtils_1 = __nccwpck_require__(3310);
const auth_1 = __nccwpck_require__(7231); const auth_1 = __nccwpck_require__(7231);
const http_client_1 = __nccwpck_require__(1283); const http_client_1 = __nccwpck_require__(1283);
const cache_twirp_1 = __nccwpck_require__(894); const cache_twirp_1 = __nccwpck_require__(894);
@ -9300,7 +9332,7 @@ class CacheServiceClient {
this.maxAttempts = 5; this.maxAttempts = 5;
this.baseRetryIntervalMilliseconds = 3000; this.baseRetryIntervalMilliseconds = 3000;
this.retryMultiplier = 1.5; this.retryMultiplier = 1.5;
const token = (0, config_1.getRuntimeToken)(); const token = (0, cacheUtils_1.getRuntimeToken)();
this.baseUrl = (0, config_1.getCacheServiceURL)(); this.baseUrl = (0, config_1.getCacheServiceURL)();
if (maxAttempts) { if (maxAttempts) {
this.maxAttempts = maxAttempts; this.maxAttempts = maxAttempts;
@ -57677,16 +57709,6 @@ DotObject._process = _process
module.exports = DotObject module.exports = DotObject
/***/ }),
/***/ 1300:
/***/ ((module) => {
"use strict";
function e(e){this.message=e}e.prototype=new Error,e.prototype.name="InvalidCharacterError";var r="undefined"!=typeof window&&window.atob&&window.atob.bind(window)||function(r){var t=String(r).replace(/=+$/,"");if(t.length%4==1)throw new e("'atob' failed: The string to be decoded is not correctly encoded.");for(var n,o,a=0,i=0,c="";o=t.charAt(i++);~o&&(n=a%4?64*n+o:o,a++%4)?c+=String.fromCharCode(255&n>>(-2*a&6)):0)o="ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/=".indexOf(o);return c};function t(e){var t=e.replace(/-/g,"+").replace(/_/g,"/");switch(t.length%4){case 0:break;case 2:t+="==";break;case 3:t+="=";break;default:throw"Illegal base64url string!"}try{return function(e){return decodeURIComponent(r(e).replace(/(.)/g,(function(e,r){var t=r.charCodeAt(0).toString(16).toUpperCase();return t.length<2&&(t="0"+t),"%"+t})))}(t)}catch(e){return r(t)}}function n(e){this.message=e}function o(e,r){if("string"!=typeof e)throw new n("Invalid token specified");var o=!0===(r=r||{}).header?0:1;try{return JSON.parse(t(e.split(".")[o]))}catch(e){throw new n("Invalid token specified: "+e.message)}}n.prototype=new Error,n.prototype.name="InvalidTokenError";const a=o;a.default=o,a.InvalidTokenError=n,module.exports=a;
//# sourceMappingURL=jwt-decode.cjs.js.map
/***/ }), /***/ }),
/***/ 2020: /***/ 2020:
@ -73498,7 +73520,7 @@ module.exports = JSON.parse('[[[0,44],"disallowed_STD3_valid"],[[45,46],"valid"]
/***/ ((module) => { /***/ ((module) => {
"use strict"; "use strict";
module.exports = JSON.parse('{"name":"@actions/cache","version":"3.3.0","preview":true,"description":"Actions cache lib","keywords":["github","actions","cache"],"homepage":"https://github.com/actions/toolkit/tree/main/packages/cache","license":"MIT","main":"lib/cache.js","types":"lib/cache.d.ts","directories":{"lib":"lib","test":"__tests__"},"files":["lib","!.DS_Store"],"publishConfig":{"access":"public"},"repository":{"type":"git","url":"git+https://github.com/actions/toolkit.git","directory":"packages/cache"},"scripts":{"audit-moderate":"npm install && npm audit --json --audit-level=moderate > audit.json","test":"echo \\"Error: run tests from root\\" && exit 1","tsc":"tsc"},"bugs":{"url":"https://github.com/actions/toolkit/issues"},"dependencies":{"@actions/core":"^1.11.1","@actions/exec":"^1.0.1","@actions/glob":"^0.1.0","@actions/http-client":"^2.1.1","@actions/io":"^1.0.1","@azure/abort-controller":"^1.1.0","@azure/ms-rest-js":"^2.6.0","@azure/storage-blob":"^12.13.0","@protobuf-ts/plugin":"^2.9.4","semver":"^6.3.1","jwt-decode":"^3.1.2","twirp-ts":"^2.5.0"},"devDependencies":{"@types/semver":"^6.0.0","typescript":"^5.2.2"}}'); module.exports = JSON.parse('{"name":"@actions/cache","version":"3.3.0","preview":true,"description":"Actions cache lib","keywords":["github","actions","cache"],"homepage":"https://github.com/actions/toolkit/tree/main/packages/cache","license":"MIT","main":"lib/cache.js","types":"lib/cache.d.ts","directories":{"lib":"lib","test":"__tests__"},"files":["lib","!.DS_Store"],"publishConfig":{"access":"public"},"repository":{"type":"git","url":"git+https://github.com/actions/toolkit.git","directory":"packages/cache"},"scripts":{"audit-moderate":"npm install && npm audit --json --audit-level=moderate > audit.json","test":"echo \\"Error: run tests from root\\" && exit 1","tsc":"tsc"},"bugs":{"url":"https://github.com/actions/toolkit/issues"},"dependencies":{"@actions/core":"^1.11.1","@actions/exec":"^1.0.1","@actions/glob":"^0.1.0","@actions/http-client":"^2.1.1","@actions/io":"^1.0.1","@azure/abort-controller":"^1.1.0","@azure/ms-rest-js":"^2.6.0","@azure/storage-blob":"^12.13.0","@protobuf-ts/plugin":"^2.9.4","semver":"^6.3.1","twirp-ts":"^2.5.0"},"devDependencies":{"@types/semver":"^6.0.0","typescript":"^5.2.2"}}');
/***/ }) /***/ })

468
dist/save/index.js vendored
View File

@ -6064,11 +6064,8 @@ function restoreCacheV2(paths, primaryKey, restoreKeys, options, enableCrossOsAr
let archivePath = ''; let archivePath = '';
try { try {
const twirpClient = cacheTwirpClient.internalCacheTwirpClient(); const twirpClient = cacheTwirpClient.internalCacheTwirpClient();
const backendIds = utils.getBackendIdsFromToken();
const compressionMethod = yield utils.getCompressionMethod(); const compressionMethod = yield utils.getCompressionMethod();
const request = { const request = {
workflowRunBackendId: backendIds.workflowRunBackendId,
workflowJobRunBackendId: backendIds.workflowJobRunBackendId,
key: primaryKey, key: primaryKey,
restoreKeys, restoreKeys,
version: utils.getCacheVersion(paths, compressionMethod, enableCrossOsArchive) version: utils.getCacheVersion(paths, compressionMethod, enableCrossOsArchive)
@ -6223,8 +6220,6 @@ function saveCacheV1(paths, key, options, enableCrossOsArchive = false) {
*/ */
function saveCacheV2(paths, key, options, enableCrossOsArchive = false) { function saveCacheV2(paths, key, options, enableCrossOsArchive = false) {
return __awaiter(this, void 0, void 0, function* () { return __awaiter(this, void 0, void 0, function* () {
// BackendIds are retrieved form the signed JWT
const backendIds = utils.getBackendIdsFromToken();
const compressionMethod = yield utils.getCompressionMethod(); const compressionMethod = yield utils.getCompressionMethod();
const twirpClient = cacheTwirpClient.internalCacheTwirpClient(); const twirpClient = cacheTwirpClient.internalCacheTwirpClient();
let cacheId = -1; let cacheId = -1;
@ -6251,8 +6246,6 @@ function saveCacheV2(paths, key, options, enableCrossOsArchive = false) {
core.debug('Reserving Cache'); core.debug('Reserving Cache');
const version = utils.getCacheVersion(paths, compressionMethod, enableCrossOsArchive); const version = utils.getCacheVersion(paths, compressionMethod, enableCrossOsArchive);
const request = { const request = {
workflowRunBackendId: backendIds.workflowRunBackendId,
workflowJobRunBackendId: backendIds.workflowJobRunBackendId,
key, key,
version version
}; };
@ -6263,8 +6256,6 @@ function saveCacheV2(paths, key, options, enableCrossOsArchive = false) {
core.debug(`Attempting to upload cache located at: ${archivePath}`); core.debug(`Attempting to upload cache located at: ${archivePath}`);
yield (0, upload_cache_1.UploadCacheFile)(response.signedUploadUrl, archivePath); yield (0, upload_cache_1.UploadCacheFile)(response.signedUploadUrl, archivePath);
const finalizeRequest = { const finalizeRequest = {
workflowRunBackendId: backendIds.workflowRunBackendId,
workflowJobRunBackendId: backendIds.workflowJobRunBackendId,
key, key,
version, version,
sizeBytes: `${archiveFileSize}` sizeBytes: `${archiveFileSize}`
@ -6456,18 +6447,18 @@ const runtime_3 = __nccwpck_require__(3503);
const runtime_4 = __nccwpck_require__(3503); const runtime_4 = __nccwpck_require__(3503);
const runtime_5 = __nccwpck_require__(3503); const runtime_5 = __nccwpck_require__(3503);
const timestamp_1 = __nccwpck_require__(8983); const timestamp_1 = __nccwpck_require__(8983);
const cachemetadata_1 = __nccwpck_require__(2773);
// @generated message type with reflection information, may provide speed optimized methods // @generated message type with reflection information, may provide speed optimized methods
class CreateCacheEntryRequest$Type extends runtime_5.MessageType { class CreateCacheEntryRequest$Type extends runtime_5.MessageType {
constructor() { constructor() {
super("github.actions.results.api.v1.CreateCacheEntryRequest", [ super("github.actions.results.api.v1.CreateCacheEntryRequest", [
{ no: 1, name: "workflow_run_backend_id", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 1, name: "metadata", kind: "message", T: () => cachemetadata_1.CacheMetadata },
{ no: 2, name: "workflow_job_run_backend_id", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 2, name: "key", kind: "scalar", T: 9 /*ScalarType.STRING*/ },
{ no: 3, name: "key", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 3, name: "version", kind: "scalar", T: 9 /*ScalarType.STRING*/ }
{ no: 4, name: "version", kind: "scalar", T: 9 /*ScalarType.STRING*/ }
]); ]);
} }
create(value) { create(value) {
const message = { workflowRunBackendId: "", workflowJobRunBackendId: "", key: "", version: "" }; const message = { key: "", version: "" };
globalThis.Object.defineProperty(message, runtime_4.MESSAGE_TYPE, { enumerable: false, value: this }); globalThis.Object.defineProperty(message, runtime_4.MESSAGE_TYPE, { enumerable: false, value: this });
if (value !== undefined) if (value !== undefined)
(0, runtime_3.reflectionMergePartial)(this, message, value); (0, runtime_3.reflectionMergePartial)(this, message, value);
@ -6478,16 +6469,13 @@ class CreateCacheEntryRequest$Type extends runtime_5.MessageType {
while (reader.pos < end) { while (reader.pos < end) {
let [fieldNo, wireType] = reader.tag(); let [fieldNo, wireType] = reader.tag();
switch (fieldNo) { switch (fieldNo) {
case /* string workflow_run_backend_id */ 1: case /* github.actions.results.entities.v1.CacheMetadata metadata */ 1:
message.workflowRunBackendId = reader.string(); message.metadata = cachemetadata_1.CacheMetadata.internalBinaryRead(reader, reader.uint32(), options, message.metadata);
break; break;
case /* string workflow_job_run_backend_id */ 2: case /* string key */ 2:
message.workflowJobRunBackendId = reader.string();
break;
case /* string key */ 3:
message.key = reader.string(); message.key = reader.string();
break; break;
case /* string version */ 4: case /* string version */ 3:
message.version = reader.string(); message.version = reader.string();
break; break;
default: default:
@ -6502,18 +6490,15 @@ class CreateCacheEntryRequest$Type extends runtime_5.MessageType {
return message; return message;
} }
internalBinaryWrite(message, writer, options) { internalBinaryWrite(message, writer, options) {
/* string workflow_run_backend_id = 1; */ /* github.actions.results.entities.v1.CacheMetadata metadata = 1; */
if (message.workflowRunBackendId !== "") if (message.metadata)
writer.tag(1, runtime_1.WireType.LengthDelimited).string(message.workflowRunBackendId); cachemetadata_1.CacheMetadata.internalBinaryWrite(message.metadata, writer.tag(1, runtime_1.WireType.LengthDelimited).fork(), options).join();
/* string workflow_job_run_backend_id = 2; */ /* string key = 2; */
if (message.workflowJobRunBackendId !== "")
writer.tag(2, runtime_1.WireType.LengthDelimited).string(message.workflowJobRunBackendId);
/* string key = 3; */
if (message.key !== "") if (message.key !== "")
writer.tag(3, runtime_1.WireType.LengthDelimited).string(message.key); writer.tag(2, runtime_1.WireType.LengthDelimited).string(message.key);
/* string version = 4; */ /* string version = 3; */
if (message.version !== "") if (message.version !== "")
writer.tag(4, runtime_1.WireType.LengthDelimited).string(message.version); writer.tag(3, runtime_1.WireType.LengthDelimited).string(message.version);
let u = options.writeUnknownFields; let u = options.writeUnknownFields;
if (u !== false) if (u !== false)
(u == true ? runtime_2.UnknownFieldHandler.onWrite : u)(this.typeName, message, writer); (u == true ? runtime_2.UnknownFieldHandler.onWrite : u)(this.typeName, message, writer);
@ -6582,15 +6567,14 @@ exports.CreateCacheEntryResponse = new CreateCacheEntryResponse$Type();
class FinalizeCacheEntryUploadRequest$Type extends runtime_5.MessageType { class FinalizeCacheEntryUploadRequest$Type extends runtime_5.MessageType {
constructor() { constructor() {
super("github.actions.results.api.v1.FinalizeCacheEntryUploadRequest", [ super("github.actions.results.api.v1.FinalizeCacheEntryUploadRequest", [
{ no: 1, name: "workflow_run_backend_id", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 1, name: "metadata", kind: "message", T: () => cachemetadata_1.CacheMetadata },
{ no: 2, name: "workflow_job_run_backend_id", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 2, name: "key", kind: "scalar", T: 9 /*ScalarType.STRING*/ },
{ no: 3, name: "key", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 3, name: "size_bytes", kind: "scalar", T: 3 /*ScalarType.INT64*/ },
{ no: 4, name: "size_bytes", kind: "scalar", T: 3 /*ScalarType.INT64*/ }, { no: 4, name: "version", kind: "scalar", T: 9 /*ScalarType.STRING*/ }
{ no: 5, name: "version", kind: "scalar", T: 9 /*ScalarType.STRING*/ }
]); ]);
} }
create(value) { create(value) {
const message = { workflowRunBackendId: "", workflowJobRunBackendId: "", key: "", sizeBytes: "0", version: "" }; const message = { key: "", sizeBytes: "0", version: "" };
globalThis.Object.defineProperty(message, runtime_4.MESSAGE_TYPE, { enumerable: false, value: this }); globalThis.Object.defineProperty(message, runtime_4.MESSAGE_TYPE, { enumerable: false, value: this });
if (value !== undefined) if (value !== undefined)
(0, runtime_3.reflectionMergePartial)(this, message, value); (0, runtime_3.reflectionMergePartial)(this, message, value);
@ -6601,19 +6585,16 @@ class FinalizeCacheEntryUploadRequest$Type extends runtime_5.MessageType {
while (reader.pos < end) { while (reader.pos < end) {
let [fieldNo, wireType] = reader.tag(); let [fieldNo, wireType] = reader.tag();
switch (fieldNo) { switch (fieldNo) {
case /* string workflow_run_backend_id */ 1: case /* github.actions.results.entities.v1.CacheMetadata metadata */ 1:
message.workflowRunBackendId = reader.string(); message.metadata = cachemetadata_1.CacheMetadata.internalBinaryRead(reader, reader.uint32(), options, message.metadata);
break; break;
case /* string workflow_job_run_backend_id */ 2: case /* string key */ 2:
message.workflowJobRunBackendId = reader.string();
break;
case /* string key */ 3:
message.key = reader.string(); message.key = reader.string();
break; break;
case /* int64 size_bytes */ 4: case /* int64 size_bytes */ 3:
message.sizeBytes = reader.int64().toString(); message.sizeBytes = reader.int64().toString();
break; break;
case /* string version */ 5: case /* string version */ 4:
message.version = reader.string(); message.version = reader.string();
break; break;
default: default:
@ -6628,21 +6609,18 @@ class FinalizeCacheEntryUploadRequest$Type extends runtime_5.MessageType {
return message; return message;
} }
internalBinaryWrite(message, writer, options) { internalBinaryWrite(message, writer, options) {
/* string workflow_run_backend_id = 1; */ /* github.actions.results.entities.v1.CacheMetadata metadata = 1; */
if (message.workflowRunBackendId !== "") if (message.metadata)
writer.tag(1, runtime_1.WireType.LengthDelimited).string(message.workflowRunBackendId); cachemetadata_1.CacheMetadata.internalBinaryWrite(message.metadata, writer.tag(1, runtime_1.WireType.LengthDelimited).fork(), options).join();
/* string workflow_job_run_backend_id = 2; */ /* string key = 2; */
if (message.workflowJobRunBackendId !== "")
writer.tag(2, runtime_1.WireType.LengthDelimited).string(message.workflowJobRunBackendId);
/* string key = 3; */
if (message.key !== "") if (message.key !== "")
writer.tag(3, runtime_1.WireType.LengthDelimited).string(message.key); writer.tag(2, runtime_1.WireType.LengthDelimited).string(message.key);
/* int64 size_bytes = 4; */ /* int64 size_bytes = 3; */
if (message.sizeBytes !== "0") if (message.sizeBytes !== "0")
writer.tag(4, runtime_1.WireType.Varint).int64(message.sizeBytes); writer.tag(3, runtime_1.WireType.Varint).int64(message.sizeBytes);
/* string version = 5; */ /* string version = 4; */
if (message.version !== "") if (message.version !== "")
writer.tag(5, runtime_1.WireType.LengthDelimited).string(message.version); writer.tag(4, runtime_1.WireType.LengthDelimited).string(message.version);
let u = options.writeUnknownFields; let u = options.writeUnknownFields;
if (u !== false) if (u !== false)
(u == true ? runtime_2.UnknownFieldHandler.onWrite : u)(this.typeName, message, writer); (u == true ? runtime_2.UnknownFieldHandler.onWrite : u)(this.typeName, message, writer);
@ -6711,15 +6689,14 @@ exports.FinalizeCacheEntryUploadResponse = new FinalizeCacheEntryUploadResponse$
class GetCacheEntryDownloadURLRequest$Type extends runtime_5.MessageType { class GetCacheEntryDownloadURLRequest$Type extends runtime_5.MessageType {
constructor() { constructor() {
super("github.actions.results.api.v1.GetCacheEntryDownloadURLRequest", [ super("github.actions.results.api.v1.GetCacheEntryDownloadURLRequest", [
{ no: 1, name: "workflow_run_backend_id", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 1, name: "metadata", kind: "message", T: () => cachemetadata_1.CacheMetadata },
{ no: 2, name: "workflow_job_run_backend_id", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 2, name: "key", kind: "scalar", T: 9 /*ScalarType.STRING*/ },
{ no: 3, name: "key", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 3, name: "restore_keys", kind: "scalar", repeat: 2 /*RepeatType.UNPACKED*/, T: 9 /*ScalarType.STRING*/ },
{ no: 4, name: "restore_keys", kind: "scalar", repeat: 2 /*RepeatType.UNPACKED*/, T: 9 /*ScalarType.STRING*/ }, { no: 4, name: "version", kind: "scalar", T: 9 /*ScalarType.STRING*/ }
{ no: 5, name: "version", kind: "scalar", T: 9 /*ScalarType.STRING*/ }
]); ]);
} }
create(value) { create(value) {
const message = { workflowRunBackendId: "", workflowJobRunBackendId: "", key: "", restoreKeys: [], version: "" }; const message = { key: "", restoreKeys: [], version: "" };
globalThis.Object.defineProperty(message, runtime_4.MESSAGE_TYPE, { enumerable: false, value: this }); globalThis.Object.defineProperty(message, runtime_4.MESSAGE_TYPE, { enumerable: false, value: this });
if (value !== undefined) if (value !== undefined)
(0, runtime_3.reflectionMergePartial)(this, message, value); (0, runtime_3.reflectionMergePartial)(this, message, value);
@ -6730,19 +6707,16 @@ class GetCacheEntryDownloadURLRequest$Type extends runtime_5.MessageType {
while (reader.pos < end) { while (reader.pos < end) {
let [fieldNo, wireType] = reader.tag(); let [fieldNo, wireType] = reader.tag();
switch (fieldNo) { switch (fieldNo) {
case /* string workflow_run_backend_id */ 1: case /* github.actions.results.entities.v1.CacheMetadata metadata */ 1:
message.workflowRunBackendId = reader.string(); message.metadata = cachemetadata_1.CacheMetadata.internalBinaryRead(reader, reader.uint32(), options, message.metadata);
break; break;
case /* string workflow_job_run_backend_id */ 2: case /* string key */ 2:
message.workflowJobRunBackendId = reader.string();
break;
case /* string key */ 3:
message.key = reader.string(); message.key = reader.string();
break; break;
case /* repeated string restore_keys */ 4: case /* repeated string restore_keys */ 3:
message.restoreKeys.push(reader.string()); message.restoreKeys.push(reader.string());
break; break;
case /* string version */ 5: case /* string version */ 4:
message.version = reader.string(); message.version = reader.string();
break; break;
default: default:
@ -6757,21 +6731,18 @@ class GetCacheEntryDownloadURLRequest$Type extends runtime_5.MessageType {
return message; return message;
} }
internalBinaryWrite(message, writer, options) { internalBinaryWrite(message, writer, options) {
/* string workflow_run_backend_id = 1; */ /* github.actions.results.entities.v1.CacheMetadata metadata = 1; */
if (message.workflowRunBackendId !== "") if (message.metadata)
writer.tag(1, runtime_1.WireType.LengthDelimited).string(message.workflowRunBackendId); cachemetadata_1.CacheMetadata.internalBinaryWrite(message.metadata, writer.tag(1, runtime_1.WireType.LengthDelimited).fork(), options).join();
/* string workflow_job_run_backend_id = 2; */ /* string key = 2; */
if (message.workflowJobRunBackendId !== "")
writer.tag(2, runtime_1.WireType.LengthDelimited).string(message.workflowJobRunBackendId);
/* string key = 3; */
if (message.key !== "") if (message.key !== "")
writer.tag(3, runtime_1.WireType.LengthDelimited).string(message.key); writer.tag(2, runtime_1.WireType.LengthDelimited).string(message.key);
/* repeated string restore_keys = 4; */ /* repeated string restore_keys = 3; */
for (let i = 0; i < message.restoreKeys.length; i++) for (let i = 0; i < message.restoreKeys.length; i++)
writer.tag(4, runtime_1.WireType.LengthDelimited).string(message.restoreKeys[i]); writer.tag(3, runtime_1.WireType.LengthDelimited).string(message.restoreKeys[i]);
/* string version = 5; */ /* string version = 4; */
if (message.version !== "") if (message.version !== "")
writer.tag(5, runtime_1.WireType.LengthDelimited).string(message.version); writer.tag(4, runtime_1.WireType.LengthDelimited).string(message.version);
let u = options.writeUnknownFields; let u = options.writeUnknownFields;
if (u !== false) if (u !== false)
(u == true ? runtime_2.UnknownFieldHandler.onWrite : u)(this.typeName, message, writer); (u == true ? runtime_2.UnknownFieldHandler.onWrite : u)(this.typeName, message, writer);
@ -6840,13 +6811,12 @@ exports.GetCacheEntryDownloadURLResponse = new GetCacheEntryDownloadURLResponse$
class DeleteCacheEntryRequest$Type extends runtime_5.MessageType { class DeleteCacheEntryRequest$Type extends runtime_5.MessageType {
constructor() { constructor() {
super("github.actions.results.api.v1.DeleteCacheEntryRequest", [ super("github.actions.results.api.v1.DeleteCacheEntryRequest", [
{ no: 1, name: "workflow_run_backend_id", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 1, name: "metadata", kind: "message", T: () => cachemetadata_1.CacheMetadata },
{ no: 2, name: "workflow_job_run_backend_id", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 2, name: "key", kind: "scalar", T: 9 /*ScalarType.STRING*/ }
{ no: 3, name: "key", kind: "scalar", T: 9 /*ScalarType.STRING*/ }
]); ]);
} }
create(value) { create(value) {
const message = { workflowRunBackendId: "", workflowJobRunBackendId: "", key: "" }; const message = { key: "" };
globalThis.Object.defineProperty(message, runtime_4.MESSAGE_TYPE, { enumerable: false, value: this }); globalThis.Object.defineProperty(message, runtime_4.MESSAGE_TYPE, { enumerable: false, value: this });
if (value !== undefined) if (value !== undefined)
(0, runtime_3.reflectionMergePartial)(this, message, value); (0, runtime_3.reflectionMergePartial)(this, message, value);
@ -6857,13 +6827,10 @@ class DeleteCacheEntryRequest$Type extends runtime_5.MessageType {
while (reader.pos < end) { while (reader.pos < end) {
let [fieldNo, wireType] = reader.tag(); let [fieldNo, wireType] = reader.tag();
switch (fieldNo) { switch (fieldNo) {
case /* string workflow_run_backend_id */ 1: case /* github.actions.results.entities.v1.CacheMetadata metadata */ 1:
message.workflowRunBackendId = reader.string(); message.metadata = cachemetadata_1.CacheMetadata.internalBinaryRead(reader, reader.uint32(), options, message.metadata);
break; break;
case /* string workflow_job_run_backend_id */ 2: case /* string key */ 2:
message.workflowJobRunBackendId = reader.string();
break;
case /* string key */ 3:
message.key = reader.string(); message.key = reader.string();
break; break;
default: default:
@ -6878,15 +6845,12 @@ class DeleteCacheEntryRequest$Type extends runtime_5.MessageType {
return message; return message;
} }
internalBinaryWrite(message, writer, options) { internalBinaryWrite(message, writer, options) {
/* string workflow_run_backend_id = 1; */ /* github.actions.results.entities.v1.CacheMetadata metadata = 1; */
if (message.workflowRunBackendId !== "") if (message.metadata)
writer.tag(1, runtime_1.WireType.LengthDelimited).string(message.workflowRunBackendId); cachemetadata_1.CacheMetadata.internalBinaryWrite(message.metadata, writer.tag(1, runtime_1.WireType.LengthDelimited).fork(), options).join();
/* string workflow_job_run_backend_id = 2; */ /* string key = 2; */
if (message.workflowJobRunBackendId !== "")
writer.tag(2, runtime_1.WireType.LengthDelimited).string(message.workflowJobRunBackendId);
/* string key = 3; */
if (message.key !== "") if (message.key !== "")
writer.tag(3, runtime_1.WireType.LengthDelimited).string(message.key); writer.tag(2, runtime_1.WireType.LengthDelimited).string(message.key);
let u = options.writeUnknownFields; let u = options.writeUnknownFields;
if (u !== false) if (u !== false)
(u == true ? runtime_2.UnknownFieldHandler.onWrite : u)(this.typeName, message, writer); (u == true ? runtime_2.UnknownFieldHandler.onWrite : u)(this.typeName, message, writer);
@ -6955,14 +6919,13 @@ exports.DeleteCacheEntryResponse = new DeleteCacheEntryResponse$Type();
class ListCacheEntriesRequest$Type extends runtime_5.MessageType { class ListCacheEntriesRequest$Type extends runtime_5.MessageType {
constructor() { constructor() {
super("github.actions.results.api.v1.ListCacheEntriesRequest", [ super("github.actions.results.api.v1.ListCacheEntriesRequest", [
{ no: 1, name: "workflow_run_backend_id", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 1, name: "metadata", kind: "message", T: () => cachemetadata_1.CacheMetadata },
{ no: 2, name: "workflow_job_run_backend_id", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 2, name: "key", kind: "scalar", T: 9 /*ScalarType.STRING*/ },
{ no: 3, name: "key", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 3, name: "restore_keys", kind: "scalar", repeat: 2 /*RepeatType.UNPACKED*/, T: 9 /*ScalarType.STRING*/ }
{ no: 4, name: "restore_keys", kind: "scalar", repeat: 2 /*RepeatType.UNPACKED*/, T: 9 /*ScalarType.STRING*/ }
]); ]);
} }
create(value) { create(value) {
const message = { workflowRunBackendId: "", workflowJobRunBackendId: "", key: "", restoreKeys: [] }; const message = { key: "", restoreKeys: [] };
globalThis.Object.defineProperty(message, runtime_4.MESSAGE_TYPE, { enumerable: false, value: this }); globalThis.Object.defineProperty(message, runtime_4.MESSAGE_TYPE, { enumerable: false, value: this });
if (value !== undefined) if (value !== undefined)
(0, runtime_3.reflectionMergePartial)(this, message, value); (0, runtime_3.reflectionMergePartial)(this, message, value);
@ -6973,16 +6936,13 @@ class ListCacheEntriesRequest$Type extends runtime_5.MessageType {
while (reader.pos < end) { while (reader.pos < end) {
let [fieldNo, wireType] = reader.tag(); let [fieldNo, wireType] = reader.tag();
switch (fieldNo) { switch (fieldNo) {
case /* string workflow_run_backend_id */ 1: case /* github.actions.results.entities.v1.CacheMetadata metadata */ 1:
message.workflowRunBackendId = reader.string(); message.metadata = cachemetadata_1.CacheMetadata.internalBinaryRead(reader, reader.uint32(), options, message.metadata);
break; break;
case /* string workflow_job_run_backend_id */ 2: case /* string key */ 2:
message.workflowJobRunBackendId = reader.string();
break;
case /* string key */ 3:
message.key = reader.string(); message.key = reader.string();
break; break;
case /* repeated string restore_keys */ 4: case /* repeated string restore_keys */ 3:
message.restoreKeys.push(reader.string()); message.restoreKeys.push(reader.string());
break; break;
default: default:
@ -6997,18 +6957,15 @@ class ListCacheEntriesRequest$Type extends runtime_5.MessageType {
return message; return message;
} }
internalBinaryWrite(message, writer, options) { internalBinaryWrite(message, writer, options) {
/* string workflow_run_backend_id = 1; */ /* github.actions.results.entities.v1.CacheMetadata metadata = 1; */
if (message.workflowRunBackendId !== "") if (message.metadata)
writer.tag(1, runtime_1.WireType.LengthDelimited).string(message.workflowRunBackendId); cachemetadata_1.CacheMetadata.internalBinaryWrite(message.metadata, writer.tag(1, runtime_1.WireType.LengthDelimited).fork(), options).join();
/* string workflow_job_run_backend_id = 2; */ /* string key = 2; */
if (message.workflowJobRunBackendId !== "")
writer.tag(2, runtime_1.WireType.LengthDelimited).string(message.workflowJobRunBackendId);
/* string key = 3; */
if (message.key !== "") if (message.key !== "")
writer.tag(3, runtime_1.WireType.LengthDelimited).string(message.key); writer.tag(2, runtime_1.WireType.LengthDelimited).string(message.key);
/* repeated string restore_keys = 4; */ /* repeated string restore_keys = 3; */
for (let i = 0; i < message.restoreKeys.length; i++) for (let i = 0; i < message.restoreKeys.length; i++)
writer.tag(4, runtime_1.WireType.LengthDelimited).string(message.restoreKeys[i]); writer.tag(3, runtime_1.WireType.LengthDelimited).string(message.restoreKeys[i]);
let u = options.writeUnknownFields; let u = options.writeUnknownFields;
if (u !== false) if (u !== false)
(u == true ? runtime_2.UnknownFieldHandler.onWrite : u)(this.typeName, message, writer); (u == true ? runtime_2.UnknownFieldHandler.onWrite : u)(this.typeName, message, writer);
@ -7166,15 +7123,14 @@ exports.ListCacheEntriesResponse_CacheEntry = new ListCacheEntriesResponse_Cache
class LookupCacheEntryRequest$Type extends runtime_5.MessageType { class LookupCacheEntryRequest$Type extends runtime_5.MessageType {
constructor() { constructor() {
super("github.actions.results.api.v1.LookupCacheEntryRequest", [ super("github.actions.results.api.v1.LookupCacheEntryRequest", [
{ no: 1, name: "workflow_run_backend_id", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 1, name: "metadata", kind: "message", T: () => cachemetadata_1.CacheMetadata },
{ no: 2, name: "workflow_job_run_backend_id", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 2, name: "key", kind: "scalar", T: 9 /*ScalarType.STRING*/ },
{ no: 3, name: "key", kind: "scalar", T: 9 /*ScalarType.STRING*/ }, { no: 3, name: "restore_keys", kind: "scalar", repeat: 2 /*RepeatType.UNPACKED*/, T: 9 /*ScalarType.STRING*/ },
{ no: 4, name: "restore_keys", kind: "scalar", repeat: 2 /*RepeatType.UNPACKED*/, T: 9 /*ScalarType.STRING*/ }, { no: 4, name: "version", kind: "scalar", T: 9 /*ScalarType.STRING*/ }
{ no: 5, name: "version", kind: "scalar", T: 9 /*ScalarType.STRING*/ }
]); ]);
} }
create(value) { create(value) {
const message = { workflowRunBackendId: "", workflowJobRunBackendId: "", key: "", restoreKeys: [], version: "" }; const message = { key: "", restoreKeys: [], version: "" };
globalThis.Object.defineProperty(message, runtime_4.MESSAGE_TYPE, { enumerable: false, value: this }); globalThis.Object.defineProperty(message, runtime_4.MESSAGE_TYPE, { enumerable: false, value: this });
if (value !== undefined) if (value !== undefined)
(0, runtime_3.reflectionMergePartial)(this, message, value); (0, runtime_3.reflectionMergePartial)(this, message, value);
@ -7185,19 +7141,16 @@ class LookupCacheEntryRequest$Type extends runtime_5.MessageType {
while (reader.pos < end) { while (reader.pos < end) {
let [fieldNo, wireType] = reader.tag(); let [fieldNo, wireType] = reader.tag();
switch (fieldNo) { switch (fieldNo) {
case /* string workflow_run_backend_id */ 1: case /* github.actions.results.entities.v1.CacheMetadata metadata */ 1:
message.workflowRunBackendId = reader.string(); message.metadata = cachemetadata_1.CacheMetadata.internalBinaryRead(reader, reader.uint32(), options, message.metadata);
break; break;
case /* string workflow_job_run_backend_id */ 2: case /* string key */ 2:
message.workflowJobRunBackendId = reader.string();
break;
case /* string key */ 3:
message.key = reader.string(); message.key = reader.string();
break; break;
case /* repeated string restore_keys */ 4: case /* repeated string restore_keys */ 3:
message.restoreKeys.push(reader.string()); message.restoreKeys.push(reader.string());
break; break;
case /* string version */ 5: case /* string version */ 4:
message.version = reader.string(); message.version = reader.string();
break; break;
default: default:
@ -7212,21 +7165,18 @@ class LookupCacheEntryRequest$Type extends runtime_5.MessageType {
return message; return message;
} }
internalBinaryWrite(message, writer, options) { internalBinaryWrite(message, writer, options) {
/* string workflow_run_backend_id = 1; */ /* github.actions.results.entities.v1.CacheMetadata metadata = 1; */
if (message.workflowRunBackendId !== "") if (message.metadata)
writer.tag(1, runtime_1.WireType.LengthDelimited).string(message.workflowRunBackendId); cachemetadata_1.CacheMetadata.internalBinaryWrite(message.metadata, writer.tag(1, runtime_1.WireType.LengthDelimited).fork(), options).join();
/* string workflow_job_run_backend_id = 2; */ /* string key = 2; */
if (message.workflowJobRunBackendId !== "")
writer.tag(2, runtime_1.WireType.LengthDelimited).string(message.workflowJobRunBackendId);
/* string key = 3; */
if (message.key !== "") if (message.key !== "")
writer.tag(3, runtime_1.WireType.LengthDelimited).string(message.key); writer.tag(2, runtime_1.WireType.LengthDelimited).string(message.key);
/* repeated string restore_keys = 4; */ /* repeated string restore_keys = 3; */
for (let i = 0; i < message.restoreKeys.length; i++) for (let i = 0; i < message.restoreKeys.length; i++)
writer.tag(4, runtime_1.WireType.LengthDelimited).string(message.restoreKeys[i]); writer.tag(3, runtime_1.WireType.LengthDelimited).string(message.restoreKeys[i]);
/* string version = 5; */ /* string version = 4; */
if (message.version !== "") if (message.version !== "")
writer.tag(5, runtime_1.WireType.LengthDelimited).string(message.version); writer.tag(4, runtime_1.WireType.LengthDelimited).string(message.version);
let u = options.writeUnknownFields; let u = options.writeUnknownFields;
if (u !== false) if (u !== false)
(u == true ? runtime_2.UnknownFieldHandler.onWrite : u)(this.typeName, message, writer); (u == true ? runtime_2.UnknownFieldHandler.onWrite : u)(this.typeName, message, writer);
@ -8004,6 +7954,147 @@ function handleCacheServiceLookupCacheEntryProtobuf(ctx, service, data, intercep
/***/ }), /***/ }),
/***/ 2773:
/***/ ((__unused_webpack_module, exports, __nccwpck_require__) => {
"use strict";
Object.defineProperty(exports, "__esModule", ({ value: true }));
exports.CacheMetadata = void 0;
const runtime_1 = __nccwpck_require__(3503);
const runtime_2 = __nccwpck_require__(3503);
const runtime_3 = __nccwpck_require__(3503);
const runtime_4 = __nccwpck_require__(3503);
const runtime_5 = __nccwpck_require__(3503);
const cachescope_1 = __nccwpck_require__(7685);
// @generated message type with reflection information, may provide speed optimized methods
class CacheMetadata$Type extends runtime_5.MessageType {
constructor() {
super("github.actions.results.entities.v1.CacheMetadata", [
{ no: 1, name: "repository_id", kind: "scalar", T: 3 /*ScalarType.INT64*/ },
{ no: 2, name: "scope", kind: "message", repeat: 1 /*RepeatType.PACKED*/, T: () => cachescope_1.CacheScope }
]);
}
create(value) {
const message = { repositoryId: "0", scope: [] };
globalThis.Object.defineProperty(message, runtime_4.MESSAGE_TYPE, { enumerable: false, value: this });
if (value !== undefined)
(0, runtime_3.reflectionMergePartial)(this, message, value);
return message;
}
internalBinaryRead(reader, length, options, target) {
let message = target !== null && target !== void 0 ? target : this.create(), end = reader.pos + length;
while (reader.pos < end) {
let [fieldNo, wireType] = reader.tag();
switch (fieldNo) {
case /* int64 repository_id */ 1:
message.repositoryId = reader.int64().toString();
break;
case /* repeated github.actions.results.entities.v1.CacheScope scope */ 2:
message.scope.push(cachescope_1.CacheScope.internalBinaryRead(reader, reader.uint32(), options));
break;
default:
let u = options.readUnknownField;
if (u === "throw")
throw new globalThis.Error(`Unknown field ${fieldNo} (wire type ${wireType}) for ${this.typeName}`);
let d = reader.skip(wireType);
if (u !== false)
(u === true ? runtime_2.UnknownFieldHandler.onRead : u)(this.typeName, message, fieldNo, wireType, d);
}
}
return message;
}
internalBinaryWrite(message, writer, options) {
/* int64 repository_id = 1; */
if (message.repositoryId !== "0")
writer.tag(1, runtime_1.WireType.Varint).int64(message.repositoryId);
/* repeated github.actions.results.entities.v1.CacheScope scope = 2; */
for (let i = 0; i < message.scope.length; i++)
cachescope_1.CacheScope.internalBinaryWrite(message.scope[i], writer.tag(2, runtime_1.WireType.LengthDelimited).fork(), options).join();
let u = options.writeUnknownFields;
if (u !== false)
(u == true ? runtime_2.UnknownFieldHandler.onWrite : u)(this.typeName, message, writer);
return writer;
}
}
/**
* @generated MessageType for protobuf message github.actions.results.entities.v1.CacheMetadata
*/
exports.CacheMetadata = new CacheMetadata$Type();
//# sourceMappingURL=cachemetadata.js.map
/***/ }),
/***/ 7685:
/***/ ((__unused_webpack_module, exports, __nccwpck_require__) => {
"use strict";
Object.defineProperty(exports, "__esModule", ({ value: true }));
exports.CacheScope = void 0;
const runtime_1 = __nccwpck_require__(3503);
const runtime_2 = __nccwpck_require__(3503);
const runtime_3 = __nccwpck_require__(3503);
const runtime_4 = __nccwpck_require__(3503);
const runtime_5 = __nccwpck_require__(3503);
// @generated message type with reflection information, may provide speed optimized methods
class CacheScope$Type extends runtime_5.MessageType {
constructor() {
super("github.actions.results.entities.v1.CacheScope", [
{ no: 1, name: "scope", kind: "scalar", T: 9 /*ScalarType.STRING*/ },
{ no: 2, name: "permission", kind: "scalar", T: 3 /*ScalarType.INT64*/ }
]);
}
create(value) {
const message = { scope: "", permission: "0" };
globalThis.Object.defineProperty(message, runtime_4.MESSAGE_TYPE, { enumerable: false, value: this });
if (value !== undefined)
(0, runtime_3.reflectionMergePartial)(this, message, value);
return message;
}
internalBinaryRead(reader, length, options, target) {
let message = target !== null && target !== void 0 ? target : this.create(), end = reader.pos + length;
while (reader.pos < end) {
let [fieldNo, wireType] = reader.tag();
switch (fieldNo) {
case /* string scope */ 1:
message.scope = reader.string();
break;
case /* int64 permission */ 2:
message.permission = reader.int64().toString();
break;
default:
let u = options.readUnknownField;
if (u === "throw")
throw new globalThis.Error(`Unknown field ${fieldNo} (wire type ${wireType}) for ${this.typeName}`);
let d = reader.skip(wireType);
if (u !== false)
(u === true ? runtime_2.UnknownFieldHandler.onRead : u)(this.typeName, message, fieldNo, wireType, d);
}
}
return message;
}
internalBinaryWrite(message, writer, options) {
/* string scope = 1; */
if (message.scope !== "")
writer.tag(1, runtime_1.WireType.LengthDelimited).string(message.scope);
/* int64 permission = 2; */
if (message.permission !== "0")
writer.tag(2, runtime_1.WireType.Varint).int64(message.permission);
let u = options.writeUnknownFields;
if (u !== false)
(u == true ? runtime_2.UnknownFieldHandler.onWrite : u)(this.typeName, message, writer);
return writer;
}
}
/**
* @generated MessageType for protobuf message github.actions.results.entities.v1.CacheScope
*/
exports.CacheScope = new CacheScope$Type();
//# sourceMappingURL=cachescope.js.map
/***/ }),
/***/ 4076: /***/ 4076:
/***/ (function(__unused_webpack_module, exports, __nccwpck_require__) { /***/ (function(__unused_webpack_module, exports, __nccwpck_require__) {
@ -8416,11 +8507,8 @@ var __asyncValues = (this && this.__asyncValues) || function (o) {
function verb(n) { i[n] = o[n] && function (v) { return new Promise(function (resolve, reject) { v = o[n](v), settle(resolve, reject, v.done, v.value); }); }; } function verb(n) { i[n] = o[n] && function (v) { return new Promise(function (resolve, reject) { v = o[n](v), settle(resolve, reject, v.done, v.value); }); }; }
function settle(resolve, reject, d, v) { Promise.resolve(v).then(function(v) { resolve({ value: v, done: d }); }, reject); } function settle(resolve, reject, d, v) { Promise.resolve(v).then(function(v) { resolve({ value: v, done: d }); }, reject); }
}; };
var __importDefault = (this && this.__importDefault) || function (mod) {
return (mod && mod.__esModule) ? mod : { "default": mod };
};
Object.defineProperty(exports, "__esModule", ({ value: true })); Object.defineProperty(exports, "__esModule", ({ value: true }));
exports.getBackendIdsFromToken = exports.getRuntimeToken = exports.getCacheVersion = exports.isGhes = exports.assertDefined = exports.getGnuTarPathOnWindows = exports.getCacheFileName = exports.getCompressionMethod = exports.unlinkFile = exports.resolvePaths = exports.getArchiveFileSizeInBytes = exports.createTempDirectory = void 0; exports.getRuntimeToken = exports.getCacheVersion = exports.isGhes = exports.assertDefined = exports.getGnuTarPathOnWindows = exports.getCacheFileName = exports.getCompressionMethod = exports.unlinkFile = exports.resolvePaths = exports.getArchiveFileSizeInBytes = exports.createTempDirectory = void 0;
const core = __importStar(__nccwpck_require__(4850)); const core = __importStar(__nccwpck_require__(4850));
const exec = __importStar(__nccwpck_require__(309)); const exec = __importStar(__nccwpck_require__(309));
const glob = __importStar(__nccwpck_require__(9590)); const glob = __importStar(__nccwpck_require__(9590));
@ -8430,7 +8518,6 @@ const fs = __importStar(__nccwpck_require__(7147));
const path = __importStar(__nccwpck_require__(1017)); const path = __importStar(__nccwpck_require__(1017));
const semver = __importStar(__nccwpck_require__(3910)); const semver = __importStar(__nccwpck_require__(3910));
const util = __importStar(__nccwpck_require__(3837)); const util = __importStar(__nccwpck_require__(3837));
const jwt_decode_1 = __importDefault(__nccwpck_require__(1300));
const constants_1 = __nccwpck_require__(4010); const constants_1 = __nccwpck_require__(4010);
const versionSalt = '1.0'; const versionSalt = '1.0';
// From https://github.com/actions/toolkit/blob/main/packages/tool-cache/src/tool-cache.ts#L23 // From https://github.com/actions/toolkit/blob/main/packages/tool-cache/src/tool-cache.ts#L23
@ -8603,54 +8690,6 @@ function getRuntimeToken() {
return token; return token;
} }
exports.getRuntimeToken = getRuntimeToken; exports.getRuntimeToken = getRuntimeToken;
const InvalidJwtError = new Error('Failed to get backend IDs: The provided JWT token is invalid and/or missing claims');
// uses the JWT token claims to get the
// workflow run and workflow job run backend ids
function getBackendIdsFromToken() {
const token = getRuntimeToken();
const decoded = (0, jwt_decode_1.default)(token);
if (!decoded.scp) {
throw InvalidJwtError;
}
/*
* example decoded:
* {
* scp: "Actions.ExampleScope Actions.Results:ce7f54c7-61c7-4aae-887f-30da475f5f1a:ca395085-040a-526b-2ce8-bdc85f692774"
* }
*/
const scpParts = decoded.scp.split(' ');
if (scpParts.length === 0) {
throw InvalidJwtError;
}
/*
* example scpParts:
* ["Actions.ExampleScope", "Actions.Results:ce7f54c7-61c7-4aae-887f-30da475f5f1a:ca395085-040a-526b-2ce8-bdc85f692774"]
*/
for (const scopes of scpParts) {
const scopeParts = scopes.split(':');
if ((scopeParts === null || scopeParts === void 0 ? void 0 : scopeParts[0]) !== 'Actions.Results') {
// not the Actions.Results scope
continue;
}
/*
* example scopeParts:
* ["Actions.Results", "ce7f54c7-61c7-4aae-887f-30da475f5f1a", "ca395085-040a-526b-2ce8-bdc85f692774"]
*/
if (scopeParts.length !== 3) {
// missing expected number of claims
throw InvalidJwtError;
}
const ids = {
workflowRunBackendId: scopeParts[1],
workflowJobRunBackendId: scopeParts[2]
};
core.debug(`Workflow Run Backend ID: ${ids.workflowRunBackendId}`);
core.debug(`Workflow Job Run Backend ID: ${ids.workflowJobRunBackendId}`);
return ids;
}
throw InvalidJwtError;
}
exports.getBackendIdsFromToken = getBackendIdsFromToken;
//# sourceMappingURL=cacheUtils.js.map //# sourceMappingURL=cacheUtils.js.map
/***/ }), /***/ }),
@ -8661,15 +8700,7 @@ exports.getBackendIdsFromToken = getBackendIdsFromToken;
"use strict"; "use strict";
Object.defineProperty(exports, "__esModule", ({ value: true })); Object.defineProperty(exports, "__esModule", ({ value: true }));
exports.getCacheServiceURL = exports.getCacheServiceVersion = exports.getRuntimeToken = void 0; exports.getCacheServiceURL = exports.getCacheServiceVersion = void 0;
function getRuntimeToken() {
const token = process.env['ACTIONS_RUNTIME_TOKEN'];
if (!token) {
throw new Error('Unable to get the ACTIONS_RUNTIME_TOKEN env variable');
}
return token;
}
exports.getRuntimeToken = getRuntimeToken;
function getCacheServiceVersion() { function getCacheServiceVersion() {
return process.env['ACTIONS_CACHE_SERVICE_V2'] ? 'v2' : 'v1'; return process.env['ACTIONS_CACHE_SERVICE_V2'] ? 'v2' : 'v1';
} }
@ -9285,6 +9316,7 @@ const core_1 = __nccwpck_require__(4850);
const user_agent_1 = __nccwpck_require__(5736); const user_agent_1 = __nccwpck_require__(5736);
const errors_1 = __nccwpck_require__(6333); const errors_1 = __nccwpck_require__(6333);
const config_1 = __nccwpck_require__(6490); const config_1 = __nccwpck_require__(6490);
const cacheUtils_1 = __nccwpck_require__(3310);
const auth_1 = __nccwpck_require__(7231); const auth_1 = __nccwpck_require__(7231);
const http_client_1 = __nccwpck_require__(1283); const http_client_1 = __nccwpck_require__(1283);
const cache_twirp_1 = __nccwpck_require__(894); const cache_twirp_1 = __nccwpck_require__(894);
@ -9300,7 +9332,7 @@ class CacheServiceClient {
this.maxAttempts = 5; this.maxAttempts = 5;
this.baseRetryIntervalMilliseconds = 3000; this.baseRetryIntervalMilliseconds = 3000;
this.retryMultiplier = 1.5; this.retryMultiplier = 1.5;
const token = (0, config_1.getRuntimeToken)(); const token = (0, cacheUtils_1.getRuntimeToken)();
this.baseUrl = (0, config_1.getCacheServiceURL)(); this.baseUrl = (0, config_1.getCacheServiceURL)();
if (maxAttempts) { if (maxAttempts) {
this.maxAttempts = maxAttempts; this.maxAttempts = maxAttempts;
@ -57677,16 +57709,6 @@ DotObject._process = _process
module.exports = DotObject module.exports = DotObject
/***/ }),
/***/ 1300:
/***/ ((module) => {
"use strict";
function e(e){this.message=e}e.prototype=new Error,e.prototype.name="InvalidCharacterError";var r="undefined"!=typeof window&&window.atob&&window.atob.bind(window)||function(r){var t=String(r).replace(/=+$/,"");if(t.length%4==1)throw new e("'atob' failed: The string to be decoded is not correctly encoded.");for(var n,o,a=0,i=0,c="";o=t.charAt(i++);~o&&(n=a%4?64*n+o:o,a++%4)?c+=String.fromCharCode(255&n>>(-2*a&6)):0)o="ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/=".indexOf(o);return c};function t(e){var t=e.replace(/-/g,"+").replace(/_/g,"/");switch(t.length%4){case 0:break;case 2:t+="==";break;case 3:t+="=";break;default:throw"Illegal base64url string!"}try{return function(e){return decodeURIComponent(r(e).replace(/(.)/g,(function(e,r){var t=r.charCodeAt(0).toString(16).toUpperCase();return t.length<2&&(t="0"+t),"%"+t})))}(t)}catch(e){return r(t)}}function n(e){this.message=e}function o(e,r){if("string"!=typeof e)throw new n("Invalid token specified");var o=!0===(r=r||{}).header?0:1;try{return JSON.parse(t(e.split(".")[o]))}catch(e){throw new n("Invalid token specified: "+e.message)}}n.prototype=new Error,n.prototype.name="InvalidTokenError";const a=o;a.default=o,a.InvalidTokenError=n,module.exports=a;
//# sourceMappingURL=jwt-decode.cjs.js.map
/***/ }), /***/ }),
/***/ 2020: /***/ 2020:
@ -73498,7 +73520,7 @@ module.exports = JSON.parse('[[[0,44],"disallowed_STD3_valid"],[[45,46],"valid"]
/***/ ((module) => { /***/ ((module) => {
"use strict"; "use strict";
module.exports = JSON.parse('{"name":"@actions/cache","version":"3.3.0","preview":true,"description":"Actions cache lib","keywords":["github","actions","cache"],"homepage":"https://github.com/actions/toolkit/tree/main/packages/cache","license":"MIT","main":"lib/cache.js","types":"lib/cache.d.ts","directories":{"lib":"lib","test":"__tests__"},"files":["lib","!.DS_Store"],"publishConfig":{"access":"public"},"repository":{"type":"git","url":"git+https://github.com/actions/toolkit.git","directory":"packages/cache"},"scripts":{"audit-moderate":"npm install && npm audit --json --audit-level=moderate > audit.json","test":"echo \\"Error: run tests from root\\" && exit 1","tsc":"tsc"},"bugs":{"url":"https://github.com/actions/toolkit/issues"},"dependencies":{"@actions/core":"^1.11.1","@actions/exec":"^1.0.1","@actions/glob":"^0.1.0","@actions/http-client":"^2.1.1","@actions/io":"^1.0.1","@azure/abort-controller":"^1.1.0","@azure/ms-rest-js":"^2.6.0","@azure/storage-blob":"^12.13.0","@protobuf-ts/plugin":"^2.9.4","semver":"^6.3.1","jwt-decode":"^3.1.2","twirp-ts":"^2.5.0"},"devDependencies":{"@types/semver":"^6.0.0","typescript":"^5.2.2"}}'); module.exports = JSON.parse('{"name":"@actions/cache","version":"3.3.0","preview":true,"description":"Actions cache lib","keywords":["github","actions","cache"],"homepage":"https://github.com/actions/toolkit/tree/main/packages/cache","license":"MIT","main":"lib/cache.js","types":"lib/cache.d.ts","directories":{"lib":"lib","test":"__tests__"},"files":["lib","!.DS_Store"],"publishConfig":{"access":"public"},"repository":{"type":"git","url":"git+https://github.com/actions/toolkit.git","directory":"packages/cache"},"scripts":{"audit-moderate":"npm install && npm audit --json --audit-level=moderate > audit.json","test":"echo \\"Error: run tests from root\\" && exit 1","tsc":"tsc"},"bugs":{"url":"https://github.com/actions/toolkit/issues"},"dependencies":{"@actions/core":"^1.11.1","@actions/exec":"^1.0.1","@actions/glob":"^0.1.0","@actions/http-client":"^2.1.1","@actions/io":"^1.0.1","@azure/abort-controller":"^1.1.0","@azure/ms-rest-js":"^2.6.0","@azure/storage-blob":"^12.13.0","@protobuf-ts/plugin":"^2.9.4","semver":"^6.3.1","twirp-ts":"^2.5.0"},"devDependencies":{"@types/semver":"^6.0.0","typescript":"^5.2.2"}}');
/***/ }) /***/ })

2
package-lock.json generated
View File

@ -49,7 +49,6 @@
"@azure/ms-rest-js": "^2.6.0", "@azure/ms-rest-js": "^2.6.0",
"@azure/storage-blob": "^12.13.0", "@azure/storage-blob": "^12.13.0",
"@protobuf-ts/plugin": "^2.9.4", "@protobuf-ts/plugin": "^2.9.4",
"jwt-decode": "^3.1.2",
"semver": "^6.3.1", "semver": "^6.3.1",
"twirp-ts": "^2.5.0" "twirp-ts": "^2.5.0"
}, },
@ -9141,7 +9140,6 @@
"@azure/storage-blob": "^12.13.0", "@azure/storage-blob": "^12.13.0",
"@protobuf-ts/plugin": "^2.9.4", "@protobuf-ts/plugin": "^2.9.4",
"@types/semver": "^6.0.0", "@types/semver": "^6.0.0",
"jwt-decode": "^3.1.2",
"semver": "^6.3.1", "semver": "^6.3.1",
"twirp-ts": "^2.5.0", "twirp-ts": "^2.5.0",
"typescript": "^5.2.2" "typescript": "^5.2.2"