mirror of
https://code.forgejo.org/actions/checkout
synced 2024-11-25 10:30:39 +00:00
.
This commit is contained in:
parent
093dbebc2e
commit
59bb2dd7cd
2 changed files with 6 additions and 59 deletions
29
dist/index.js
vendored
29
dist/index.js
vendored
|
@ -8376,12 +8376,12 @@ function downloadRepository(accessToken, owner, repo, ref, commit, repositoryPat
|
||||||
// Download the archive
|
// Download the archive
|
||||||
let archiveData = yield retryHelper.execute(() => __awaiter(this, void 0, void 0, function* () {
|
let archiveData = yield retryHelper.execute(() => __awaiter(this, void 0, void 0, function* () {
|
||||||
core.info('Downloading the archive using the REST API');
|
core.info('Downloading the archive using the REST API');
|
||||||
yield yield downloadArchive(accessToken, owner, repo, ref, commit);
|
return yield downloadArchive(accessToken, owner, repo, ref, commit);
|
||||||
}));
|
}));
|
||||||
// Write archive to disk
|
// Write archive to disk
|
||||||
core.info('Writing archive to disk');
|
core.info('Writing archive to disk');
|
||||||
yield fs.promises.writeFile(archivePath, archiveData);
|
yield fs.promises.writeFile(archivePath, archiveData);
|
||||||
archiveData = undefined;
|
archiveData = Buffer.from(''); // Free memory
|
||||||
// // Get the archive URL using the REST API
|
// // Get the archive URL using the REST API
|
||||||
// await retryHelper.execute(async () => {
|
// await retryHelper.execute(async () => {
|
||||||
// // Prepare the archive stream
|
// // Prepare the archive stream
|
||||||
|
@ -8407,28 +8407,12 @@ function downloadRepository(accessToken, owner, repo, ref, commit, repositoryPat
|
||||||
// await fileStreamClosed
|
// await fileStreamClosed
|
||||||
// }
|
// }
|
||||||
// })
|
// })
|
||||||
// return Buffer.from(response.data) // response.data is ArrayBuffer
|
|
||||||
// // Download the archive
|
|
||||||
// core.info('Downloading the archive') // Do not print the URL since it contains a token to download the archive
|
|
||||||
// await downloadFile(archiveUrl, archivePath)
|
|
||||||
// // console.log(`status=${response.status}`)
|
|
||||||
// // console.log(`headers=${JSON.stringify(response.headers)}`)
|
|
||||||
// // console.log(`data=${response.data}`)
|
|
||||||
// // console.log(`data=${JSON.stringify(response.data)}`)
|
|
||||||
// // for (const key of Object.keys(response.data)) {
|
|
||||||
// // console.log(`data['${key}']=${response.data[key]}`)
|
|
||||||
// // }
|
|
||||||
// // Write archive to file
|
|
||||||
// const runnerTemp = process.env['RUNNER_TEMP'] as string
|
|
||||||
// assert.ok(runnerTemp, 'RUNNER_TEMP not defined')
|
|
||||||
// const archivePath = path.join(runnerTemp, 'checkout.tar.gz')
|
|
||||||
// await io.rmRF(archivePath)
|
|
||||||
// await fs.promises.writeFile(archivePath, raw)
|
// await fs.promises.writeFile(archivePath, raw)
|
||||||
// // await exec.exec(`ls -la "${archiveFile}"`, [], {
|
// // await exec.exec(`ls -la "${archiveFile}"`, [], {
|
||||||
// // cwd: repositoryPath
|
// // cwd: repositoryPath
|
||||||
// // } as ExecOptions)
|
// // } as ExecOptions)
|
||||||
// Extract archive
|
// Extract archive
|
||||||
const extractPath = path.join(runnerTemp, `checkout-archive${IS_WINDOWS ? '.zip' : '.tar.gz'}`);
|
const extractPath = path.join(runnerTemp, `checkout`);
|
||||||
yield io.rmRF(extractPath);
|
yield io.rmRF(extractPath);
|
||||||
yield io.mkdirP(extractPath);
|
yield io.mkdirP(extractPath);
|
||||||
if (IS_WINDOWS) {
|
if (IS_WINDOWS) {
|
||||||
|
@ -8476,13 +8460,6 @@ function downloadArchive(accessToken, owner, repo, ref, commit) {
|
||||||
throw new Error(`Unexpected response from GitHub API. Status: '${response.status}'`);
|
throw new Error(`Unexpected response from GitHub API. Status: '${response.status}'`);
|
||||||
}
|
}
|
||||||
return Buffer.from(response.data); // response.data is ArrayBuffer
|
return Buffer.from(response.data); // response.data is ArrayBuffer
|
||||||
// console.log('GETTING THE LOCATION')
|
|
||||||
// const archiveUrl = response.headers['Location'] // Do not print the archive URL because it has an embedded token
|
|
||||||
// assert.ok(
|
|
||||||
// archiveUrl,
|
|
||||||
// `Expected GitHub API response to contain 'Location' header`
|
|
||||||
// )
|
|
||||||
// return archiveUrl
|
|
||||||
});
|
});
|
||||||
}
|
}
|
||||||
// async function getArchiveUrl(
|
// async function getArchiveUrl(
|
||||||
|
|
|
@ -36,13 +36,13 @@ export async function downloadRepository(
|
||||||
// Download the archive
|
// Download the archive
|
||||||
let archiveData = await retryHelper.execute(async () => {
|
let archiveData = await retryHelper.execute(async () => {
|
||||||
core.info('Downloading the archive using the REST API')
|
core.info('Downloading the archive using the REST API')
|
||||||
await await downloadArchive(accessToken, owner, repo, ref, commit)
|
return await downloadArchive(accessToken, owner, repo, ref, commit)
|
||||||
})
|
})
|
||||||
|
|
||||||
// Write archive to disk
|
// Write archive to disk
|
||||||
core.info('Writing archive to disk')
|
core.info('Writing archive to disk')
|
||||||
await fs.promises.writeFile(archivePath, archiveData)
|
await fs.promises.writeFile(archivePath, archiveData)
|
||||||
archiveData = undefined
|
archiveData = Buffer.from('') // Free memory
|
||||||
|
|
||||||
// // Get the archive URL using the REST API
|
// // Get the archive URL using the REST API
|
||||||
// await retryHelper.execute(async () => {
|
// await retryHelper.execute(async () => {
|
||||||
|
@ -72,35 +72,13 @@ export async function downloadRepository(
|
||||||
// }
|
// }
|
||||||
// })
|
// })
|
||||||
|
|
||||||
// return Buffer.from(response.data) // response.data is ArrayBuffer
|
|
||||||
|
|
||||||
// // Download the archive
|
|
||||||
// core.info('Downloading the archive') // Do not print the URL since it contains a token to download the archive
|
|
||||||
// await downloadFile(archiveUrl, archivePath)
|
|
||||||
|
|
||||||
// // console.log(`status=${response.status}`)
|
|
||||||
// // console.log(`headers=${JSON.stringify(response.headers)}`)
|
|
||||||
// // console.log(`data=${response.data}`)
|
|
||||||
// // console.log(`data=${JSON.stringify(response.data)}`)
|
|
||||||
// // for (const key of Object.keys(response.data)) {
|
|
||||||
// // console.log(`data['${key}']=${response.data[key]}`)
|
|
||||||
// // }
|
|
||||||
|
|
||||||
// // Write archive to file
|
|
||||||
// const runnerTemp = process.env['RUNNER_TEMP'] as string
|
|
||||||
// assert.ok(runnerTemp, 'RUNNER_TEMP not defined')
|
|
||||||
// const archivePath = path.join(runnerTemp, 'checkout.tar.gz')
|
|
||||||
// await io.rmRF(archivePath)
|
|
||||||
// await fs.promises.writeFile(archivePath, raw)
|
// await fs.promises.writeFile(archivePath, raw)
|
||||||
// // await exec.exec(`ls -la "${archiveFile}"`, [], {
|
// // await exec.exec(`ls -la "${archiveFile}"`, [], {
|
||||||
// // cwd: repositoryPath
|
// // cwd: repositoryPath
|
||||||
// // } as ExecOptions)
|
// // } as ExecOptions)
|
||||||
|
|
||||||
// Extract archive
|
// Extract archive
|
||||||
const extractPath = path.join(
|
const extractPath = path.join(runnerTemp, `checkout`)
|
||||||
runnerTemp,
|
|
||||||
`checkout-archive${IS_WINDOWS ? '.zip' : '.tar.gz'}`
|
|
||||||
)
|
|
||||||
await io.rmRF(extractPath)
|
await io.rmRF(extractPath)
|
||||||
await io.mkdirP(extractPath)
|
await io.mkdirP(extractPath)
|
||||||
if (IS_WINDOWS) {
|
if (IS_WINDOWS) {
|
||||||
|
@ -160,14 +138,6 @@ async function downloadArchive(
|
||||||
}
|
}
|
||||||
|
|
||||||
return Buffer.from(response.data) // response.data is ArrayBuffer
|
return Buffer.from(response.data) // response.data is ArrayBuffer
|
||||||
|
|
||||||
// console.log('GETTING THE LOCATION')
|
|
||||||
// const archiveUrl = response.headers['Location'] // Do not print the archive URL because it has an embedded token
|
|
||||||
// assert.ok(
|
|
||||||
// archiveUrl,
|
|
||||||
// `Expected GitHub API response to contain 'Location' header`
|
|
||||||
// )
|
|
||||||
// return archiveUrl
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// async function getArchiveUrl(
|
// async function getArchiveUrl(
|
||||||
|
|
Loading…
Reference in a new issue