Migrate S3 bucket
This commit is contained in:
parent
0e5fd407b9
commit
67299ce6fe
6
Makefile
6
Makefile
|
@ -8,7 +8,7 @@ export
|
|||
|
||||
BUILD := build/$(T)/$(L)
|
||||
DEB := riju-$(T)-$(L).deb
|
||||
S3_DEBS := s3://$(S3_BUCKET)-debs
|
||||
S3_DEBS := s3://$(S3_BUCKET)
|
||||
S3_DEB := $(S3_DEBS)/debs/$(DEB)
|
||||
S3_HASH := $(S3_DEBS)/hashes/riju-$(T)-$(L)
|
||||
|
||||
|
@ -146,9 +146,9 @@ pkg-deb: # L=<lang> T=<type> [Z=gzip|xz] : Build .deb from packaging environment
|
|||
@: $${L} $${T}
|
||||
fakeroot dpkg-deb --build -Z$(Z) $(BUILD)/pkg $(BUILD)/$(DEB)
|
||||
|
||||
## This is equivalent to the sequence 'pkg-clean', 'pkg-build', 'pkg-deb'.
|
||||
## This is equivalent to the sequence 'script', 'pkg-clean', 'pkg-build', 'pkg-deb'.
|
||||
|
||||
pkg: pkg-clean pkg-build pkg-deb # L=<lang> T=<type> [Z=gzip|xz] : Build fresh .deb
|
||||
pkg: script pkg-clean pkg-build pkg-deb # L=<lang> T=<type> [Z=gzip|xz] : Build fresh .deb
|
||||
|
||||
## This is equivalent to 'make pkg T=lang', 'make pkg T=config'. For
|
||||
## shared dependencies, use 'make pkg T=shared' directly.
|
||||
|
|
20
tf/infra.tf
20
tf/infra.tf
|
@ -43,7 +43,7 @@ data "aws_iam_policy_document" "deploy" {
|
|||
]
|
||||
|
||||
resources = [
|
||||
"arn:aws:s3:::${aws_s3_bucket.riju_debs.bucket}",
|
||||
"arn:aws:s3:::${aws_s3_bucket.riju.bucket}",
|
||||
]
|
||||
}
|
||||
|
||||
|
@ -53,7 +53,7 @@ data "aws_iam_policy_document" "deploy" {
|
|||
]
|
||||
|
||||
resources = [
|
||||
"arn:aws:s3:::${aws_s3_bucket.riju_debs.bucket}/*",
|
||||
"arn:aws:s3:::${aws_s3_bucket.riju.bucket}/*",
|
||||
]
|
||||
}
|
||||
}
|
||||
|
@ -69,7 +69,7 @@ resource "aws_iam_user_policy_attachment" "deploy" {
|
|||
policy_arn = aws_iam_policy.deploy.arn
|
||||
}
|
||||
|
||||
data "aws_iam_policy_document" "riju_debs" {
|
||||
data "aws_iam_policy_document" "riju" {
|
||||
statement {
|
||||
principals {
|
||||
type = "*"
|
||||
|
@ -81,7 +81,7 @@ data "aws_iam_policy_document" "riju_debs" {
|
|||
]
|
||||
|
||||
resources = [
|
||||
"arn:aws:s3:::${aws_s3_bucket.riju_debs.bucket}",
|
||||
"arn:aws:s3:::${aws_s3_bucket.riju.bucket}",
|
||||
]
|
||||
}
|
||||
|
||||
|
@ -96,19 +96,19 @@ data "aws_iam_policy_document" "riju_debs" {
|
|||
]
|
||||
|
||||
resources = [
|
||||
"arn:aws:s3:::${aws_s3_bucket.riju_debs.bucket}/*",
|
||||
"arn:aws:s3:::${aws_s3_bucket.riju.bucket}/*",
|
||||
]
|
||||
}
|
||||
}
|
||||
|
||||
resource "aws_s3_bucket" "riju_debs" {
|
||||
bucket = "${data.external.env.result.S3_BUCKET}-debs"
|
||||
resource "aws_s3_bucket" "riju" {
|
||||
bucket = data.external.env.result.S3_BUCKET
|
||||
tags = local.tags
|
||||
}
|
||||
|
||||
resource "aws_s3_bucket_policy" "riju_debs" {
|
||||
bucket = aws_s3_bucket.riju_debs.id
|
||||
policy = data.aws_iam_policy_document.riju_debs.json
|
||||
resource "aws_s3_bucket_policy" "riju" {
|
||||
bucket = aws_s3_bucket.riju.id
|
||||
policy = data.aws_iam_policy_document.riju.json
|
||||
}
|
||||
|
||||
data "aws_ami" "server" {
|
||||
|
|
|
@ -37,7 +37,7 @@ function getInformationalDependencies() {
|
|||
JSON.parse(
|
||||
(
|
||||
await runCommand(
|
||||
`aws s3api list-objects-v2 --bucket riju-debs --prefix hashes`,
|
||||
`aws s3api list-objects-v2 --bucket '${getS3Bucket()}' --prefix hashes`,
|
||||
{ getStdout: true }
|
||||
)
|
||||
).stdout || '{"Contents": []}'
|
||||
|
@ -52,7 +52,7 @@ function getInformationalDependencies() {
|
|||
JSON.parse(
|
||||
(
|
||||
await runCommand(
|
||||
`aws s3api list-objects-v2 --bucket riju-debs --prefix test-hashes/lang`,
|
||||
`aws s3api list-objects-v2 --bucket '${getS3Bucket()}' --prefix test-hashes/lang`,
|
||||
{ getStdout: true }
|
||||
)
|
||||
).stdout || '{"Contents": []}'
|
||||
|
|
|
@ -349,9 +349,7 @@ async function main() {
|
|||
"package category (lang, config, shared)"
|
||||
);
|
||||
program.parse(process.argv);
|
||||
console.log(
|
||||
await generateBuildScript({ lang: program.lang, type: program.type })
|
||||
);
|
||||
console.log(await generateBuildScript(program.opts()));
|
||||
process.exit(0);
|
||||
}
|
||||
|
||||
|
|
Loading…
Reference in New Issue