Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
10 changes: 7 additions & 3 deletions docker/scripts/app-configure-blobstore.py
Original file line number Diff line number Diff line change
Expand Up @@ -24,10 +24,13 @@ def init_blobstore():
),
)
for bucketname in ARTIFACT_STORAGE_NAMES:
adjusted_bucket_name = (
os.environ.get("BLOB_STORE_BUCKET_PREFIX", "")
+ bucketname
+ os.environ.get("BLOB_STORE_BUCKET_SUFFIX", "")
).strip()
try:
blobstore.create_bucket(
Bucket=f"{os.environ.get('BLOB_STORE_BUCKET_PREFIX', '')}{bucketname}".strip()
)
blobstore.create_bucket(Bucket=adjusted_bucket_name)
except botocore.exceptions.ClientError as err:
if err.response["Error"]["Code"] == "BucketAlreadyExists":
print(f"Bucket {bucketname} already exists")
Expand All @@ -36,5 +39,6 @@ def init_blobstore():
else:
print(f"Bucket {bucketname} created")


if __name__ == "__main__":
sys.exit(init_blobstore())
22 changes: 16 additions & 6 deletions k8s/settings_local.py
Original file line number Diff line number Diff line change
Expand Up @@ -18,7 +18,7 @@


def _multiline_to_list(s):
"""Helper to split at newlines and conver to list"""
"""Helper to split at newlines and convert to list"""
return [item.strip() for item in s.split("\n")]


Expand Down Expand Up @@ -80,13 +80,19 @@ def _multiline_to_list(s):
else:
raise RuntimeError("DATATRACKER_API_PRIVATE_KEY_PEM_B64 must be set")

_RED_PRECOMPUTER_TRIGGER_RETRY_DELAY = os.environ.get("DATATRACKER_RED_PRECOMPUTER_TRIGGER_RETRY_DELAY", None)
_RED_PRECOMPUTER_TRIGGER_RETRY_DELAY = os.environ.get(
"DATATRACKER_RED_PRECOMPUTER_TRIGGER_RETRY_DELAY", None
)
if _RED_PRECOMPUTER_TRIGGER_RETRY_DELAY is not None:
RED_PRECOMPUTER_TRIGGER_RETRY_DELAY = _RED_PRECOMPUTER_TRIGGER_RETRY_DELAY
_RED_PRECOMPUTER_TRIGGER_MAX_RETRIES = os.environ.get("DATATRACKER_RED_PRECOMPUTER_TRIGGER_MAX_RETRIES", None)
RED_PRECOMPUTER_TRIGGER_RETRY_DELAY = _RED_PRECOMPUTER_TRIGGER_RETRY_DELAY
_RED_PRECOMPUTER_TRIGGER_MAX_RETRIES = os.environ.get(
"DATATRACKER_RED_PRECOMPUTER_TRIGGER_MAX_RETRIES", None
)
if _RED_PRECOMPUTER_TRIGGER_MAX_RETRIES is not None:
RED_PRECOMPUTER_TRIGGER_MAX_RETRIES = _RED_PRECOMPUTER_TRIGGER_MAX_RETRIES
_TRIGGER_RED_PRECOMPUTE_MULTIPLE_URL = os.environ.get("DATATRACKER_TRIGGER_RED_PRECOMPUTE_MULTIPLE_URL", None)
_TRIGGER_RED_PRECOMPUTE_MULTIPLE_URL = os.environ.get(
"DATATRACKER_TRIGGER_RED_PRECOMPUTE_MULTIPLE_URL", None
)
if _TRIGGER_RED_PRECOMPUTE_MULTIPLE_URL is not None:
TRIGGER_RED_PRECOMPUTE_MULTIPLE_URL = _TRIGGER_RED_PRECOMPUTE_MULTIPLE_URL

Expand Down Expand Up @@ -387,6 +393,7 @@ def _multiline_to_list(s):
"and DATATRACKER_BLOB_STORE_SECRET_KEY must be set"
)
_blob_store_bucket_prefix = os.environ.get("DATATRACKER_BLOB_STORE_BUCKET_PREFIX", "")
_blob_store_bucket_suffix = os.environ.get("DATATRACKER_BLOB_STORE_BUCKET_SUFFIX", "")
_blob_store_enable_profiling = (
os.environ.get("DATATRACKER_BLOB_STORE_ENABLE_PROFILING", "false").lower() == "true"
)
Expand All @@ -406,6 +413,9 @@ def _multiline_to_list(s):
if storagename in ["staging"]:
continue
replica_storagename = f"r2-{storagename}"
adjusted_bucket_name = (
_blob_store_bucket_prefix + storagename + _blob_store_bucket_suffix
).strip()
STORAGES[replica_storagename] = {
"BACKEND": "ietf.doc.storage.MetadataS3Storage",
"OPTIONS": dict(
Expand All @@ -422,7 +432,7 @@ def _multiline_to_list(s):
retries={"total_max_attempts": _blob_store_max_attempts},
),
verify=False,
bucket_name=f"{_blob_store_bucket_prefix}{storagename}".strip(),
bucket_name=adjusted_bucket_name,
ietf_log_blob_timing=_blob_store_enable_profiling,
),
}
Expand Down
Loading