feat(etl): Add release 2506 86/42986/2
authorPeter Mikus <peter.mikus@icloud.com>
Tue, 20 May 2025 10:15:36 +0000 (12:15 +0200)
committerPeter Mikus <peter.mikus@icloud.com>
Thu, 22 May 2025 04:55:07 +0000 (04:55 +0000)
Signed-off-by: Peter Mikus <peter.mikus@icloud.com>
Change-Id: Iaecddb1156ef899526b7228f9448983d1892a9c5

csit.infra.etl/coverage_hoststack.py
csit.infra.etl/coverage_mrr.py
csit.infra.etl/coverage_ndrpdr.py
csit.infra.etl/coverage_reconf.py
csit.infra.etl/coverage_soak.py
csit.infra.etl/iterative_hoststack.py
csit.infra.etl/iterative_mrr.py
csit.infra.etl/iterative_ndrpdr.py
csit.infra.etl/iterative_reconf.py
csit.infra.etl/iterative_soak.py
fdio.infra.terraform/terraform-nomad-pyspark-etl/variables.tf

index 559dd71..1b4c87e 100644 (file)
@@ -141,7 +141,7 @@ paths = wr.s3.list_objects(
     ignore_empty=True
 )
 
-filtered_paths = [path for path in paths if "report-coverage-2502" in path]
+filtered_paths = [path for path in paths if "report-coverage-2506" in path]
 
 out_sdf = process_json_to_dataframe("hoststack", filtered_paths)
 out_sdf.printSchema()
@@ -163,7 +163,7 @@ except KeyError:
 try:
     wr.s3.to_parquet(
         df=out_sdf.toPandas(),
-        path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/coverage_rls2502",
+        path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/coverage_rls2506",
         dataset=True,
         partition_cols=["test_type", "year", "month", "day"],
         compression="snappy",
index 5cbb982..c371229 100644 (file)
@@ -141,7 +141,7 @@ paths = wr.s3.list_objects(
     ignore_empty=True
 )
 
-filtered_paths = [path for path in paths if "report-coverage-2502" in path]
+filtered_paths = [path for path in paths if "report-coverage-2506" in path]
 
 out_sdf = process_json_to_dataframe("mrr", filtered_paths)
 out_sdf.printSchema()
@@ -163,7 +163,7 @@ except KeyError:
 try:
     wr.s3.to_parquet(
         df=out_sdf.toPandas(),
-        path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/coverage_rls2502",
+        path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/coverage_rls2506",
         dataset=True,
         partition_cols=["test_type", "year", "month", "day"],
         compression="snappy",
index f3c9b63..8a4b0a9 100644 (file)
@@ -141,7 +141,7 @@ paths = wr.s3.list_objects(
     ignore_empty=True
 )
 
-filtered_paths = [path for path in paths if "report-coverage-2502" in path]
+filtered_paths = [path for path in paths if "report-coverage-2506" in path]
 
 out_sdf = process_json_to_dataframe("ndrpdr", filtered_paths)
 out_sdf.printSchema()
@@ -163,7 +163,7 @@ except KeyError:
 try:
     wr.s3.to_parquet(
         df=out_sdf.toPandas(),
-        path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/coverage_rls2502",
+        path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/coverage_rls2506",
         dataset=True,
         partition_cols=["test_type", "year", "month", "day"],
         compression="snappy",
index 307c50f..f24ffa9 100644 (file)
@@ -141,7 +141,7 @@ paths = wr.s3.list_objects(
     ignore_empty=True
 )
 
-filtered_paths = [path for path in paths if "report-coverage-2502" in path]
+filtered_paths = [path for path in paths if "report-coverage-2506" in path]
 
 out_sdf = process_json_to_dataframe("reconf", filtered_paths)
 out_sdf.printSchema()
@@ -163,7 +163,7 @@ except KeyError:
 try:
     wr.s3.to_parquet(
         df=out_sdf.toPandas(),
-        path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/coverage_rls2502",
+        path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/coverage_rls2506",
         dataset=True,
         partition_cols=["test_type", "year", "month", "day"],
         compression="snappy",
index bb6e9fb..82b6f4f 100644 (file)
@@ -141,7 +141,7 @@ paths = wr.s3.list_objects(
     ignore_empty=True
 )
 
-filtered_paths = [path for path in paths if "report-coverage-2502" in path]
+filtered_paths = [path for path in paths if "report-coverage-2506" in path]
 
 out_sdf = process_json_to_dataframe("soak", filtered_paths)
 out_sdf.printSchema()
@@ -163,7 +163,7 @@ except KeyError:
 try:
     wr.s3.to_parquet(
         df=out_sdf.toPandas(),
-        path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/coverage_rls2502",
+        path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/coverage_rls2506",
         dataset=True,
         partition_cols=["test_type", "year", "month", "day"],
         compression="snappy",
index e0d9284..1a97962 100644 (file)
@@ -141,7 +141,7 @@ paths = wr.s3.list_objects(
     ignore_empty=True
 )
 
-filtered_paths = [path for path in paths if "report-iterative-2502" in path]
+filtered_paths = [path for path in paths if "report-iterative-2506" in path]
 
 out_sdf = process_json_to_dataframe("hoststack", filtered_paths)
 out_sdf.printSchema()
@@ -163,7 +163,7 @@ except KeyError:
 try:
     wr.s3.to_parquet(
         df=out_sdf.toPandas(),
-        path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/iterative_rls2502",
+        path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/iterative_rls2506",
         dataset=True,
         partition_cols=["test_type", "year", "month", "day"],
         compression="snappy",
index b1a1e00..c3c8ee3 100644 (file)
@@ -141,7 +141,7 @@ paths = wr.s3.list_objects(
     ignore_empty=True
 )
 
-filtered_paths = [path for path in paths if "report-iterative-2502" in path]
+filtered_paths = [path for path in paths if "report-iterative-2506" in path]
 
 out_sdf = process_json_to_dataframe("mrr", filtered_paths)
 out_sdf.printSchema()
@@ -163,7 +163,7 @@ except KeyError:
 try:
     wr.s3.to_parquet(
         df=out_sdf.toPandas(),
-        path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/iterative_rls2502",
+        path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/iterative_rls2506",
         dataset=True,
         partition_cols=["test_type", "year", "month", "day"],
         compression="snappy",
index f86a32c..40d783d 100644 (file)
@@ -141,7 +141,7 @@ paths = wr.s3.list_objects(
     ignore_empty=True
 )
 
-filtered_paths = [path for path in paths if "report-iterative-2502" in path]
+filtered_paths = [path for path in paths if "report-iterative-2506" in path]
 
 out_sdf = process_json_to_dataframe("ndrpdr", filtered_paths)
 out_sdf.printSchema()
@@ -163,7 +163,7 @@ except KeyError:
 try:
     wr.s3.to_parquet(
         df=out_sdf.toPandas(),
-        path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/iterative_rls2502",
+        path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/iterative_rls2506",
         dataset=True,
         partition_cols=["test_type", "year", "month", "day"],
         compression="snappy",
index 68ee8f0..e5eb2ef 100644 (file)
@@ -141,7 +141,7 @@ paths = wr.s3.list_objects(
     ignore_empty=True
 )
 
-filtered_paths = [path for path in paths if "report-iterative-2502" in path]
+filtered_paths = [path for path in paths if "report-iterative-2506" in path]
 
 out_sdf = process_json_to_dataframe("reconf", filtered_paths)
 out_sdf.show(truncate=False)
@@ -164,7 +164,7 @@ except KeyError:
 try:
     wr.s3.to_parquet(
         df=out_sdf.toPandas(),
-        path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/iterative_rls2502",
+        path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/iterative_rls2506",
         dataset=True,
         partition_cols=["test_type", "year", "month", "day"],
         compression="snappy",
index e213c31..5b77157 100644 (file)
@@ -141,7 +141,7 @@ paths = wr.s3.list_objects(
     ignore_empty=True
 )
 
-filtered_paths = [path for path in paths if "report-iterative-2502" in path]
+filtered_paths = [path for path in paths if "report-iterative-2506" in path]
 
 out_sdf = process_json_to_dataframe("soak", filtered_paths)
 out_sdf.printSchema()
@@ -163,7 +163,7 @@ except KeyError:
 try:
     wr.s3.to_parquet(
         df=out_sdf.toPandas(),
-        path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/iterative_rls2502",
+        path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/iterative_rls2506",
         dataset=True,
         partition_cols=["test_type", "year", "month", "day"],
         compression="snappy",
index dddb405..d0d9fa1 100644 (file)
@@ -54,46 +54,46 @@ variable "nomad_jobs" {
       job_name = "etl-stats"
       memory = 50000
     },
-    #{
-    #  job_name = "etl-iterative-hoststack"
-    #  memory = 50000
-    #},
-    #{
-    #  job_name = "etl-iterative-mrr"
-    #  memory = 50000
-    #},
-    #{
-    #  job_name = "etl-iterative-ndrpdr"
-    #  memory = 50000
-    #},
-    #{
-    #  job_name = "etl-iterative-reconf"
-    #  memory = 50000
-    #},
-    #{
-    #  job_name = "etl-iterative-soak"
-    #  memory = 50000
-    #},
-    #{
-    #  job_name = "etl-coverage-hoststack"
-    #  memory = 50000
-    #},
-    #{
-    #  job_name = "etl-coverage-mrr"
-    #  memory = 50000
-    #},
-    #{
-    #  job_name = "etl-coverage-ndrpdr"
-    #  memory = 50000
-    #},
-    #{
-    #  job_name = "etl-coverage-reconf"
-    #  memory = 50000
-    #},
-    #{
-    #  job_name = "etl-coverage-soak"
-    #  memory = 50000
-    #},
+    {
+      job_name = "etl-iterative-hoststack"
+      memory = 50000
+    },
+    {
+      job_name = "etl-iterative-mrr"
+      memory = 50000
+    },
+    {
+      job_name = "etl-iterative-ndrpdr"
+      memory = 50000
+    },
+    {
+      job_name = "etl-iterative-reconf"
+      memory = 50000
+    },
+    {
+      job_name = "etl-iterative-soak"
+      memory = 50000
+    },
+    {
+      job_name = "etl-coverage-hoststack"
+      memory = 50000
+    },
+    {
+      job_name = "etl-coverage-mrr"
+      memory = 50000
+    },
+    {
+      job_name = "etl-coverage-ndrpdr"
+      memory = 50000
+    },
+    {
+      job_name = "etl-coverage-reconf"
+      memory = 50000
+    },
+    {
+      job_name = "etl-coverage-soak"
+      memory = 50000
+    },
     {
       job_name = "etl-trending-hoststack"
       memory = 50000