From 7870a5678fa81941315c8b468ba7313d1a19ecb3 Mon Sep 17 00:00:00 2001 From: Techassi Date: Mon, 18 Nov 2024 16:12:00 +0100 Subject: [PATCH 01/14] chore(release): Update stackableRelease to 24.11 --- stacks/stacks-v2.yaml | 34 +++++++++++++++++----------------- 1 file changed, 17 insertions(+), 17 deletions(-) diff --git a/stacks/stacks-v2.yaml b/stacks/stacks-v2.yaml index 95f5ca29..b08ae29d 100644 --- a/stacks/stacks-v2.yaml +++ b/stacks/stacks-v2.yaml @@ -2,7 +2,7 @@ stacks: monitoring: description: Stack containing Prometheus and Grafana - stackableRelease: dev + stackableRelease: 24.11 stackableOperators: - commons - listener @@ -25,7 +25,7 @@ stacks: default: adminadmin logging: description: Stack containing OpenSearch, OpenSearch Dashboards (Kibana) and Vector aggregator - stackableRelease: dev + stackableRelease: 24.11 stackableOperators: - commons - listener @@ -60,7 +60,7 @@ stacks: observability: description: >- An observability stack with auto-injection of the opentelemetry-collector sidecar to receive traces/logs/metrics via OTLP, and send them to Jaeger/Tempo/Loki. - stackableRelease: dev + stackableRelease: 24.11 stackableOperators: - commons - listener @@ -85,7 +85,7 @@ stacks: default: adminadmin airflow: description: Stack containing Airflow scheduling platform - stackableRelease: dev + stackableRelease: 24.11 stackableOperators: - commons - listener @@ -112,7 +112,7 @@ stacks: default: airflowSecretKey data-lakehouse-iceberg-trino-spark: description: Data lakehouse using Iceberg lakehouse on S3, Trino as query engine, Spark for streaming ingest and Superset for data visualization - stackableRelease: dev + stackableRelease: 24.11 stackableOperators: - commons - listener @@ -169,7 +169,7 @@ stacks: default: supersetSecretKey hdfs-hbase: description: HBase cluster using HDFS as underlying storage - stackableRelease: dev + stackableRelease: 24.11 stackableOperators: - commons - listener @@ -192,7 +192,7 @@ stacks: parameters: [] nifi-kafka-druid-superset-s3: description: Stack containing NiFi, Kafka, Druid, MinIO and Superset for data visualization - stackableRelease: dev + stackableRelease: 24.11 stackableOperators: - commons - listener @@ -238,7 +238,7 @@ stacks: default: adminadmin spark-trino-superset-s3: description: Stack containing MinIO, Trino and Superset for data visualization - stackableRelease: dev + stackableRelease: 24.11 stackableOperators: - commons - listener @@ -283,7 +283,7 @@ stacks: default: supersetSecretKey trino-superset-s3: description: Stack containing MinIO, Trino and Superset for data visualization - stackableRelease: dev + stackableRelease: 24.11 stackableOperators: - commons - listener @@ -325,7 +325,7 @@ stacks: default: supersetSecretKey trino-iceberg: description: Stack containing Trino using Apache Iceberg as a S3 data lakehouse - stackableRelease: dev + stackableRelease: 24.11 stackableOperators: - commons - listener @@ -359,7 +359,7 @@ stacks: default: adminadmin jupyterhub-pyspark-hdfs: description: Jupyterhub with PySpark and HDFS integration - stackableRelease: dev + stackableRelease: 24.11 stackableOperators: - commons - listener @@ -389,7 +389,7 @@ stacks: default: adminadmin dual-hive-hdfs-s3: description: Dual stack Hive on HDFS and S3 for Hadoop/Hive to Trino migration - stackableRelease: dev + stackableRelease: 24.11 stackableOperators: - commons - listener @@ -426,7 +426,7 @@ stacks: The bind user credentials are: ldapadmin:ldapadminpassword. No AuthenticationClass is configured, The AuthenticationClass is created manually in the tutorial. Use the 'openldap' Stack for an OpenLDAD with an AuthenticationClass already installed. - stackableRelease: dev + stackableRelease: 24.11 stackableOperators: - commons - listener @@ -449,7 +449,7 @@ stacks: The bind user credentials are: ldapadmin:ldapadminpassword. The LDAP AuthenticationClass is called 'ldap' and the SecretClass for the bind credentials is called 'ldap-bind-credentials'. The stack already creates an appropriate Secret, so referring to the 'ldap' AuthenticationClass in your ProductCluster should be enough. - stackableRelease: dev + stackableRelease: 24.11 stackableOperators: - commons - listener @@ -475,7 +475,7 @@ stacks: 3 users are created in Keycloak: admin:adminadmin, alice:alicealice, bob:bobbob. admin and alice are admins with full authorization in Druid and Trino, bob is not authorized. This is a proof-of-concept and the mechanisms used here are subject to change. - stackableRelease: dev + stackableRelease: 24.11 stackableOperators: - commons - listener @@ -541,7 +541,7 @@ stacks: Note that this stack is tightly coupled with the demo. So if you install the stack you will get demo-specific parts (such as Keycloak users or regorules). - stackableRelease: dev + stackableRelease: 24.11 stackableOperators: - commons - listener @@ -611,7 +611,7 @@ stacks: signal-processing: description: >- A stack used for creating, streaming and processing in-flight data and persisting it to TimescaleDB before it is displayed in Grafana - stackableRelease: dev + stackableRelease: 24.11 stackableOperators: - commons - listener From e9ea7772d865ed017fbbdf40d42773a967de891c Mon Sep 17 00:00:00 2001 From: Techassi Date: Mon, 18 Nov 2024 16:12:02 +0100 Subject: [PATCH 02/14] chore(release): Update image references with stackable24.11.0 --- .../03-enable-and-run-spark-dag.yaml | 2 +- .../airflow-scheduled-job/04-enable-and-run-date-dag.yaml | 2 +- .../create-nifi-ingestion-job.yaml | 4 ++-- .../create-spark-ingestion-job.yaml | 4 ++-- .../create-trino-tables.yaml | 4 ++-- .../setup-superset.yaml | 2 +- demos/end-to-end-security/create-spark-report.yaml | 4 ++-- demos/end-to-end-security/create-trino-tables.yaml | 2 +- .../create-hfile-and-import-to-hbase.yaml | 2 +- .../load-test-data.yaml | 2 +- .../create-druid-ingestion-job.yaml | 2 +- .../create-nifi-ingestion-job.yaml | 2 +- .../nifi-kafka-druid-earthquake-data/setup-superset.yaml | 2 +- .../create-druid-ingestion-job.yaml | 2 +- .../create-nifi-ingestion-job.yaml | 2 +- .../nifi-kafka-druid-water-level-data/setup-superset.yaml | 2 +- demos/signal-processing/Dockerfile-nifi | 2 +- demos/signal-processing/create-nifi-ingestion-job.yaml | 4 ++-- demos/signal-processing/create-timescale-tables.yaml | 2 +- .../create-spark-anomaly-detection-job.yaml | 4 ++-- .../setup-superset.yaml | 2 +- demos/trino-taxi-data/create-table-in-trino.yaml | 2 +- demos/trino-taxi-data/setup-superset.yaml | 2 +- stacks/_templates/jupyterhub.yaml | 2 +- stacks/_templates/keycloak.yaml | 2 +- stacks/end-to-end-security/krb5.yaml | 8 ++++---- stacks/end-to-end-security/superset.yaml | 2 +- stacks/keycloak-opa-poc/keycloak.yaml | 2 +- stacks/keycloak-opa-poc/setup-keycloak.yaml | 2 +- stacks/logging/setup-opensearch-dashboards.yaml | 2 +- stacks/signal-processing/jupyterhub.yaml | 4 ++-- stacks/signal-processing/nifi.yaml | 2 +- 32 files changed, 42 insertions(+), 42 deletions(-) diff --git a/demos/airflow-scheduled-job/03-enable-and-run-spark-dag.yaml b/demos/airflow-scheduled-job/03-enable-and-run-spark-dag.yaml index dd650853..a64c4e7b 100644 --- a/demos/airflow-scheduled-job/03-enable-and-run-spark-dag.yaml +++ b/demos/airflow-scheduled-job/03-enable-and-run-spark-dag.yaml @@ -8,7 +8,7 @@ spec: spec: containers: - name: start-pyspark-job - image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.7.0 + image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.11.0 # N.B. it is possible for the scheduler to report that a DAG exists, only for the worker task to fail if a pod is unexpectedly # restarted. Additionally, the db-init job takes a few minutes to complete before the cluster is deployed. The wait/watch steps # below are not "water-tight" but add a layer of stability by at least ensuring that the db is initialized and ready and that diff --git a/demos/airflow-scheduled-job/04-enable-and-run-date-dag.yaml b/demos/airflow-scheduled-job/04-enable-and-run-date-dag.yaml index b5e9ba8d..fc0dff9e 100644 --- a/demos/airflow-scheduled-job/04-enable-and-run-date-dag.yaml +++ b/demos/airflow-scheduled-job/04-enable-and-run-date-dag.yaml @@ -8,7 +8,7 @@ spec: spec: containers: - name: start-date-job - image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.7.0 + image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.11.0 # N.B. it is possible for the scheduler to report that a DAG exists, only for the worker task to fail if a pod is unexpectedly # restarted. Additionally, the db-init job takes a few minutes to complete before the cluster is deployed. The wait/watch steps # below are not "water-tight" but add a layer of stability by at least ensuring that the db is initialized and ready and that diff --git a/demos/data-lakehouse-iceberg-trino-spark/create-nifi-ingestion-job.yaml b/demos/data-lakehouse-iceberg-trino-spark/create-nifi-ingestion-job.yaml index 277c6005..9d019ce8 100644 --- a/demos/data-lakehouse-iceberg-trino-spark/create-nifi-ingestion-job.yaml +++ b/demos/data-lakehouse-iceberg-trino-spark/create-nifi-ingestion-job.yaml @@ -9,11 +9,11 @@ spec: serviceAccountName: demo-serviceaccount initContainers: - name: wait-for-kafka - image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.7.0 + image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.11.0 command: ["bash", "-c", "echo 'Waiting for all kafka brokers to be ready' && kubectl wait --for=condition=ready --timeout=30m pod -l app.kubernetes.io/instance=kafka -l app.kubernetes.io/name=kafka"] containers: - name: create-nifi-ingestion-job - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.7.0 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0 command: ["bash", "-c", "curl -O https://raw.githubusercontent.com/stackabletech/demos/main/demos/data-lakehouse-iceberg-trino-spark/LakehouseKafkaIngest.xml && python -u /tmp/script/script.py"] volumeMounts: - name: script diff --git a/demos/data-lakehouse-iceberg-trino-spark/create-spark-ingestion-job.yaml b/demos/data-lakehouse-iceberg-trino-spark/create-spark-ingestion-job.yaml index 423f0fad..5e567e3f 100644 --- a/demos/data-lakehouse-iceberg-trino-spark/create-spark-ingestion-job.yaml +++ b/demos/data-lakehouse-iceberg-trino-spark/create-spark-ingestion-job.yaml @@ -12,11 +12,11 @@ spec: serviceAccountName: demo-serviceaccount initContainers: - name: wait-for-kafka - image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.7.0 + image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.11.0 command: ["bash", "-c", "echo 'Waiting for all kafka brokers to be ready' && kubectl wait --for=condition=ready --timeout=30m pod -l app.kubernetes.io/name=kafka -l app.kubernetes.io/instance=kafka"] containers: - name: create-spark-ingestion-job - image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.7.0 + image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.11.0 command: ["bash", "-c", "echo 'Submitting Spark job' && kubectl apply -f /tmp/manifest/spark-ingestion-job.yaml"] volumeMounts: - name: manifest diff --git a/demos/data-lakehouse-iceberg-trino-spark/create-trino-tables.yaml b/demos/data-lakehouse-iceberg-trino-spark/create-trino-tables.yaml index 8cb6e3da..b3d744e0 100644 --- a/demos/data-lakehouse-iceberg-trino-spark/create-trino-tables.yaml +++ b/demos/data-lakehouse-iceberg-trino-spark/create-trino-tables.yaml @@ -9,11 +9,11 @@ spec: serviceAccountName: demo-serviceaccount initContainers: - name: wait-for-testdata - image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.7.0 + image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.11.0 command: ["bash", "-c", "echo 'Waiting for job load-test-data to finish' && kubectl wait --for=condition=complete --timeout=30m job/load-test-data"] containers: - name: create-tables-in-trino - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.7.0 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0 command: ["bash", "-c", "python -u /tmp/script/script.py"] volumeMounts: - name: script diff --git a/demos/data-lakehouse-iceberg-trino-spark/setup-superset.yaml b/demos/data-lakehouse-iceberg-trino-spark/setup-superset.yaml index d5fdff67..e76139b8 100644 --- a/demos/data-lakehouse-iceberg-trino-spark/setup-superset.yaml +++ b/demos/data-lakehouse-iceberg-trino-spark/setup-superset.yaml @@ -8,7 +8,7 @@ spec: spec: containers: - name: setup-superset - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.7.0 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0 command: ["bash", "-c", "curl -o superset-assets.zip https://raw.githubusercontent.com/stackabletech/demos/main/demos/data-lakehouse-iceberg-trino-spark/superset-assets.zip && python -u /tmp/script/script.py"] volumeMounts: - name: script diff --git a/demos/end-to-end-security/create-spark-report.yaml b/demos/end-to-end-security/create-spark-report.yaml index 2c8c4df5..a5f25af4 100644 --- a/demos/end-to-end-security/create-spark-report.yaml +++ b/demos/end-to-end-security/create-spark-report.yaml @@ -12,7 +12,7 @@ spec: serviceAccountName: demo-serviceaccount initContainers: - name: wait-for-trino-tables - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.7.0 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0 command: - bash - -euo @@ -23,7 +23,7 @@ spec: kubectl wait --timeout=30m --for=condition=complete job/create-tables-in-trino containers: - name: create-spark-report - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.7.0 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0 command: - bash - -euo diff --git a/demos/end-to-end-security/create-trino-tables.yaml b/demos/end-to-end-security/create-trino-tables.yaml index 7c488d5f..6c117160 100644 --- a/demos/end-to-end-security/create-trino-tables.yaml +++ b/demos/end-to-end-security/create-trino-tables.yaml @@ -8,7 +8,7 @@ spec: spec: containers: - name: create-tables-in-trino - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.7.0 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0 command: ["bash", "-c", "python -u /tmp/script/script.py"] volumeMounts: - name: script diff --git a/demos/hbase-hdfs-load-cycling-data/create-hfile-and-import-to-hbase.yaml b/demos/hbase-hdfs-load-cycling-data/create-hfile-and-import-to-hbase.yaml index 7c561ed3..92f031f6 100644 --- a/demos/hbase-hdfs-load-cycling-data/create-hfile-and-import-to-hbase.yaml +++ b/demos/hbase-hdfs-load-cycling-data/create-hfile-and-import-to-hbase.yaml @@ -9,7 +9,7 @@ spec: spec: containers: - name: create-hfile-and-import-to-hbase - image: docker.stackable.tech/stackable/hbase:2.4.18-stackable24.7.0 + image: docker.stackable.tech/stackable/hbase:2.4.18-stackable24.11.0 env: - name: HADOOP_USER_NAME value: stackable diff --git a/demos/jupyterhub-pyspark-hdfs-anomaly-detection-taxi-data/load-test-data.yaml b/demos/jupyterhub-pyspark-hdfs-anomaly-detection-taxi-data/load-test-data.yaml index d02c5088..7dd44f89 100644 --- a/demos/jupyterhub-pyspark-hdfs-anomaly-detection-taxi-data/load-test-data.yaml +++ b/demos/jupyterhub-pyspark-hdfs-anomaly-detection-taxi-data/load-test-data.yaml @@ -8,7 +8,7 @@ spec: spec: containers: - name: load-ny-taxi-data - image: docker.stackable.tech/stackable/hadoop:3.4.0-stackable0.0.0-dev + image: docker.stackable.tech/stackable/hadoop:3.4.0-stackable24.11.0 # yamllint disable rule:line-length command: ["bash", "-c", "/stackable/hadoop/bin/hdfs dfs -mkdir -p /ny-taxi-data/raw \ && cd /tmp \ diff --git a/demos/nifi-kafka-druid-earthquake-data/create-druid-ingestion-job.yaml b/demos/nifi-kafka-druid-earthquake-data/create-druid-ingestion-job.yaml index 3416ed91..c7afa9a0 100644 --- a/demos/nifi-kafka-druid-earthquake-data/create-druid-ingestion-job.yaml +++ b/demos/nifi-kafka-druid-earthquake-data/create-druid-ingestion-job.yaml @@ -8,7 +8,7 @@ spec: spec: containers: - name: create-druid-ingestion-job - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.7.0 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0 command: ["bash", "-c", "curl -X POST --insecure -H 'Content-Type: application/json' -d @/tmp/ingestion-job-spec/ingestion-job-spec.json https://druid-coordinator:8281/druid/indexer/v1/supervisor"] volumeMounts: - name: ingestion-job-spec diff --git a/demos/nifi-kafka-druid-earthquake-data/create-nifi-ingestion-job.yaml b/demos/nifi-kafka-druid-earthquake-data/create-nifi-ingestion-job.yaml index 231d8818..8bdac11b 100644 --- a/demos/nifi-kafka-druid-earthquake-data/create-nifi-ingestion-job.yaml +++ b/demos/nifi-kafka-druid-earthquake-data/create-nifi-ingestion-job.yaml @@ -8,7 +8,7 @@ spec: spec: containers: - name: create-nifi-ingestion-job - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.7.0 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0 command: ["bash", "-c", "curl -O https://raw.githubusercontent.com/stackabletech/demos/main/demos/nifi-kafka-druid-earthquake-data/IngestEarthquakesToKafka.xml && python -u /tmp/script/script.py"] volumeMounts: - name: script diff --git a/demos/nifi-kafka-druid-earthquake-data/setup-superset.yaml b/demos/nifi-kafka-druid-earthquake-data/setup-superset.yaml index b52a2adf..3febaf6d 100644 --- a/demos/nifi-kafka-druid-earthquake-data/setup-superset.yaml +++ b/demos/nifi-kafka-druid-earthquake-data/setup-superset.yaml @@ -8,7 +8,7 @@ spec: spec: containers: - name: setup-superset - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.7.0 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0 command: ["bash", "-c", "curl -o superset-assets.zip https://raw.githubusercontent.com/stackabletech/demos/main/demos/nifi-kafka-druid-earthquake-data/superset-assets.zip && python -u /tmp/script/script.py"] volumeMounts: - name: script diff --git a/demos/nifi-kafka-druid-water-level-data/create-druid-ingestion-job.yaml b/demos/nifi-kafka-druid-water-level-data/create-druid-ingestion-job.yaml index 3c2d6208..b1099c8e 100644 --- a/demos/nifi-kafka-druid-water-level-data/create-druid-ingestion-job.yaml +++ b/demos/nifi-kafka-druid-water-level-data/create-druid-ingestion-job.yaml @@ -8,7 +8,7 @@ spec: spec: containers: - name: create-druid-ingestion-job - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.7.0 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0 command: ["bash", "-c", "curl -X POST --insecure -H 'Content-Type: application/json' -d @/tmp/ingestion-job-spec/stations-ingestion-job-spec.json https://druid-coordinator:8281/druid/indexer/v1/supervisor && curl -X POST --insecure -H 'Content-Type: application/json' -d @/tmp/ingestion-job-spec/measurements-ingestion-job-spec.json https://druid-coordinator:8281/druid/indexer/v1/supervisor && curl -X POST --insecure -H 'Content-Type: application/json' -d @/tmp/ingestion-job-spec/measurements-compaction-job-spec.json https://druid-coordinator:8281/druid/coordinator/v1/config/compaction"] volumeMounts: - name: ingestion-job-spec diff --git a/demos/nifi-kafka-druid-water-level-data/create-nifi-ingestion-job.yaml b/demos/nifi-kafka-druid-water-level-data/create-nifi-ingestion-job.yaml index 6795a681..a42c7c3e 100644 --- a/demos/nifi-kafka-druid-water-level-data/create-nifi-ingestion-job.yaml +++ b/demos/nifi-kafka-druid-water-level-data/create-nifi-ingestion-job.yaml @@ -8,7 +8,7 @@ spec: spec: containers: - name: create-nifi-ingestion-job - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.7.0 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0 command: ["bash", "-c", "curl -O https://raw.githubusercontent.com/stackabletech/demos/main/demos/nifi-kafka-druid-water-level-data/IngestWaterLevelsToKafka.xml && python -u /tmp/script/script.py"] volumeMounts: - name: script diff --git a/demos/nifi-kafka-druid-water-level-data/setup-superset.yaml b/demos/nifi-kafka-druid-water-level-data/setup-superset.yaml index 6cf44c53..b756d04f 100644 --- a/demos/nifi-kafka-druid-water-level-data/setup-superset.yaml +++ b/demos/nifi-kafka-druid-water-level-data/setup-superset.yaml @@ -8,7 +8,7 @@ spec: spec: containers: - name: setup-superset - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.7.0 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0 command: ["bash", "-c", "curl -o superset-assets.zip https://raw.githubusercontent.com/stackabletech/demos/main/demos/nifi-kafka-druid-water-level-data/superset-assets.zip && python -u /tmp/script/script.py"] volumeMounts: - name: script diff --git a/demos/signal-processing/Dockerfile-nifi b/demos/signal-processing/Dockerfile-nifi index db643c3f..1bd50f50 100644 --- a/demos/signal-processing/Dockerfile-nifi +++ b/demos/signal-processing/Dockerfile-nifi @@ -1,3 +1,3 @@ -FROM docker.stackable.tech/stackable/nifi:1.27.0-stackable24.7.0 +FROM docker.stackable.tech/stackable/nifi:1.27.0-stackable24.11.0 RUN curl --fail -o /stackable/nifi/postgresql-42.6.0.jar "https://repo.stackable.tech/repository/misc/postgresql-timescaledb/postgresql-42.6.0.jar" diff --git a/demos/signal-processing/create-nifi-ingestion-job.yaml b/demos/signal-processing/create-nifi-ingestion-job.yaml index 51179a50..52924903 100644 --- a/demos/signal-processing/create-nifi-ingestion-job.yaml +++ b/demos/signal-processing/create-nifi-ingestion-job.yaml @@ -9,13 +9,13 @@ spec: serviceAccountName: demo-serviceaccount initContainers: - name: wait-for-timescale-job - image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.7.0 + image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.11.0 command: ["bash", "-c", "echo 'Waiting for timescaleDB tables to be ready' && kubectl wait --for=condition=complete job/create-timescale-tables-job" ] containers: - name: create-nifi-ingestion-job - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.7.0 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0 command: ["bash", "-c", "export PGPASSWORD=$(cat /timescale-admin-credentials/password) && \ curl -O https://raw.githubusercontent.com/stackabletech/demos/main/demos/signal-processing/DownloadAndWriteToDB.xml && \ sed -i \"s/PLACEHOLDERPGPASSWORD/$PGPASSWORD/g\" DownloadAndWriteToDB.xml && \ diff --git a/demos/signal-processing/create-timescale-tables.yaml b/demos/signal-processing/create-timescale-tables.yaml index 61089f34..0804fa71 100644 --- a/demos/signal-processing/create-timescale-tables.yaml +++ b/demos/signal-processing/create-timescale-tables.yaml @@ -9,7 +9,7 @@ spec: serviceAccountName: demo-serviceaccount initContainers: - name: wait-for-timescale - image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.7.0 + image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.11.0 command: ["bash", "-c", "echo 'Waiting for timescaleDB to be ready' && kubectl wait --for=condition=ready --timeout=30m pod -l app.kubernetes.io/name=postgresql-timescaledb" ] diff --git a/demos/spark-k8s-anomaly-detection-taxi-data/create-spark-anomaly-detection-job.yaml b/demos/spark-k8s-anomaly-detection-taxi-data/create-spark-anomaly-detection-job.yaml index 5dce76c3..3cafbec2 100644 --- a/demos/spark-k8s-anomaly-detection-taxi-data/create-spark-anomaly-detection-job.yaml +++ b/demos/spark-k8s-anomaly-detection-taxi-data/create-spark-anomaly-detection-job.yaml @@ -8,11 +8,11 @@ spec: spec: initContainers: - name: wait-for-testdata - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.7.0 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0 command: ["bash", "-c", "echo 'Waiting for job load-ny-taxi-data to finish' && kubectl wait --for=condition=complete --timeout=30m job/load-ny-taxi-data"] containers: - name: create-spark-anomaly-detection-job - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.7.0 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0 command: ["bash", "-c", "echo 'Submitting Spark job' && kubectl apply -f /tmp/manifest/spark-ad-job.yaml"] volumeMounts: - name: manifest diff --git a/demos/spark-k8s-anomaly-detection-taxi-data/setup-superset.yaml b/demos/spark-k8s-anomaly-detection-taxi-data/setup-superset.yaml index 36aba951..26767455 100644 --- a/demos/spark-k8s-anomaly-detection-taxi-data/setup-superset.yaml +++ b/demos/spark-k8s-anomaly-detection-taxi-data/setup-superset.yaml @@ -8,7 +8,7 @@ spec: spec: containers: - name: setup-superset - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.7.0 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0 command: ["bash", "-c", "curl -o superset-assets.zip https://raw.githubusercontent.com/stackabletech/demos/main/demos/spark-k8s-anomaly-detection-taxi-data/superset-assets.zip && python -u /tmp/script/script.py"] volumeMounts: - name: script diff --git a/demos/trino-taxi-data/create-table-in-trino.yaml b/demos/trino-taxi-data/create-table-in-trino.yaml index d45ce7d9..7e02c16a 100644 --- a/demos/trino-taxi-data/create-table-in-trino.yaml +++ b/demos/trino-taxi-data/create-table-in-trino.yaml @@ -8,7 +8,7 @@ spec: spec: containers: - name: create-ny-taxi-data-table-in-trino - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.7.0 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0 command: ["bash", "-c", "python -u /tmp/script/script.py"] volumeMounts: - name: script diff --git a/demos/trino-taxi-data/setup-superset.yaml b/demos/trino-taxi-data/setup-superset.yaml index 2c94efda..78b24240 100644 --- a/demos/trino-taxi-data/setup-superset.yaml +++ b/demos/trino-taxi-data/setup-superset.yaml @@ -8,7 +8,7 @@ spec: spec: containers: - name: setup-superset - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.7.0 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0 command: ["bash", "-c", "curl -o superset-assets.zip https://raw.githubusercontent.com/stackabletech/demos/main/demos/trino-taxi-data/superset-assets.zip && python -u /tmp/script/script.py"] volumeMounts: - name: script diff --git a/stacks/_templates/jupyterhub.yaml b/stacks/_templates/jupyterhub.yaml index fd4bbd81..55452569 100644 --- a/stacks/_templates/jupyterhub.yaml +++ b/stacks/_templates/jupyterhub.yaml @@ -50,7 +50,7 @@ options: HADOOP_CONF_DIR: "/home/jovyan/hdfs" initContainers: - name: download-notebook - image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.7.0 + image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.11.0 command: ['sh', '-c', 'curl https://raw.githubusercontent.com/stackabletech/demos/main/stacks/jupyterhub-pyspark-hdfs/notebook.ipynb -o /notebook/notebook.ipynb'] volumeMounts: - mountPath: /notebook diff --git a/stacks/_templates/keycloak.yaml b/stacks/_templates/keycloak.yaml index ecc9a9fb..b620e50f 100644 --- a/stacks/_templates/keycloak.yaml +++ b/stacks/_templates/keycloak.yaml @@ -48,7 +48,7 @@ spec: - name: tls mountPath: /tls/ - name: create-auth-class - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.7.0 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0 command: ["/bin/bash", "-c"] args: - | diff --git a/stacks/end-to-end-security/krb5.yaml b/stacks/end-to-end-security/krb5.yaml index 19657231..c9dd6018 100644 --- a/stacks/end-to-end-security/krb5.yaml +++ b/stacks/end-to-end-security/krb5.yaml @@ -14,7 +14,7 @@ spec: spec: initContainers: - name: init - image: docker.stackable.tech/stackable/krb5:1.21.1-stackable24.7.0 + image: docker.stackable.tech/stackable/krb5:1.21.1-stackable24.11.0 args: - sh - -euo @@ -35,7 +35,7 @@ spec: name: data containers: - name: kdc - image: docker.stackable.tech/stackable/krb5:1.21.1-stackable24.7.0 + image: docker.stackable.tech/stackable/krb5:1.21.1-stackable24.11.0 args: - krb5kdc - -n @@ -48,7 +48,7 @@ spec: - mountPath: /var/kerberos/krb5kdc name: data - name: kadmind - image: docker.stackable.tech/stackable/krb5:1.21.1-stackable24.7.0 + image: docker.stackable.tech/stackable/krb5:1.21.1-stackable24.11.0 args: - kadmind - -nofork @@ -61,7 +61,7 @@ spec: - mountPath: /var/kerberos/krb5kdc name: data - name: client - image: docker.stackable.tech/stackable/krb5:1.21.1-stackable24.7.0 + image: docker.stackable.tech/stackable/krb5:1.21.1-stackable24.11.0 tty: true stdin: true env: diff --git a/stacks/end-to-end-security/superset.yaml b/stacks/end-to-end-security/superset.yaml index 0577245e..09533874 100644 --- a/stacks/end-to-end-security/superset.yaml +++ b/stacks/end-to-end-security/superset.yaml @@ -25,7 +25,7 @@ spec: initContainers: # The postgres image does not contain curl or wget... - name: download-dump - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.7.0 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0 command: - bash - -c diff --git a/stacks/keycloak-opa-poc/keycloak.yaml b/stacks/keycloak-opa-poc/keycloak.yaml index a6c2e225..0e484a96 100644 --- a/stacks/keycloak-opa-poc/keycloak.yaml +++ b/stacks/keycloak-opa-poc/keycloak.yaml @@ -70,7 +70,7 @@ spec: spec: containers: - name: propagate-keycloak-address - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.7.0 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0 command: - bash - -x diff --git a/stacks/keycloak-opa-poc/setup-keycloak.yaml b/stacks/keycloak-opa-poc/setup-keycloak.yaml index f21d64a2..499f8632 100644 --- a/stacks/keycloak-opa-poc/setup-keycloak.yaml +++ b/stacks/keycloak-opa-poc/setup-keycloak.yaml @@ -29,7 +29,7 @@ spec: spec: containers: - name: setup-keycloak - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.7.0 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0 env: - name: KEYCLOAK_ADMIN_PASSWORD valueFrom: diff --git a/stacks/logging/setup-opensearch-dashboards.yaml b/stacks/logging/setup-opensearch-dashboards.yaml index c3b4330d..ed565227 100644 --- a/stacks/logging/setup-opensearch-dashboards.yaml +++ b/stacks/logging/setup-opensearch-dashboards.yaml @@ -8,7 +8,7 @@ spec: spec: containers: - name: setup-opensearch-dashboards - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.7.0 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0 env: - name: OPEN_SEARCH_ADMIN_PASSWORD valueFrom: diff --git a/stacks/signal-processing/jupyterhub.yaml b/stacks/signal-processing/jupyterhub.yaml index f26e5988..4557865c 100644 --- a/stacks/signal-processing/jupyterhub.yaml +++ b/stacks/signal-processing/jupyterhub.yaml @@ -30,7 +30,7 @@ options: singleuser: cmd: null image: - # TODO (@NickLarsenNZ): Use a versioned image with stackable0.0.0-dev or stackableXX.X.X so that + # TODO (@NickLarsenNZ): Use a versioned image with stackable24.11.0 or stackableXX.X.X so that # the demo is reproducable for the release and it will be automatically replaced for the release branch. name: docker.stackable.tech/demos/jupyter-pyspark-with-alibi-detect tag: python-3.9 @@ -41,7 +41,7 @@ options: stackable.tech/vendor: Stackable initContainers: - name: download-notebook - image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.7.0 + image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.11.0 command: ['sh', '-c', 'curl https://raw.githubusercontent.com/stackabletech/demos/main/stacks/signal-processing/tsdb.ipynb -o /notebook/tsdb.ipynb'] volumeMounts: - mountPath: /notebook diff --git a/stacks/signal-processing/nifi.yaml b/stacks/signal-processing/nifi.yaml index 48d7c39d..5e758282 100644 --- a/stacks/signal-processing/nifi.yaml +++ b/stacks/signal-processing/nifi.yaml @@ -6,7 +6,7 @@ metadata: spec: image: productVersion: 1.27.0 - # TODO (@NickLarsenNZ): Use a versioned image with stackable0.0.0-dev or stackableXX.X.X so that + # TODO (@NickLarsenNZ): Use a versioned image with stackable24.11.0 or stackableXX.X.X so that # the demo is reproducable for the release and it will be automatically replaced for the release branch. custom: docker.stackable.tech/demos/nifi:1.27.0-postgresql clusterConfig: From d3be4baed22a045cd137c51c1578eec34e32abf2 Mon Sep 17 00:00:00 2001 From: Techassi Date: Mon, 18 Nov 2024 16:12:02 +0100 Subject: [PATCH 03/14] chore(release): Replace githubusercontent references main->release-24.11 --- .../create-nifi-ingestion-job.yaml | 2 +- .../setup-superset.yaml | 2 +- demos/demos-v2.yaml | 66 ++--- .../create-nifi-ingestion-job.yaml | 2 +- .../setup-superset.yaml | 2 +- .../create-nifi-ingestion-job.yaml | 2 +- .../setup-superset.yaml | 2 +- .../create-nifi-ingestion-job.yaml | 2 +- .../setup-superset.yaml | 2 +- demos/trino-taxi-data/setup-superset.yaml | 2 +- stacks/_templates/jupyterhub.yaml | 2 +- stacks/end-to-end-security/superset.yaml | 2 +- stacks/signal-processing/jupyterhub.yaml | 2 +- stacks/stacks-v2.yaml | 226 +++++++++--------- 14 files changed, 158 insertions(+), 158 deletions(-) diff --git a/demos/data-lakehouse-iceberg-trino-spark/create-nifi-ingestion-job.yaml b/demos/data-lakehouse-iceberg-trino-spark/create-nifi-ingestion-job.yaml index 9d019ce8..8ae380ec 100644 --- a/demos/data-lakehouse-iceberg-trino-spark/create-nifi-ingestion-job.yaml +++ b/demos/data-lakehouse-iceberg-trino-spark/create-nifi-ingestion-job.yaml @@ -14,7 +14,7 @@ spec: containers: - name: create-nifi-ingestion-job image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0 - command: ["bash", "-c", "curl -O https://raw.githubusercontent.com/stackabletech/demos/main/demos/data-lakehouse-iceberg-trino-spark/LakehouseKafkaIngest.xml && python -u /tmp/script/script.py"] + command: ["bash", "-c", "curl -O https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/data-lakehouse-iceberg-trino-spark/LakehouseKafkaIngest.xml && python -u /tmp/script/script.py"] volumeMounts: - name: script mountPath: /tmp/script diff --git a/demos/data-lakehouse-iceberg-trino-spark/setup-superset.yaml b/demos/data-lakehouse-iceberg-trino-spark/setup-superset.yaml index e76139b8..c225fe88 100644 --- a/demos/data-lakehouse-iceberg-trino-spark/setup-superset.yaml +++ b/demos/data-lakehouse-iceberg-trino-spark/setup-superset.yaml @@ -9,7 +9,7 @@ spec: containers: - name: setup-superset image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0 - command: ["bash", "-c", "curl -o superset-assets.zip https://raw.githubusercontent.com/stackabletech/demos/main/demos/data-lakehouse-iceberg-trino-spark/superset-assets.zip && python -u /tmp/script/script.py"] + command: ["bash", "-c", "curl -o superset-assets.zip https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/data-lakehouse-iceberg-trino-spark/superset-assets.zip && python -u /tmp/script/script.py"] volumeMounts: - name: script mountPath: /tmp/script diff --git a/demos/demos-v2.yaml b/demos/demos-v2.yaml index a31ff1ea..9b52292b 100644 --- a/demos/demos-v2.yaml +++ b/demos/demos-v2.yaml @@ -7,10 +7,10 @@ demos: - airflow - job-scheduling manifests: - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/airflow-scheduled-job/01-airflow-spark-clusterrole.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/airflow-scheduled-job/02-airflow-spark-clusterrolebinding.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/airflow-scheduled-job/03-enable-and-run-spark-dag.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/airflow-scheduled-job/04-enable-and-run-date-dag.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/airflow-scheduled-job/01-airflow-spark-clusterrole.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/airflow-scheduled-job/02-airflow-spark-clusterrolebinding.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/airflow-scheduled-job/03-enable-and-run-spark-dag.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/airflow-scheduled-job/04-enable-and-run-date-dag.yaml supportedNamespaces: [] resourceRequests: cpu: 2401m @@ -24,8 +24,8 @@ demos: - hdfs - cycling-tripdata manifests: - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/hbase-hdfs-load-cycling-data/distcp-cycling-data.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/hbase-hdfs-load-cycling-data/create-hfile-and-import-to-hbase.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/hbase-hdfs-load-cycling-data/distcp-cycling-data.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/hbase-hdfs-load-cycling-data/create-hfile-and-import-to-hbase.yaml supportedNamespaces: [] resourceRequests: cpu: "3" @@ -43,9 +43,9 @@ demos: - opa - keycloak manifests: - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/end-to-end-security/create-trino-tables.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/end-to-end-security/serviceaccount.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/end-to-end-security/create-spark-report.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/end-to-end-security/create-trino-tables.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/end-to-end-security/serviceaccount.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/end-to-end-security/create-spark-report.yaml supportedNamespaces: [] resourceRequests: cpu: 9000m @@ -64,9 +64,9 @@ demos: - s3 - earthquakes manifests: - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/nifi-kafka-druid-earthquake-data/create-nifi-ingestion-job.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/nifi-kafka-druid-earthquake-data/create-druid-ingestion-job.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/nifi-kafka-druid-earthquake-data/setup-superset.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/nifi-kafka-druid-earthquake-data/create-nifi-ingestion-job.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/nifi-kafka-druid-earthquake-data/create-druid-ingestion-job.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/nifi-kafka-druid-earthquake-data/setup-superset.yaml supportedNamespaces: ["default"] resourceRequests: cpu: 8700m @@ -85,9 +85,9 @@ demos: - s3 - water-levels manifests: - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/nifi-kafka-druid-water-level-data/create-nifi-ingestion-job.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/nifi-kafka-druid-water-level-data/create-druid-ingestion-job.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/nifi-kafka-druid-water-level-data/setup-superset.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/nifi-kafka-druid-water-level-data/create-nifi-ingestion-job.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/nifi-kafka-druid-water-level-data/create-druid-ingestion-job.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/nifi-kafka-druid-water-level-data/setup-superset.yaml supportedNamespaces: ["default"] resourceRequests: cpu: 8900m @@ -104,10 +104,10 @@ demos: - s3 - ny-taxi-data manifests: - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/spark-k8s-anomaly-detection-taxi-data/serviceaccount.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/spark-k8s-anomaly-detection-taxi-data/load-test-data.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/spark-k8s-anomaly-detection-taxi-data/create-spark-anomaly-detection-job.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/spark-k8s-anomaly-detection-taxi-data/setup-superset.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/spark-k8s-anomaly-detection-taxi-data/serviceaccount.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/spark-k8s-anomaly-detection-taxi-data/load-test-data.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/spark-k8s-anomaly-detection-taxi-data/create-spark-anomaly-detection-job.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/spark-k8s-anomaly-detection-taxi-data/setup-superset.yaml supportedNamespaces: [] resourceRequests: cpu: 6400m @@ -139,9 +139,9 @@ demos: - s3 - ny-taxi-data manifests: - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/trino-taxi-data/load-test-data.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/trino-taxi-data/create-table-in-trino.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/trino-taxi-data/setup-superset.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/trino-taxi-data/load-test-data.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/trino-taxi-data/create-table-in-trino.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/trino-taxi-data/setup-superset.yaml supportedNamespaces: [] resourceRequests: cpu: 6800m @@ -164,12 +164,12 @@ demos: - water-levels - earthquakes manifests: - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/data-lakehouse-iceberg-trino-spark/serviceaccount.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/data-lakehouse-iceberg-trino-spark/load-test-data.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/data-lakehouse-iceberg-trino-spark/create-trino-tables.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/data-lakehouse-iceberg-trino-spark/create-nifi-ingestion-job.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/data-lakehouse-iceberg-trino-spark/create-spark-ingestion-job.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/data-lakehouse-iceberg-trino-spark/setup-superset.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/data-lakehouse-iceberg-trino-spark/serviceaccount.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/data-lakehouse-iceberg-trino-spark/load-test-data.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/data-lakehouse-iceberg-trino-spark/create-trino-tables.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/data-lakehouse-iceberg-trino-spark/create-nifi-ingestion-job.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/data-lakehouse-iceberg-trino-spark/create-spark-ingestion-job.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/data-lakehouse-iceberg-trino-spark/setup-superset.yaml supportedNamespaces: ["default"] resourceRequests: cpu: "80" @@ -185,7 +185,7 @@ demos: - pyspark - ny-taxi-data manifests: - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/jupyterhub-pyspark-hdfs-anomaly-detection-taxi-data/load-test-data.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/jupyterhub-pyspark-hdfs-anomaly-detection-taxi-data/load-test-data.yaml supportedNamespaces: [] resourceRequests: cpu: 3350m @@ -202,7 +202,7 @@ demos: - vector - zookeeper manifests: - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/logging/zookeeper.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/logging/zookeeper.yaml supportedNamespaces: [] resourceRequests: cpu: 6500m @@ -218,9 +218,9 @@ demos: - grafana-dashboards - zookeeper manifests: - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/signal-processing/serviceaccount.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/signal-processing/create-timescale-tables.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/signal-processing/create-nifi-ingestion-job.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/signal-processing/serviceaccount.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/signal-processing/create-timescale-tables.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/signal-processing/create-nifi-ingestion-job.yaml supportedNamespaces: [] resourceRequests: cpu: "3" diff --git a/demos/nifi-kafka-druid-earthquake-data/create-nifi-ingestion-job.yaml b/demos/nifi-kafka-druid-earthquake-data/create-nifi-ingestion-job.yaml index 8bdac11b..ddec687b 100644 --- a/demos/nifi-kafka-druid-earthquake-data/create-nifi-ingestion-job.yaml +++ b/demos/nifi-kafka-druid-earthquake-data/create-nifi-ingestion-job.yaml @@ -9,7 +9,7 @@ spec: containers: - name: create-nifi-ingestion-job image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0 - command: ["bash", "-c", "curl -O https://raw.githubusercontent.com/stackabletech/demos/main/demos/nifi-kafka-druid-earthquake-data/IngestEarthquakesToKafka.xml && python -u /tmp/script/script.py"] + command: ["bash", "-c", "curl -O https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/nifi-kafka-druid-earthquake-data/IngestEarthquakesToKafka.xml && python -u /tmp/script/script.py"] volumeMounts: - name: script mountPath: /tmp/script diff --git a/demos/nifi-kafka-druid-earthquake-data/setup-superset.yaml b/demos/nifi-kafka-druid-earthquake-data/setup-superset.yaml index 3febaf6d..76a2de68 100644 --- a/demos/nifi-kafka-druid-earthquake-data/setup-superset.yaml +++ b/demos/nifi-kafka-druid-earthquake-data/setup-superset.yaml @@ -9,7 +9,7 @@ spec: containers: - name: setup-superset image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0 - command: ["bash", "-c", "curl -o superset-assets.zip https://raw.githubusercontent.com/stackabletech/demos/main/demos/nifi-kafka-druid-earthquake-data/superset-assets.zip && python -u /tmp/script/script.py"] + command: ["bash", "-c", "curl -o superset-assets.zip https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/nifi-kafka-druid-earthquake-data/superset-assets.zip && python -u /tmp/script/script.py"] volumeMounts: - name: script mountPath: /tmp/script diff --git a/demos/nifi-kafka-druid-water-level-data/create-nifi-ingestion-job.yaml b/demos/nifi-kafka-druid-water-level-data/create-nifi-ingestion-job.yaml index a42c7c3e..26520651 100644 --- a/demos/nifi-kafka-druid-water-level-data/create-nifi-ingestion-job.yaml +++ b/demos/nifi-kafka-druid-water-level-data/create-nifi-ingestion-job.yaml @@ -9,7 +9,7 @@ spec: containers: - name: create-nifi-ingestion-job image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0 - command: ["bash", "-c", "curl -O https://raw.githubusercontent.com/stackabletech/demos/main/demos/nifi-kafka-druid-water-level-data/IngestWaterLevelsToKafka.xml && python -u /tmp/script/script.py"] + command: ["bash", "-c", "curl -O https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/nifi-kafka-druid-water-level-data/IngestWaterLevelsToKafka.xml && python -u /tmp/script/script.py"] volumeMounts: - name: script mountPath: /tmp/script diff --git a/demos/nifi-kafka-druid-water-level-data/setup-superset.yaml b/demos/nifi-kafka-druid-water-level-data/setup-superset.yaml index b756d04f..fc73baa8 100644 --- a/demos/nifi-kafka-druid-water-level-data/setup-superset.yaml +++ b/demos/nifi-kafka-druid-water-level-data/setup-superset.yaml @@ -9,7 +9,7 @@ spec: containers: - name: setup-superset image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0 - command: ["bash", "-c", "curl -o superset-assets.zip https://raw.githubusercontent.com/stackabletech/demos/main/demos/nifi-kafka-druid-water-level-data/superset-assets.zip && python -u /tmp/script/script.py"] + command: ["bash", "-c", "curl -o superset-assets.zip https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/nifi-kafka-druid-water-level-data/superset-assets.zip && python -u /tmp/script/script.py"] volumeMounts: - name: script mountPath: /tmp/script diff --git a/demos/signal-processing/create-nifi-ingestion-job.yaml b/demos/signal-processing/create-nifi-ingestion-job.yaml index 52924903..0d59c7ba 100644 --- a/demos/signal-processing/create-nifi-ingestion-job.yaml +++ b/demos/signal-processing/create-nifi-ingestion-job.yaml @@ -17,7 +17,7 @@ spec: - name: create-nifi-ingestion-job image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0 command: ["bash", "-c", "export PGPASSWORD=$(cat /timescale-admin-credentials/password) && \ - curl -O https://raw.githubusercontent.com/stackabletech/demos/main/demos/signal-processing/DownloadAndWriteToDB.xml && \ + curl -O https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/signal-processing/DownloadAndWriteToDB.xml && \ sed -i \"s/PLACEHOLDERPGPASSWORD/$PGPASSWORD/g\" DownloadAndWriteToDB.xml && \ python -u /tmp/script/script.py"] volumeMounts: diff --git a/demos/spark-k8s-anomaly-detection-taxi-data/setup-superset.yaml b/demos/spark-k8s-anomaly-detection-taxi-data/setup-superset.yaml index 26767455..a2a704f6 100644 --- a/demos/spark-k8s-anomaly-detection-taxi-data/setup-superset.yaml +++ b/demos/spark-k8s-anomaly-detection-taxi-data/setup-superset.yaml @@ -9,7 +9,7 @@ spec: containers: - name: setup-superset image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0 - command: ["bash", "-c", "curl -o superset-assets.zip https://raw.githubusercontent.com/stackabletech/demos/main/demos/spark-k8s-anomaly-detection-taxi-data/superset-assets.zip && python -u /tmp/script/script.py"] + command: ["bash", "-c", "curl -o superset-assets.zip https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/spark-k8s-anomaly-detection-taxi-data/superset-assets.zip && python -u /tmp/script/script.py"] volumeMounts: - name: script mountPath: /tmp/script diff --git a/demos/trino-taxi-data/setup-superset.yaml b/demos/trino-taxi-data/setup-superset.yaml index 78b24240..9902ba7f 100644 --- a/demos/trino-taxi-data/setup-superset.yaml +++ b/demos/trino-taxi-data/setup-superset.yaml @@ -9,7 +9,7 @@ spec: containers: - name: setup-superset image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0 - command: ["bash", "-c", "curl -o superset-assets.zip https://raw.githubusercontent.com/stackabletech/demos/main/demos/trino-taxi-data/superset-assets.zip && python -u /tmp/script/script.py"] + command: ["bash", "-c", "curl -o superset-assets.zip https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/trino-taxi-data/superset-assets.zip && python -u /tmp/script/script.py"] volumeMounts: - name: script mountPath: /tmp/script diff --git a/stacks/_templates/jupyterhub.yaml b/stacks/_templates/jupyterhub.yaml index 55452569..1939afd9 100644 --- a/stacks/_templates/jupyterhub.yaml +++ b/stacks/_templates/jupyterhub.yaml @@ -51,7 +51,7 @@ options: initContainers: - name: download-notebook image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.11.0 - command: ['sh', '-c', 'curl https://raw.githubusercontent.com/stackabletech/demos/main/stacks/jupyterhub-pyspark-hdfs/notebook.ipynb -o /notebook/notebook.ipynb'] + command: ['sh', '-c', 'curl https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/jupyterhub-pyspark-hdfs/notebook.ipynb -o /notebook/notebook.ipynb'] volumeMounts: - mountPath: /notebook name: notebook diff --git a/stacks/end-to-end-security/superset.yaml b/stacks/end-to-end-security/superset.yaml index 09533874..c80fd82f 100644 --- a/stacks/end-to-end-security/superset.yaml +++ b/stacks/end-to-end-security/superset.yaml @@ -31,7 +31,7 @@ spec: - -c - | cd /tmp - curl --fail -O https://raw.githubusercontent.com/stackabletech/demos/main/stacks/end-to-end-security/postgres_superset_dump.sql.gz + curl --fail -O https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/end-to-end-security/postgres_superset_dump.sql.gz gunzip postgres_superset_dump.sql.gz # We need to omit changing the users password, as otherwise the content in the Secrets does not match diff --git a/stacks/signal-processing/jupyterhub.yaml b/stacks/signal-processing/jupyterhub.yaml index 4557865c..8bf3a23d 100644 --- a/stacks/signal-processing/jupyterhub.yaml +++ b/stacks/signal-processing/jupyterhub.yaml @@ -42,7 +42,7 @@ options: initContainers: - name: download-notebook image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.11.0 - command: ['sh', '-c', 'curl https://raw.githubusercontent.com/stackabletech/demos/main/stacks/signal-processing/tsdb.ipynb -o /notebook/tsdb.ipynb'] + command: ['sh', '-c', 'curl https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/signal-processing/tsdb.ipynb -o /notebook/tsdb.ipynb'] volumeMounts: - mountPath: /notebook name: notebook diff --git a/stacks/stacks-v2.yaml b/stacks/stacks-v2.yaml index b08ae29d..21224a60 100644 --- a/stacks/stacks-v2.yaml +++ b/stacks/stacks-v2.yaml @@ -11,9 +11,9 @@ stacks: - prometheus - grafana manifests: - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/monitoring/grafana-dashboards.yaml - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/prometheus.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/prometheus-service-monitor.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/monitoring/grafana-dashboards.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/prometheus.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/prometheus-service-monitor.yaml supportedNamespaces: [] resourceRequests: cpu: 1750m @@ -37,11 +37,11 @@ stacks: - opensearch-dashboards - vector manifests: - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/opensearch.yaml - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/opensearch-dashboards.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/logging/setup-opensearch-dashboards.yaml - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/vector-aggregator.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/vector-aggregator-discovery.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/opensearch.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/opensearch-dashboards.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/logging/setup-opensearch-dashboards.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/vector-aggregator.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/vector-aggregator-discovery.yaml supportedNamespaces: [] resourceRequests: cpu: 5150m @@ -71,14 +71,14 @@ stacks: - observability - tracing manifests: - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/observability/jaeger.yaml - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/observability/opentelemetry-operator.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/observability/grafana-admin-credentials.yaml - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/observability/grafana.yaml - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/observability/grafana-tempo.yaml - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/observability/grafana-loki.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/observability/opentelemetry-collector-sidecar.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/observability/opentelemetry-collector-deployment.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/observability/jaeger.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/observability/opentelemetry-operator.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/observability/grafana-admin-credentials.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/observability/grafana.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/observability/grafana-tempo.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/observability/grafana-loki.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/observability/opentelemetry-collector-sidecar.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/observability/opentelemetry-collector-deployment.yaml parameters: - name: grafanaAdminPassword description: Password of the Grafana admin user @@ -95,9 +95,9 @@ stacks: labels: - airflow manifests: - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/postgresql-airflow.yaml - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/redis-airflow.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/airflow/airflow.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/postgresql-airflow.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/redis-airflow.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/airflow/airflow.yaml supportedNamespaces: [] resourceRequests: cpu: 3400m @@ -135,17 +135,17 @@ stacks: - minio - s3 manifests: - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/minio-distributed.yaml - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/postgresql-hive.yaml - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/postgresql-hive-iceberg.yaml - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/postgresql-superset.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/data-lakehouse-iceberg-trino-spark/s3-connection.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/data-lakehouse-iceberg-trino-spark/hive-metastores.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/data-lakehouse-iceberg-trino-spark/trino.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/data-lakehouse-iceberg-trino-spark/zookeeper.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/data-lakehouse-iceberg-trino-spark/kafka.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/data-lakehouse-iceberg-trino-spark/nifi.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/nifi-kafka-druid-superset-s3/superset.yaml # Reuse + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/minio-distributed.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/postgresql-hive.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/postgresql-hive-iceberg.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/postgresql-superset.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/data-lakehouse-iceberg-trino-spark/s3-connection.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/data-lakehouse-iceberg-trino-spark/hive-metastores.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/data-lakehouse-iceberg-trino-spark/trino.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/data-lakehouse-iceberg-trino-spark/zookeeper.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/data-lakehouse-iceberg-trino-spark/kafka.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/data-lakehouse-iceberg-trino-spark/nifi.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/nifi-kafka-druid-superset-s3/superset.yaml # Reuse supportedNamespaces: [] resourceRequests: cpu: "71" @@ -181,9 +181,9 @@ stacks: - hbase - hdfs manifests: - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/hdfs-hbase/zookeeper.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/hdfs-hbase/hdfs.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/hdfs-hbase/hbase.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/hdfs-hbase/zookeeper.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/hdfs-hbase/hdfs.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/hdfs-hbase/hbase.yaml supportedNamespaces: [] resourceRequests: cpu: 4200m @@ -210,14 +210,14 @@ stacks: - minio - s3 manifests: - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/minio.yaml - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/postgresql-druid.yaml - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/postgresql-superset.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/nifi-kafka-druid-superset-s3/zookeeper.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/nifi-kafka-druid-superset-s3/kafka.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/nifi-kafka-druid-superset-s3/druid.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/nifi-kafka-druid-superset-s3/superset.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/nifi-kafka-druid-superset-s3/nifi.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/minio.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/postgresql-druid.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/postgresql-superset.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/nifi-kafka-druid-superset-s3/zookeeper.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/nifi-kafka-druid-superset-s3/kafka.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/nifi-kafka-druid-superset-s3/druid.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/nifi-kafka-druid-superset-s3/superset.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/nifi-kafka-druid-superset-s3/nifi.yaml supportedNamespaces: [] resourceRequests: cpu: 8900m @@ -254,15 +254,15 @@ stacks: - minio - s3 manifests: - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/minio.yaml - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/postgresql-hive.yaml - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/postgresql-hive-iceberg.yaml - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/postgresql-superset.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/trino-superset-s3/s3-connection.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/spark-trino-superset-s3/hive-metastore.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/trino-superset-s3/trino.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/spark-trino-superset-s3/trino-prediction-catalog.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/trino-superset-s3/superset.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/minio.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/postgresql-hive.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/postgresql-hive-iceberg.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/postgresql-superset.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/trino-superset-s3/s3-connection.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/spark-trino-superset-s3/hive-metastore.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/trino-superset-s3/trino.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/spark-trino-superset-s3/trino-prediction-catalog.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/trino-superset-s3/superset.yaml supportedNamespaces: [] resourceRequests: cpu: 7100m @@ -298,13 +298,13 @@ stacks: - minio - s3 manifests: - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/minio.yaml - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/postgresql-hive.yaml - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/postgresql-superset.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/trino-superset-s3/s3-connection.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/trino-superset-s3/hive-metastore.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/trino-superset-s3/trino.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/trino-superset-s3/superset.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/minio.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/postgresql-hive.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/postgresql-superset.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/trino-superset-s3/s3-connection.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/trino-superset-s3/hive-metastore.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/trino-superset-s3/trino.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/trino-superset-s3/superset.yaml supportedNamespaces: [] resourceRequests: cpu: 6800m @@ -340,11 +340,11 @@ stacks: - minio - s3 manifests: - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/minio-distributed-small.yaml - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/postgresql-hive-iceberg.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/trino-iceberg/s3-connection.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/trino-iceberg/hive-metastores.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/trino-iceberg/trino.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/minio-distributed-small.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/postgresql-hive-iceberg.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/trino-iceberg/s3-connection.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/trino-iceberg/hive-metastores.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/trino-iceberg/trino.yaml supportedNamespaces: [] resourceRequests: cpu: 6000m # Measured 5600m @@ -372,12 +372,12 @@ stacks: - hdfs - pyspark manifests: - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/jupyterhub.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/jupyterhub-pyspark-hdfs/zookeeper.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/jupyterhub-pyspark-hdfs/hdfs.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/jupyterhub-pyspark-hdfs/serviceaccount.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/jupyterhub.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/jupyterhub-pyspark-hdfs/zookeeper.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/jupyterhub-pyspark-hdfs/hdfs.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/jupyterhub-pyspark-hdfs/serviceaccount.yaml # TODO Use patched JHub that created service for us from customer setup (ask Sebastian) - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/jupyterhub-pyspark-hdfs/spark_driver_service.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/jupyterhub-pyspark-hdfs/spark_driver_service.yaml supportedNamespaces: [] resourceRequests: cpu: 3350m @@ -405,12 +405,12 @@ stacks: - hdfs - s3 manifests: - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/postgresql-hivehdfs.yaml - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/postgresql-hives3.yaml - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/minio.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/dual-hive-hdfs-s3/hdfs.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/dual-hive-hdfs-s3/hive.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/dual-hive-hdfs-s3/trino.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/postgresql-hivehdfs.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/postgresql-hives3.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/minio.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/dual-hive-hdfs-s3/hdfs.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/dual-hive-hdfs-s3/hive.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/dual-hive-hdfs-s3/trino.yaml supportedNamespaces: [] resourceRequests: cpu: 7750m @@ -435,7 +435,7 @@ stacks: - authentication - ldap manifests: - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/authentication/openldap-tls.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/authentication/openldap-tls.yaml supportedNamespaces: ["default"] resourceRequests: cpu: 1950m @@ -458,8 +458,8 @@ stacks: - authentication - ldap manifests: - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/authentication/openldap-tls.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/authentication/openldap-tls-authenticationclass.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/authentication/openldap-tls.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/authentication/openldap-tls-authenticationclass.yaml supportedNamespaces: [] resourceRequests: cpu: 1950m @@ -490,18 +490,18 @@ stacks: - authentication - sso manifests: - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/postgresql-superset.yaml - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/postgresql-druid.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/keycloak-opa-poc/serviceaccount.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/keycloak-opa-poc/keycloak.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/keycloak-opa-poc/setup-keycloak.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/keycloak-opa-poc/opa.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/keycloak-opa-poc/policies.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/keycloak-opa-poc/zookeeper.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/keycloak-opa-poc/hdfs.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/keycloak-opa-poc/druid.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/keycloak-opa-poc/trino.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/keycloak-opa-poc/superset.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/postgresql-superset.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/postgresql-druid.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/keycloak-opa-poc/serviceaccount.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/keycloak-opa-poc/keycloak.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/keycloak-opa-poc/setup-keycloak.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/keycloak-opa-poc/opa.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/keycloak-opa-poc/policies.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/keycloak-opa-poc/zookeeper.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/keycloak-opa-poc/hdfs.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/keycloak-opa-poc/druid.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/keycloak-opa-poc/trino.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/keycloak-opa-poc/superset.yaml supportedNamespaces: ["default"] # ClusterRoleBinding needs explicit namespace resourceRequests: cpu: 7850m @@ -567,22 +567,22 @@ stacks: memory: 19586Mi pvc: 40Gi manifests: - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/keycloak-serviceaccount.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/end-to-end-security/keycloak-realm-config.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/keycloak.yaml - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/postgresql-hive-iceberg.yaml - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/postgresql-superset.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/end-to-end-security/krb5.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/end-to-end-security/kerberos-secretclass.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/end-to-end-security/opa.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/end-to-end-security/zookeeper.yaml # TODO: Add authentication - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/end-to-end-security/hdfs.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/end-to-end-security/hdfs-regorules.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/end-to-end-security/hive-metastore.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/end-to-end-security/trino.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/end-to-end-security/trino-regorules.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/end-to-end-security/trino-policies.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/end-to-end-security/superset.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/keycloak-serviceaccount.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/end-to-end-security/keycloak-realm-config.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/keycloak.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/postgresql-hive-iceberg.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/postgresql-superset.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/end-to-end-security/krb5.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/end-to-end-security/kerberos-secretclass.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/end-to-end-security/opa.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/end-to-end-security/zookeeper.yaml # TODO: Add authentication + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/end-to-end-security/hdfs.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/end-to-end-security/hdfs-regorules.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/end-to-end-security/hive-metastore.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/end-to-end-security/trino.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/end-to-end-security/trino-regorules.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/end-to-end-security/trino-policies.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/end-to-end-security/superset.yaml parameters: - name: keycloakAdminPassword description: Password of the Keycloak admin user @@ -624,15 +624,15 @@ stacks: - jupyterhub - grafana manifests: - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/signal-processing/secrets.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/signal-processing/grafana-dashboards.yaml - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/signal-processing/grafana.yaml - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/postgresql-timescaledb.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/nifi-kafka-druid-superset-s3/zookeeper.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/signal-processing/nifi.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/jupyterhub-pyspark-hdfs/serviceaccount.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/jupyterhub-pyspark-hdfs/spark_driver_service.yaml - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/signal-processing/jupyterhub.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/signal-processing/secrets.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/signal-processing/grafana-dashboards.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/signal-processing/grafana.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/postgresql-timescaledb.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/nifi-kafka-druid-superset-s3/zookeeper.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/signal-processing/nifi.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/jupyterhub-pyspark-hdfs/serviceaccount.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/jupyterhub-pyspark-hdfs/spark_driver_service.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/signal-processing/jupyterhub.yaml parameters: - name: nifiAdminPassword description: Password of the NiFI admin user From 06db1dab3e0e3d34cd62555342d2d86654dde0d9 Mon Sep 17 00:00:00 2001 From: Techassi Date: Tue, 26 Nov 2024 16:42:28 +0100 Subject: [PATCH 04/14] chore: Change docs version from nightly to 24.11 --- docs/antora.yml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/docs/antora.yml b/docs/antora.yml index 189f7d75..2d6b5aae 100644 --- a/docs/antora.yml +++ b/docs/antora.yml @@ -2,4 +2,4 @@ # Use 'home' here so that the versioning is picked up automatically based # on SDP releases name: home -version: "nightly" +version: "24.11" From 541f044c630a495dfb7499b848c6644d23c8b2be Mon Sep 17 00:00:00 2001 From: Techassi Date: Tue, 26 Nov 2024 17:29:49 +0100 Subject: [PATCH 05/14] chore: Remove superfluous page-aliases --- docs/modules/demos/pages/airflow-scheduled-job.adoc | 1 - docs/modules/demos/pages/data-lakehouse-iceberg-trino-spark.adoc | 1 - docs/modules/demos/pages/hbase-hdfs-load-cycling-data.adoc | 1 - docs/modules/demos/pages/index.adoc | 1 - .../jupyterhub-pyspark-hdfs-anomaly-detection-taxi-data.adoc | 1 - docs/modules/demos/pages/logging.adoc | 1 - docs/modules/demos/pages/nifi-kafka-druid-earthquake-data.adoc | 1 - docs/modules/demos/pages/nifi-kafka-druid-water-level-data.adoc | 1 - .../demos/pages/spark-k8s-anomaly-detection-taxi-data.adoc | 1 - docs/modules/demos/pages/trino-iceberg.adoc | 1 - docs/modules/demos/pages/trino-taxi-data.adoc | 1 - 11 files changed, 11 deletions(-) diff --git a/docs/modules/demos/pages/airflow-scheduled-job.adoc b/docs/modules/demos/pages/airflow-scheduled-job.adoc index 1ecf3d80..933de6f1 100644 --- a/docs/modules/demos/pages/airflow-scheduled-job.adoc +++ b/docs/modules/demos/pages/airflow-scheduled-job.adoc @@ -1,5 +1,4 @@ = airflow-scheduled-job -:page-aliases: stable@stackablectl::demos/airflow-scheduled-job.adoc :description: This demo installs Airflow with Postgres and Redis on Kubernetes, showcasing DAG scheduling, job runs, and status verification via the Airflow UI. Install this demo on an existing Kubernetes cluster: diff --git a/docs/modules/demos/pages/data-lakehouse-iceberg-trino-spark.adoc b/docs/modules/demos/pages/data-lakehouse-iceberg-trino-spark.adoc index f79dd363..96429e96 100644 --- a/docs/modules/demos/pages/data-lakehouse-iceberg-trino-spark.adoc +++ b/docs/modules/demos/pages/data-lakehouse-iceberg-trino-spark.adoc @@ -1,5 +1,4 @@ = data-lakehouse-iceberg-trino-spark -:page-aliases: stable@stackablectl::demos/data-lakehouse-iceberg-trino-spark.adoc :description: This demo shows a data workload with real-world data volumes using Trino, Kafka, Spark, NiFi, Superset and OPA. :demo-code: https://github.com/stackabletech/demos/blob/main/demos/data-lakehouse-iceberg-trino-spark/create-spark-ingestion-job.yaml diff --git a/docs/modules/demos/pages/hbase-hdfs-load-cycling-data.adoc b/docs/modules/demos/pages/hbase-hdfs-load-cycling-data.adoc index 1d00ddd3..8421491c 100644 --- a/docs/modules/demos/pages/hbase-hdfs-load-cycling-data.adoc +++ b/docs/modules/demos/pages/hbase-hdfs-load-cycling-data.adoc @@ -1,5 +1,4 @@ = hbase-hdfs-cycling-data -:page-aliases: stable@stackablectl::demos/hbase-hdfs-load-cycling-data.adoc :description: Load cyclist data from HDFS to HBase on Kubernetes using Stackable's demo. Install, copy data, create HFiles, and query efficiently. :kaggle: https://www.kaggle.com/datasets/timgid/cyclistic-dataset-google-certificate-capstone?select=Divvy_Trips_2020_Q1.csv diff --git a/docs/modules/demos/pages/index.adoc b/docs/modules/demos/pages/index.adoc index 65ada08a..0e8b83cc 100644 --- a/docs/modules/demos/pages/index.adoc +++ b/docs/modules/demos/pages/index.adoc @@ -1,5 +1,4 @@ = Demos -:page-aliases: stable@stackablectl::demos/index.adoc :description: Explore Stackable demos showcasing data platform architectures. Includes external components for evaluation. The pages in this section guide you on how to use the demos provided by Stackable. diff --git a/docs/modules/demos/pages/jupyterhub-pyspark-hdfs-anomaly-detection-taxi-data.adoc b/docs/modules/demos/pages/jupyterhub-pyspark-hdfs-anomaly-detection-taxi-data.adoc index a732b82e..10f04d62 100644 --- a/docs/modules/demos/pages/jupyterhub-pyspark-hdfs-anomaly-detection-taxi-data.adoc +++ b/docs/modules/demos/pages/jupyterhub-pyspark-hdfs-anomaly-detection-taxi-data.adoc @@ -1,5 +1,4 @@ = jupyterhub-pyspark-hdfs-anomaly-detection-taxi-data -:page-aliases: stable@stackablectl::demos/jupyterhub-pyspark-hdfs-anomaly-detection-taxi-data.adoc :scikit-lib: https://scikit-learn.org/stable/modules/generated/sklearn.ensemble.IsolationForest.html :k8s-cpu: https://kubernetes.io/docs/tasks/debug/debug-cluster/resource-metrics-pipeline/#cpu diff --git a/docs/modules/demos/pages/logging.adoc b/docs/modules/demos/pages/logging.adoc index 0466d0eb..8d28082b 100644 --- a/docs/modules/demos/pages/logging.adoc +++ b/docs/modules/demos/pages/logging.adoc @@ -1,5 +1,4 @@ = logging -:page-aliases: stable@stackablectl::demos/logging.adoc :description: Deploy a logging stack with OpenSearch, Vector, and Zookeeper for log data analysis using OpenSearch Dashboards in Kubernetes. :k8s-cpu: https://kubernetes.io/docs/tasks/debug/debug-cluster/resource-metrics-pipeline/#cpu diff --git a/docs/modules/demos/pages/nifi-kafka-druid-earthquake-data.adoc b/docs/modules/demos/pages/nifi-kafka-druid-earthquake-data.adoc index dff15bf5..f49a40c8 100644 --- a/docs/modules/demos/pages/nifi-kafka-druid-earthquake-data.adoc +++ b/docs/modules/demos/pages/nifi-kafka-druid-earthquake-data.adoc @@ -1,5 +1,4 @@ = nifi-kafka-druid-earthquake-data -:page-aliases: stable@stackablectl::demos/nifi-kafka-druid-earthquake-data.adoc :description: Install this demo for a showcase of using Kafka, Druid and Superset to view the global earthquake distribution. :superset-docs: https://superset.apache.org/docs/using-superset/creating-your-first-dashboard/#creating-charts-in-explore-view diff --git a/docs/modules/demos/pages/nifi-kafka-druid-water-level-data.adoc b/docs/modules/demos/pages/nifi-kafka-druid-water-level-data.adoc index 0d25daa6..090ea66c 100644 --- a/docs/modules/demos/pages/nifi-kafka-druid-water-level-data.adoc +++ b/docs/modules/demos/pages/nifi-kafka-druid-water-level-data.adoc @@ -1,5 +1,4 @@ = nifi-kafka-druid-water-level-data -:page-aliases: stable@stackablectl::demos/nifi-kafka-druid-water-level-data.adoc :description: Install this demo for a showcase of using Kafka, Druid and Superset to visualize water levels in across Germany. :superset: https://superset.apache.org/docs/using-superset/creating-your-first-dashboard/#creating-charts-in-explore-view diff --git a/docs/modules/demos/pages/spark-k8s-anomaly-detection-taxi-data.adoc b/docs/modules/demos/pages/spark-k8s-anomaly-detection-taxi-data.adoc index 319c523e..fcd12656 100644 --- a/docs/modules/demos/pages/spark-k8s-anomaly-detection-taxi-data.adoc +++ b/docs/modules/demos/pages/spark-k8s-anomaly-detection-taxi-data.adoc @@ -1,5 +1,4 @@ = spark-k8s-anomaly-detection-taxi-data -:page-aliases: stable@stackablectl::demos/spark-k8s-anomaly-detection-taxi-data.adoc :description: Deploy a Kubernetes-based Spark demo for anomaly detection using the popular New York taxi dataset, featuring Trino, Spark, MinIO, and Superset. :scikit-lib: https://scikit-learn.org/stable/modules/generated/sklearn.ensemble.IsolationForest.html diff --git a/docs/modules/demos/pages/trino-iceberg.adoc b/docs/modules/demos/pages/trino-iceberg.adoc index 34af036d..433e3a70 100644 --- a/docs/modules/demos/pages/trino-iceberg.adoc +++ b/docs/modules/demos/pages/trino-iceberg.adoc @@ -1,5 +1,4 @@ = trino-iceberg -:page-aliases: stable@stackablectl::demos/trino-iceberg.adoc :description: Install and explore Trino with Apache Iceberg for efficient SQL queries and scalable data management in a demo environment. :k8s-cpu: https://kubernetes.io/docs/tasks/debug/debug-cluster/resource-metrics-pipeline/#cpu diff --git a/docs/modules/demos/pages/trino-taxi-data.adoc b/docs/modules/demos/pages/trino-taxi-data.adoc index e8e9328f..541bb3ae 100644 --- a/docs/modules/demos/pages/trino-taxi-data.adoc +++ b/docs/modules/demos/pages/trino-taxi-data.adoc @@ -1,5 +1,4 @@ = trino-taxi-data -:page-aliases: stable@stackablectl::demos/trino-taxi-data.adoc :description: Install and demo Trino with NYC taxi data: Query with SQL, visualize with Superset, and explore data in MinIO and Trino on Kubernetes. :superset-docs: https://superset.apache.org/docs/creating-charts-dashboards/creating-your-first-dashboard#creating-charts-in-explore-view From 1757cdac69103eba9a1c6b35a6eddded6c803355 Mon Sep 17 00:00:00 2001 From: Nick Larsen Date: Thu, 16 Jan 2025 10:29:35 +0100 Subject: [PATCH 06/14] chore: Explicitly bump 24.11.0 to 24.11.1 These changes should be pulled into `main`, but this script needs more work. --- .scripts/update_refs.sh | 8 ++++---- 1 file changed, 4 insertions(+), 4 deletions(-) diff --git a/.scripts/update_refs.sh b/.scripts/update_refs.sh index abe1f622..32cd97ed 100755 --- a/.scripts/update_refs.sh +++ b/.scripts/update_refs.sh @@ -35,10 +35,10 @@ function prepend { function maybe_commit { [ "$COMMIT" == "true" ] || return 0 local MESSAGE="$1" - PATCH=$(mktemp) + PATCH=$(mktemp --suffix=.diff) git add -u git diff --staged > "$PATCH" - git commit -S -m "$MESSAGE" --no-verify + git diff-index --quiet HEAD -- || git commit -S -m "$MESSAGE" --no-verify echo "patch written to: $PATCH" | prepend "\t" } @@ -55,8 +55,8 @@ if [[ "$CURRENT_BRANCH" == release-* ]]; then # Replace 0.0.0-dev refs with ${STACKABLE_RELEASE}.0 # TODO (@NickLarsenNZ): handle patches later, and what about release-candidates? - SEARCH='stackable(0\.0\.0-dev|24\.7\.[0-9]+)' # TODO (@NickLarsenNZ): After https://github.com/stackabletech/stackable-cockpit/issues/310, only search for 0.0.0-dev - REPLACEMENT="stackable${STACKABLE_RELEASE}.0" # TODO (@NickLarsenNZ): Be a bit smarter about patch releases. + SEARCH='stackable(0\.0\.0-dev|24\.7\.[0-9]+|24\.11\.0)' # TODO (@NickLarsenNZ): After https://github.com/stackabletech/stackable-cockpit/issues/310, only search for 0.0.0-dev + REPLACEMENT="stackable${STACKABLE_RELEASE}.1" # TODO (@NickLarsenNZ): Be a bit smarter about patch releases. MESSAGE="Update image references with $REPLACEMENT" echo "$MESSAGE" find demos stacks -type f \ From 2f5212fb2cb0de5aa9d15d74630cb8b9d86ebf9a Mon Sep 17 00:00:00 2001 From: Nick Larsen Date: Thu, 16 Jan 2025 10:39:49 +0100 Subject: [PATCH 07/14] chore(release): Update image references with stackable24.11.1 --- .../03-enable-and-run-spark-dag.yaml | 2 +- .../airflow-scheduled-job/04-enable-and-run-date-dag.yaml | 2 +- .../create-nifi-ingestion-job.yaml | 4 ++-- .../create-spark-ingestion-job.yaml | 4 ++-- .../create-trino-tables.yaml | 4 ++-- .../setup-superset.yaml | 2 +- demos/end-to-end-security/create-spark-report.yaml | 4 ++-- demos/end-to-end-security/create-trino-tables.yaml | 2 +- .../create-hfile-and-import-to-hbase.yaml | 2 +- .../load-test-data.yaml | 2 +- .../create-druid-ingestion-job.yaml | 2 +- .../create-nifi-ingestion-job.yaml | 2 +- .../nifi-kafka-druid-earthquake-data/setup-superset.yaml | 2 +- .../create-druid-ingestion-job.yaml | 2 +- .../create-nifi-ingestion-job.yaml | 2 +- .../nifi-kafka-druid-water-level-data/setup-superset.yaml | 2 +- demos/signal-processing/Dockerfile-nifi | 2 +- demos/signal-processing/create-nifi-ingestion-job.yaml | 4 ++-- demos/signal-processing/create-timescale-tables.yaml | 2 +- .../create-spark-anomaly-detection-job.yaml | 4 ++-- .../setup-superset.yaml | 2 +- demos/trino-taxi-data/create-table-in-trino.yaml | 2 +- demos/trino-taxi-data/setup-superset.yaml | 2 +- stacks/_templates/jupyterhub.yaml | 2 +- stacks/_templates/keycloak.yaml | 2 +- stacks/end-to-end-security/krb5.yaml | 8 ++++---- stacks/end-to-end-security/superset.yaml | 2 +- stacks/keycloak-opa-poc/keycloak.yaml | 2 +- stacks/keycloak-opa-poc/setup-keycloak.yaml | 2 +- stacks/logging/setup-opensearch-dashboards.yaml | 2 +- stacks/signal-processing/jupyterhub.yaml | 4 ++-- stacks/signal-processing/nifi.yaml | 2 +- 32 files changed, 42 insertions(+), 42 deletions(-) diff --git a/demos/airflow-scheduled-job/03-enable-and-run-spark-dag.yaml b/demos/airflow-scheduled-job/03-enable-and-run-spark-dag.yaml index a64c4e7b..46e2760f 100644 --- a/demos/airflow-scheduled-job/03-enable-and-run-spark-dag.yaml +++ b/demos/airflow-scheduled-job/03-enable-and-run-spark-dag.yaml @@ -8,7 +8,7 @@ spec: spec: containers: - name: start-pyspark-job - image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.11.0 + image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.11.1 # N.B. it is possible for the scheduler to report that a DAG exists, only for the worker task to fail if a pod is unexpectedly # restarted. Additionally, the db-init job takes a few minutes to complete before the cluster is deployed. The wait/watch steps # below are not "water-tight" but add a layer of stability by at least ensuring that the db is initialized and ready and that diff --git a/demos/airflow-scheduled-job/04-enable-and-run-date-dag.yaml b/demos/airflow-scheduled-job/04-enable-and-run-date-dag.yaml index fc0dff9e..84c42550 100644 --- a/demos/airflow-scheduled-job/04-enable-and-run-date-dag.yaml +++ b/demos/airflow-scheduled-job/04-enable-and-run-date-dag.yaml @@ -8,7 +8,7 @@ spec: spec: containers: - name: start-date-job - image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.11.0 + image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.11.1 # N.B. it is possible for the scheduler to report that a DAG exists, only for the worker task to fail if a pod is unexpectedly # restarted. Additionally, the db-init job takes a few minutes to complete before the cluster is deployed. The wait/watch steps # below are not "water-tight" but add a layer of stability by at least ensuring that the db is initialized and ready and that diff --git a/demos/data-lakehouse-iceberg-trino-spark/create-nifi-ingestion-job.yaml b/demos/data-lakehouse-iceberg-trino-spark/create-nifi-ingestion-job.yaml index 8ae380ec..98472784 100644 --- a/demos/data-lakehouse-iceberg-trino-spark/create-nifi-ingestion-job.yaml +++ b/demos/data-lakehouse-iceberg-trino-spark/create-nifi-ingestion-job.yaml @@ -9,11 +9,11 @@ spec: serviceAccountName: demo-serviceaccount initContainers: - name: wait-for-kafka - image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.11.0 + image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.11.1 command: ["bash", "-c", "echo 'Waiting for all kafka brokers to be ready' && kubectl wait --for=condition=ready --timeout=30m pod -l app.kubernetes.io/instance=kafka -l app.kubernetes.io/name=kafka"] containers: - name: create-nifi-ingestion-job - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1 command: ["bash", "-c", "curl -O https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/data-lakehouse-iceberg-trino-spark/LakehouseKafkaIngest.xml && python -u /tmp/script/script.py"] volumeMounts: - name: script diff --git a/demos/data-lakehouse-iceberg-trino-spark/create-spark-ingestion-job.yaml b/demos/data-lakehouse-iceberg-trino-spark/create-spark-ingestion-job.yaml index 5e567e3f..526a8867 100644 --- a/demos/data-lakehouse-iceberg-trino-spark/create-spark-ingestion-job.yaml +++ b/demos/data-lakehouse-iceberg-trino-spark/create-spark-ingestion-job.yaml @@ -12,11 +12,11 @@ spec: serviceAccountName: demo-serviceaccount initContainers: - name: wait-for-kafka - image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.11.0 + image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.11.1 command: ["bash", "-c", "echo 'Waiting for all kafka brokers to be ready' && kubectl wait --for=condition=ready --timeout=30m pod -l app.kubernetes.io/name=kafka -l app.kubernetes.io/instance=kafka"] containers: - name: create-spark-ingestion-job - image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.11.0 + image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.11.1 command: ["bash", "-c", "echo 'Submitting Spark job' && kubectl apply -f /tmp/manifest/spark-ingestion-job.yaml"] volumeMounts: - name: manifest diff --git a/demos/data-lakehouse-iceberg-trino-spark/create-trino-tables.yaml b/demos/data-lakehouse-iceberg-trino-spark/create-trino-tables.yaml index b3d744e0..4a8a2e5d 100644 --- a/demos/data-lakehouse-iceberg-trino-spark/create-trino-tables.yaml +++ b/demos/data-lakehouse-iceberg-trino-spark/create-trino-tables.yaml @@ -9,11 +9,11 @@ spec: serviceAccountName: demo-serviceaccount initContainers: - name: wait-for-testdata - image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.11.0 + image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.11.1 command: ["bash", "-c", "echo 'Waiting for job load-test-data to finish' && kubectl wait --for=condition=complete --timeout=30m job/load-test-data"] containers: - name: create-tables-in-trino - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1 command: ["bash", "-c", "python -u /tmp/script/script.py"] volumeMounts: - name: script diff --git a/demos/data-lakehouse-iceberg-trino-spark/setup-superset.yaml b/demos/data-lakehouse-iceberg-trino-spark/setup-superset.yaml index c225fe88..53b0bb54 100644 --- a/demos/data-lakehouse-iceberg-trino-spark/setup-superset.yaml +++ b/demos/data-lakehouse-iceberg-trino-spark/setup-superset.yaml @@ -8,7 +8,7 @@ spec: spec: containers: - name: setup-superset - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1 command: ["bash", "-c", "curl -o superset-assets.zip https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/data-lakehouse-iceberg-trino-spark/superset-assets.zip && python -u /tmp/script/script.py"] volumeMounts: - name: script diff --git a/demos/end-to-end-security/create-spark-report.yaml b/demos/end-to-end-security/create-spark-report.yaml index a5f25af4..ded96dea 100644 --- a/demos/end-to-end-security/create-spark-report.yaml +++ b/demos/end-to-end-security/create-spark-report.yaml @@ -12,7 +12,7 @@ spec: serviceAccountName: demo-serviceaccount initContainers: - name: wait-for-trino-tables - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1 command: - bash - -euo @@ -23,7 +23,7 @@ spec: kubectl wait --timeout=30m --for=condition=complete job/create-tables-in-trino containers: - name: create-spark-report - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1 command: - bash - -euo diff --git a/demos/end-to-end-security/create-trino-tables.yaml b/demos/end-to-end-security/create-trino-tables.yaml index 6c117160..469d8925 100644 --- a/demos/end-to-end-security/create-trino-tables.yaml +++ b/demos/end-to-end-security/create-trino-tables.yaml @@ -8,7 +8,7 @@ spec: spec: containers: - name: create-tables-in-trino - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1 command: ["bash", "-c", "python -u /tmp/script/script.py"] volumeMounts: - name: script diff --git a/demos/hbase-hdfs-load-cycling-data/create-hfile-and-import-to-hbase.yaml b/demos/hbase-hdfs-load-cycling-data/create-hfile-and-import-to-hbase.yaml index 92f031f6..b822d886 100644 --- a/demos/hbase-hdfs-load-cycling-data/create-hfile-and-import-to-hbase.yaml +++ b/demos/hbase-hdfs-load-cycling-data/create-hfile-and-import-to-hbase.yaml @@ -9,7 +9,7 @@ spec: spec: containers: - name: create-hfile-and-import-to-hbase - image: docker.stackable.tech/stackable/hbase:2.4.18-stackable24.11.0 + image: docker.stackable.tech/stackable/hbase:2.4.18-stackable24.11.1 env: - name: HADOOP_USER_NAME value: stackable diff --git a/demos/jupyterhub-pyspark-hdfs-anomaly-detection-taxi-data/load-test-data.yaml b/demos/jupyterhub-pyspark-hdfs-anomaly-detection-taxi-data/load-test-data.yaml index 7dd44f89..22abb7b1 100644 --- a/demos/jupyterhub-pyspark-hdfs-anomaly-detection-taxi-data/load-test-data.yaml +++ b/demos/jupyterhub-pyspark-hdfs-anomaly-detection-taxi-data/load-test-data.yaml @@ -8,7 +8,7 @@ spec: spec: containers: - name: load-ny-taxi-data - image: docker.stackable.tech/stackable/hadoop:3.4.0-stackable24.11.0 + image: docker.stackable.tech/stackable/hadoop:3.4.0-stackable24.11.1 # yamllint disable rule:line-length command: ["bash", "-c", "/stackable/hadoop/bin/hdfs dfs -mkdir -p /ny-taxi-data/raw \ && cd /tmp \ diff --git a/demos/nifi-kafka-druid-earthquake-data/create-druid-ingestion-job.yaml b/demos/nifi-kafka-druid-earthquake-data/create-druid-ingestion-job.yaml index c7afa9a0..82dd92b8 100644 --- a/demos/nifi-kafka-druid-earthquake-data/create-druid-ingestion-job.yaml +++ b/demos/nifi-kafka-druid-earthquake-data/create-druid-ingestion-job.yaml @@ -8,7 +8,7 @@ spec: spec: containers: - name: create-druid-ingestion-job - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1 command: ["bash", "-c", "curl -X POST --insecure -H 'Content-Type: application/json' -d @/tmp/ingestion-job-spec/ingestion-job-spec.json https://druid-coordinator:8281/druid/indexer/v1/supervisor"] volumeMounts: - name: ingestion-job-spec diff --git a/demos/nifi-kafka-druid-earthquake-data/create-nifi-ingestion-job.yaml b/demos/nifi-kafka-druid-earthquake-data/create-nifi-ingestion-job.yaml index ddec687b..0d2a1e95 100644 --- a/demos/nifi-kafka-druid-earthquake-data/create-nifi-ingestion-job.yaml +++ b/demos/nifi-kafka-druid-earthquake-data/create-nifi-ingestion-job.yaml @@ -8,7 +8,7 @@ spec: spec: containers: - name: create-nifi-ingestion-job - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1 command: ["bash", "-c", "curl -O https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/nifi-kafka-druid-earthquake-data/IngestEarthquakesToKafka.xml && python -u /tmp/script/script.py"] volumeMounts: - name: script diff --git a/demos/nifi-kafka-druid-earthquake-data/setup-superset.yaml b/demos/nifi-kafka-druid-earthquake-data/setup-superset.yaml index 76a2de68..b3a72a3e 100644 --- a/demos/nifi-kafka-druid-earthquake-data/setup-superset.yaml +++ b/demos/nifi-kafka-druid-earthquake-data/setup-superset.yaml @@ -8,7 +8,7 @@ spec: spec: containers: - name: setup-superset - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1 command: ["bash", "-c", "curl -o superset-assets.zip https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/nifi-kafka-druid-earthquake-data/superset-assets.zip && python -u /tmp/script/script.py"] volumeMounts: - name: script diff --git a/demos/nifi-kafka-druid-water-level-data/create-druid-ingestion-job.yaml b/demos/nifi-kafka-druid-water-level-data/create-druid-ingestion-job.yaml index b1099c8e..ce8a0a28 100644 --- a/demos/nifi-kafka-druid-water-level-data/create-druid-ingestion-job.yaml +++ b/demos/nifi-kafka-druid-water-level-data/create-druid-ingestion-job.yaml @@ -8,7 +8,7 @@ spec: spec: containers: - name: create-druid-ingestion-job - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1 command: ["bash", "-c", "curl -X POST --insecure -H 'Content-Type: application/json' -d @/tmp/ingestion-job-spec/stations-ingestion-job-spec.json https://druid-coordinator:8281/druid/indexer/v1/supervisor && curl -X POST --insecure -H 'Content-Type: application/json' -d @/tmp/ingestion-job-spec/measurements-ingestion-job-spec.json https://druid-coordinator:8281/druid/indexer/v1/supervisor && curl -X POST --insecure -H 'Content-Type: application/json' -d @/tmp/ingestion-job-spec/measurements-compaction-job-spec.json https://druid-coordinator:8281/druid/coordinator/v1/config/compaction"] volumeMounts: - name: ingestion-job-spec diff --git a/demos/nifi-kafka-druid-water-level-data/create-nifi-ingestion-job.yaml b/demos/nifi-kafka-druid-water-level-data/create-nifi-ingestion-job.yaml index 26520651..f46f64dc 100644 --- a/demos/nifi-kafka-druid-water-level-data/create-nifi-ingestion-job.yaml +++ b/demos/nifi-kafka-druid-water-level-data/create-nifi-ingestion-job.yaml @@ -8,7 +8,7 @@ spec: spec: containers: - name: create-nifi-ingestion-job - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1 command: ["bash", "-c", "curl -O https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/nifi-kafka-druid-water-level-data/IngestWaterLevelsToKafka.xml && python -u /tmp/script/script.py"] volumeMounts: - name: script diff --git a/demos/nifi-kafka-druid-water-level-data/setup-superset.yaml b/demos/nifi-kafka-druid-water-level-data/setup-superset.yaml index fc73baa8..a6195157 100644 --- a/demos/nifi-kafka-druid-water-level-data/setup-superset.yaml +++ b/demos/nifi-kafka-druid-water-level-data/setup-superset.yaml @@ -8,7 +8,7 @@ spec: spec: containers: - name: setup-superset - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1 command: ["bash", "-c", "curl -o superset-assets.zip https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/nifi-kafka-druid-water-level-data/superset-assets.zip && python -u /tmp/script/script.py"] volumeMounts: - name: script diff --git a/demos/signal-processing/Dockerfile-nifi b/demos/signal-processing/Dockerfile-nifi index 1bd50f50..0e89c9c7 100644 --- a/demos/signal-processing/Dockerfile-nifi +++ b/demos/signal-processing/Dockerfile-nifi @@ -1,3 +1,3 @@ -FROM docker.stackable.tech/stackable/nifi:1.27.0-stackable24.11.0 +FROM docker.stackable.tech/stackable/nifi:1.27.0-stackable24.11.1 RUN curl --fail -o /stackable/nifi/postgresql-42.6.0.jar "https://repo.stackable.tech/repository/misc/postgresql-timescaledb/postgresql-42.6.0.jar" diff --git a/demos/signal-processing/create-nifi-ingestion-job.yaml b/demos/signal-processing/create-nifi-ingestion-job.yaml index 0d59c7ba..866023b0 100644 --- a/demos/signal-processing/create-nifi-ingestion-job.yaml +++ b/demos/signal-processing/create-nifi-ingestion-job.yaml @@ -9,13 +9,13 @@ spec: serviceAccountName: demo-serviceaccount initContainers: - name: wait-for-timescale-job - image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.11.0 + image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.11.1 command: ["bash", "-c", "echo 'Waiting for timescaleDB tables to be ready' && kubectl wait --for=condition=complete job/create-timescale-tables-job" ] containers: - name: create-nifi-ingestion-job - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1 command: ["bash", "-c", "export PGPASSWORD=$(cat /timescale-admin-credentials/password) && \ curl -O https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/signal-processing/DownloadAndWriteToDB.xml && \ sed -i \"s/PLACEHOLDERPGPASSWORD/$PGPASSWORD/g\" DownloadAndWriteToDB.xml && \ diff --git a/demos/signal-processing/create-timescale-tables.yaml b/demos/signal-processing/create-timescale-tables.yaml index 0804fa71..d3e2afd9 100644 --- a/demos/signal-processing/create-timescale-tables.yaml +++ b/demos/signal-processing/create-timescale-tables.yaml @@ -9,7 +9,7 @@ spec: serviceAccountName: demo-serviceaccount initContainers: - name: wait-for-timescale - image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.11.0 + image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.11.1 command: ["bash", "-c", "echo 'Waiting for timescaleDB to be ready' && kubectl wait --for=condition=ready --timeout=30m pod -l app.kubernetes.io/name=postgresql-timescaledb" ] diff --git a/demos/spark-k8s-anomaly-detection-taxi-data/create-spark-anomaly-detection-job.yaml b/demos/spark-k8s-anomaly-detection-taxi-data/create-spark-anomaly-detection-job.yaml index 3cafbec2..a8854e4c 100644 --- a/demos/spark-k8s-anomaly-detection-taxi-data/create-spark-anomaly-detection-job.yaml +++ b/demos/spark-k8s-anomaly-detection-taxi-data/create-spark-anomaly-detection-job.yaml @@ -8,11 +8,11 @@ spec: spec: initContainers: - name: wait-for-testdata - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1 command: ["bash", "-c", "echo 'Waiting for job load-ny-taxi-data to finish' && kubectl wait --for=condition=complete --timeout=30m job/load-ny-taxi-data"] containers: - name: create-spark-anomaly-detection-job - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1 command: ["bash", "-c", "echo 'Submitting Spark job' && kubectl apply -f /tmp/manifest/spark-ad-job.yaml"] volumeMounts: - name: manifest diff --git a/demos/spark-k8s-anomaly-detection-taxi-data/setup-superset.yaml b/demos/spark-k8s-anomaly-detection-taxi-data/setup-superset.yaml index a2a704f6..790a1f82 100644 --- a/demos/spark-k8s-anomaly-detection-taxi-data/setup-superset.yaml +++ b/demos/spark-k8s-anomaly-detection-taxi-data/setup-superset.yaml @@ -8,7 +8,7 @@ spec: spec: containers: - name: setup-superset - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1 command: ["bash", "-c", "curl -o superset-assets.zip https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/spark-k8s-anomaly-detection-taxi-data/superset-assets.zip && python -u /tmp/script/script.py"] volumeMounts: - name: script diff --git a/demos/trino-taxi-data/create-table-in-trino.yaml b/demos/trino-taxi-data/create-table-in-trino.yaml index 7e02c16a..64dafcdb 100644 --- a/demos/trino-taxi-data/create-table-in-trino.yaml +++ b/demos/trino-taxi-data/create-table-in-trino.yaml @@ -8,7 +8,7 @@ spec: spec: containers: - name: create-ny-taxi-data-table-in-trino - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1 command: ["bash", "-c", "python -u /tmp/script/script.py"] volumeMounts: - name: script diff --git a/demos/trino-taxi-data/setup-superset.yaml b/demos/trino-taxi-data/setup-superset.yaml index 9902ba7f..9a2933af 100644 --- a/demos/trino-taxi-data/setup-superset.yaml +++ b/demos/trino-taxi-data/setup-superset.yaml @@ -8,7 +8,7 @@ spec: spec: containers: - name: setup-superset - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1 command: ["bash", "-c", "curl -o superset-assets.zip https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/trino-taxi-data/superset-assets.zip && python -u /tmp/script/script.py"] volumeMounts: - name: script diff --git a/stacks/_templates/jupyterhub.yaml b/stacks/_templates/jupyterhub.yaml index 1939afd9..05d56a3f 100644 --- a/stacks/_templates/jupyterhub.yaml +++ b/stacks/_templates/jupyterhub.yaml @@ -50,7 +50,7 @@ options: HADOOP_CONF_DIR: "/home/jovyan/hdfs" initContainers: - name: download-notebook - image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.11.0 + image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.11.1 command: ['sh', '-c', 'curl https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/jupyterhub-pyspark-hdfs/notebook.ipynb -o /notebook/notebook.ipynb'] volumeMounts: - mountPath: /notebook diff --git a/stacks/_templates/keycloak.yaml b/stacks/_templates/keycloak.yaml index b620e50f..3f7a9cb8 100644 --- a/stacks/_templates/keycloak.yaml +++ b/stacks/_templates/keycloak.yaml @@ -48,7 +48,7 @@ spec: - name: tls mountPath: /tls/ - name: create-auth-class - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1 command: ["/bin/bash", "-c"] args: - | diff --git a/stacks/end-to-end-security/krb5.yaml b/stacks/end-to-end-security/krb5.yaml index c9dd6018..76c4635a 100644 --- a/stacks/end-to-end-security/krb5.yaml +++ b/stacks/end-to-end-security/krb5.yaml @@ -14,7 +14,7 @@ spec: spec: initContainers: - name: init - image: docker.stackable.tech/stackable/krb5:1.21.1-stackable24.11.0 + image: docker.stackable.tech/stackable/krb5:1.21.1-stackable24.11.1 args: - sh - -euo @@ -35,7 +35,7 @@ spec: name: data containers: - name: kdc - image: docker.stackable.tech/stackable/krb5:1.21.1-stackable24.11.0 + image: docker.stackable.tech/stackable/krb5:1.21.1-stackable24.11.1 args: - krb5kdc - -n @@ -48,7 +48,7 @@ spec: - mountPath: /var/kerberos/krb5kdc name: data - name: kadmind - image: docker.stackable.tech/stackable/krb5:1.21.1-stackable24.11.0 + image: docker.stackable.tech/stackable/krb5:1.21.1-stackable24.11.1 args: - kadmind - -nofork @@ -61,7 +61,7 @@ spec: - mountPath: /var/kerberos/krb5kdc name: data - name: client - image: docker.stackable.tech/stackable/krb5:1.21.1-stackable24.11.0 + image: docker.stackable.tech/stackable/krb5:1.21.1-stackable24.11.1 tty: true stdin: true env: diff --git a/stacks/end-to-end-security/superset.yaml b/stacks/end-to-end-security/superset.yaml index c80fd82f..f5ca1ff3 100644 --- a/stacks/end-to-end-security/superset.yaml +++ b/stacks/end-to-end-security/superset.yaml @@ -25,7 +25,7 @@ spec: initContainers: # The postgres image does not contain curl or wget... - name: download-dump - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1 command: - bash - -c diff --git a/stacks/keycloak-opa-poc/keycloak.yaml b/stacks/keycloak-opa-poc/keycloak.yaml index 0e484a96..3da32e87 100644 --- a/stacks/keycloak-opa-poc/keycloak.yaml +++ b/stacks/keycloak-opa-poc/keycloak.yaml @@ -70,7 +70,7 @@ spec: spec: containers: - name: propagate-keycloak-address - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1 command: - bash - -x diff --git a/stacks/keycloak-opa-poc/setup-keycloak.yaml b/stacks/keycloak-opa-poc/setup-keycloak.yaml index 499f8632..ab30c381 100644 --- a/stacks/keycloak-opa-poc/setup-keycloak.yaml +++ b/stacks/keycloak-opa-poc/setup-keycloak.yaml @@ -29,7 +29,7 @@ spec: spec: containers: - name: setup-keycloak - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1 env: - name: KEYCLOAK_ADMIN_PASSWORD valueFrom: diff --git a/stacks/logging/setup-opensearch-dashboards.yaml b/stacks/logging/setup-opensearch-dashboards.yaml index ed565227..c3632f94 100644 --- a/stacks/logging/setup-opensearch-dashboards.yaml +++ b/stacks/logging/setup-opensearch-dashboards.yaml @@ -8,7 +8,7 @@ spec: spec: containers: - name: setup-opensearch-dashboards - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.0 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1 env: - name: OPEN_SEARCH_ADMIN_PASSWORD valueFrom: diff --git a/stacks/signal-processing/jupyterhub.yaml b/stacks/signal-processing/jupyterhub.yaml index 8bf3a23d..5dc54a7f 100644 --- a/stacks/signal-processing/jupyterhub.yaml +++ b/stacks/signal-processing/jupyterhub.yaml @@ -30,7 +30,7 @@ options: singleuser: cmd: null image: - # TODO (@NickLarsenNZ): Use a versioned image with stackable24.11.0 or stackableXX.X.X so that + # TODO (@NickLarsenNZ): Use a versioned image with stackable24.11.1 or stackableXX.X.X so that # the demo is reproducable for the release and it will be automatically replaced for the release branch. name: docker.stackable.tech/demos/jupyter-pyspark-with-alibi-detect tag: python-3.9 @@ -41,7 +41,7 @@ options: stackable.tech/vendor: Stackable initContainers: - name: download-notebook - image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.11.0 + image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.11.1 command: ['sh', '-c', 'curl https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/signal-processing/tsdb.ipynb -o /notebook/tsdb.ipynb'] volumeMounts: - mountPath: /notebook diff --git a/stacks/signal-processing/nifi.yaml b/stacks/signal-processing/nifi.yaml index 5e758282..3bb28b09 100644 --- a/stacks/signal-processing/nifi.yaml +++ b/stacks/signal-processing/nifi.yaml @@ -6,7 +6,7 @@ metadata: spec: image: productVersion: 1.27.0 - # TODO (@NickLarsenNZ): Use a versioned image with stackable24.11.0 or stackableXX.X.X so that + # TODO (@NickLarsenNZ): Use a versioned image with stackable24.11.1 or stackableXX.X.X so that # the demo is reproducable for the release and it will be automatically replaced for the release branch. custom: docker.stackable.tech/demos/nifi:1.27.0-postgresql clusterConfig: From f16f7b0f714ad0878d2b59a8edaea76e4c9813e5 Mon Sep 17 00:00:00 2001 From: Nick <10092581+NickLarsenNZ@users.noreply.github.com> Date: Thu, 16 Jan 2025 14:48:16 +0100 Subject: [PATCH 08/14] fix(stack/end-to-end-security): Skip DB restore if the DB exists (#139) Otherwise it breaks with: ``` ERROR [flask_migrate] Error: Requested revision 17fcea065655 overlaps with other requested revisions b7851ee5522f ``` The latter revision being the one that exists in the uploaded dump. --- stacks/end-to-end-security/superset.yaml | 5 +++++ 1 file changed, 5 insertions(+) diff --git a/stacks/end-to-end-security/superset.yaml b/stacks/end-to-end-security/superset.yaml index f5ca1ff3..e1bb378f 100644 --- a/stacks/end-to-end-security/superset.yaml +++ b/stacks/end-to-end-security/superset.yaml @@ -46,6 +46,11 @@ spec: - bash - -c - | + if psql --host postgresql-superset --user postgres --csv -c "SELECT datname FROM pg_database where datname = 'superset' limit 1" | grep -q superset; then + # The flask app will do any necesary migrations. + echo "Skip restoring the DB as it already exists" + exit 0 + fi psql --host postgresql-superset --user postgres < /dump/postgres_superset_dump.sql env: - name: PGPASSWORD From 70849b9ce7dd2520162ee5092d43e67a5f95c9ef Mon Sep 17 00:00:00 2001 From: Andrew Kenworthy Date: Thu, 30 Jan 2025 16:14:36 +0100 Subject: [PATCH 09/14] change references back to main etc. --- .../create-nifi-ingestion-job.yaml | 2 +- .../setup-superset.yaml | 2 +- demos/demos-v2.yaml | 66 ++--- .../create-nifi-ingestion-job.yaml | 2 +- .../setup-superset.yaml | 2 +- .../create-nifi-ingestion-job.yaml | 2 +- .../setup-superset.yaml | 2 +- .../create-nifi-ingestion-job.yaml | 2 +- .../setup-superset.yaml | 2 +- demos/trino-taxi-data/setup-superset.yaml | 2 +- docs/antora.yml | 2 +- stacks/_templates/jupyterhub.yaml | 2 +- stacks/end-to-end-security/superset.yaml | 2 +- stacks/signal-processing/jupyterhub.yaml | 2 +- stacks/stacks-v2.yaml | 260 +++++++++--------- 15 files changed, 176 insertions(+), 176 deletions(-) diff --git a/demos/data-lakehouse-iceberg-trino-spark/create-nifi-ingestion-job.yaml b/demos/data-lakehouse-iceberg-trino-spark/create-nifi-ingestion-job.yaml index 98472784..173a4881 100644 --- a/demos/data-lakehouse-iceberg-trino-spark/create-nifi-ingestion-job.yaml +++ b/demos/data-lakehouse-iceberg-trino-spark/create-nifi-ingestion-job.yaml @@ -14,7 +14,7 @@ spec: containers: - name: create-nifi-ingestion-job image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1 - command: ["bash", "-c", "curl -O https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/data-lakehouse-iceberg-trino-spark/LakehouseKafkaIngest.xml && python -u /tmp/script/script.py"] + command: ["bash", "-c", "curl -O https://raw.githubusercontent.com/stackabletech/demos/main/demos/data-lakehouse-iceberg-trino-spark/LakehouseKafkaIngest.xml && python -u /tmp/script/script.py"] volumeMounts: - name: script mountPath: /tmp/script diff --git a/demos/data-lakehouse-iceberg-trino-spark/setup-superset.yaml b/demos/data-lakehouse-iceberg-trino-spark/setup-superset.yaml index 53b0bb54..f85d98f6 100644 --- a/demos/data-lakehouse-iceberg-trino-spark/setup-superset.yaml +++ b/demos/data-lakehouse-iceberg-trino-spark/setup-superset.yaml @@ -9,7 +9,7 @@ spec: containers: - name: setup-superset image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1 - command: ["bash", "-c", "curl -o superset-assets.zip https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/data-lakehouse-iceberg-trino-spark/superset-assets.zip && python -u /tmp/script/script.py"] + command: ["bash", "-c", "curl -o superset-assets.zip https://raw.githubusercontent.com/stackabletech/demos/main/demos/data-lakehouse-iceberg-trino-spark/superset-assets.zip && python -u /tmp/script/script.py"] volumeMounts: - name: script mountPath: /tmp/script diff --git a/demos/demos-v2.yaml b/demos/demos-v2.yaml index 9b52292b..a31ff1ea 100644 --- a/demos/demos-v2.yaml +++ b/demos/demos-v2.yaml @@ -7,10 +7,10 @@ demos: - airflow - job-scheduling manifests: - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/airflow-scheduled-job/01-airflow-spark-clusterrole.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/airflow-scheduled-job/02-airflow-spark-clusterrolebinding.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/airflow-scheduled-job/03-enable-and-run-spark-dag.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/airflow-scheduled-job/04-enable-and-run-date-dag.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/airflow-scheduled-job/01-airflow-spark-clusterrole.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/airflow-scheduled-job/02-airflow-spark-clusterrolebinding.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/airflow-scheduled-job/03-enable-and-run-spark-dag.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/airflow-scheduled-job/04-enable-and-run-date-dag.yaml supportedNamespaces: [] resourceRequests: cpu: 2401m @@ -24,8 +24,8 @@ demos: - hdfs - cycling-tripdata manifests: - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/hbase-hdfs-load-cycling-data/distcp-cycling-data.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/hbase-hdfs-load-cycling-data/create-hfile-and-import-to-hbase.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/hbase-hdfs-load-cycling-data/distcp-cycling-data.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/hbase-hdfs-load-cycling-data/create-hfile-and-import-to-hbase.yaml supportedNamespaces: [] resourceRequests: cpu: "3" @@ -43,9 +43,9 @@ demos: - opa - keycloak manifests: - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/end-to-end-security/create-trino-tables.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/end-to-end-security/serviceaccount.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/end-to-end-security/create-spark-report.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/end-to-end-security/create-trino-tables.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/end-to-end-security/serviceaccount.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/end-to-end-security/create-spark-report.yaml supportedNamespaces: [] resourceRequests: cpu: 9000m @@ -64,9 +64,9 @@ demos: - s3 - earthquakes manifests: - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/nifi-kafka-druid-earthquake-data/create-nifi-ingestion-job.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/nifi-kafka-druid-earthquake-data/create-druid-ingestion-job.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/nifi-kafka-druid-earthquake-data/setup-superset.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/nifi-kafka-druid-earthquake-data/create-nifi-ingestion-job.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/nifi-kafka-druid-earthquake-data/create-druid-ingestion-job.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/nifi-kafka-druid-earthquake-data/setup-superset.yaml supportedNamespaces: ["default"] resourceRequests: cpu: 8700m @@ -85,9 +85,9 @@ demos: - s3 - water-levels manifests: - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/nifi-kafka-druid-water-level-data/create-nifi-ingestion-job.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/nifi-kafka-druid-water-level-data/create-druid-ingestion-job.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/nifi-kafka-druid-water-level-data/setup-superset.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/nifi-kafka-druid-water-level-data/create-nifi-ingestion-job.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/nifi-kafka-druid-water-level-data/create-druid-ingestion-job.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/nifi-kafka-druid-water-level-data/setup-superset.yaml supportedNamespaces: ["default"] resourceRequests: cpu: 8900m @@ -104,10 +104,10 @@ demos: - s3 - ny-taxi-data manifests: - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/spark-k8s-anomaly-detection-taxi-data/serviceaccount.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/spark-k8s-anomaly-detection-taxi-data/load-test-data.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/spark-k8s-anomaly-detection-taxi-data/create-spark-anomaly-detection-job.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/spark-k8s-anomaly-detection-taxi-data/setup-superset.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/spark-k8s-anomaly-detection-taxi-data/serviceaccount.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/spark-k8s-anomaly-detection-taxi-data/load-test-data.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/spark-k8s-anomaly-detection-taxi-data/create-spark-anomaly-detection-job.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/spark-k8s-anomaly-detection-taxi-data/setup-superset.yaml supportedNamespaces: [] resourceRequests: cpu: 6400m @@ -139,9 +139,9 @@ demos: - s3 - ny-taxi-data manifests: - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/trino-taxi-data/load-test-data.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/trino-taxi-data/create-table-in-trino.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/trino-taxi-data/setup-superset.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/trino-taxi-data/load-test-data.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/trino-taxi-data/create-table-in-trino.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/trino-taxi-data/setup-superset.yaml supportedNamespaces: [] resourceRequests: cpu: 6800m @@ -164,12 +164,12 @@ demos: - water-levels - earthquakes manifests: - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/data-lakehouse-iceberg-trino-spark/serviceaccount.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/data-lakehouse-iceberg-trino-spark/load-test-data.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/data-lakehouse-iceberg-trino-spark/create-trino-tables.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/data-lakehouse-iceberg-trino-spark/create-nifi-ingestion-job.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/data-lakehouse-iceberg-trino-spark/create-spark-ingestion-job.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/data-lakehouse-iceberg-trino-spark/setup-superset.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/data-lakehouse-iceberg-trino-spark/serviceaccount.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/data-lakehouse-iceberg-trino-spark/load-test-data.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/data-lakehouse-iceberg-trino-spark/create-trino-tables.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/data-lakehouse-iceberg-trino-spark/create-nifi-ingestion-job.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/data-lakehouse-iceberg-trino-spark/create-spark-ingestion-job.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/data-lakehouse-iceberg-trino-spark/setup-superset.yaml supportedNamespaces: ["default"] resourceRequests: cpu: "80" @@ -185,7 +185,7 @@ demos: - pyspark - ny-taxi-data manifests: - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/jupyterhub-pyspark-hdfs-anomaly-detection-taxi-data/load-test-data.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/jupyterhub-pyspark-hdfs-anomaly-detection-taxi-data/load-test-data.yaml supportedNamespaces: [] resourceRequests: cpu: 3350m @@ -202,7 +202,7 @@ demos: - vector - zookeeper manifests: - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/logging/zookeeper.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/logging/zookeeper.yaml supportedNamespaces: [] resourceRequests: cpu: 6500m @@ -218,9 +218,9 @@ demos: - grafana-dashboards - zookeeper manifests: - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/signal-processing/serviceaccount.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/signal-processing/create-timescale-tables.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/signal-processing/create-nifi-ingestion-job.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/signal-processing/serviceaccount.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/signal-processing/create-timescale-tables.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/signal-processing/create-nifi-ingestion-job.yaml supportedNamespaces: [] resourceRequests: cpu: "3" diff --git a/demos/nifi-kafka-druid-earthquake-data/create-nifi-ingestion-job.yaml b/demos/nifi-kafka-druid-earthquake-data/create-nifi-ingestion-job.yaml index 0d2a1e95..d6098aad 100644 --- a/demos/nifi-kafka-druid-earthquake-data/create-nifi-ingestion-job.yaml +++ b/demos/nifi-kafka-druid-earthquake-data/create-nifi-ingestion-job.yaml @@ -9,7 +9,7 @@ spec: containers: - name: create-nifi-ingestion-job image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1 - command: ["bash", "-c", "curl -O https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/nifi-kafka-druid-earthquake-data/IngestEarthquakesToKafka.xml && python -u /tmp/script/script.py"] + command: ["bash", "-c", "curl -O https://raw.githubusercontent.com/stackabletech/demos/main/demos/nifi-kafka-druid-earthquake-data/IngestEarthquakesToKafka.xml && python -u /tmp/script/script.py"] volumeMounts: - name: script mountPath: /tmp/script diff --git a/demos/nifi-kafka-druid-earthquake-data/setup-superset.yaml b/demos/nifi-kafka-druid-earthquake-data/setup-superset.yaml index b3a72a3e..f0b1714b 100644 --- a/demos/nifi-kafka-druid-earthquake-data/setup-superset.yaml +++ b/demos/nifi-kafka-druid-earthquake-data/setup-superset.yaml @@ -9,7 +9,7 @@ spec: containers: - name: setup-superset image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1 - command: ["bash", "-c", "curl -o superset-assets.zip https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/nifi-kafka-druid-earthquake-data/superset-assets.zip && python -u /tmp/script/script.py"] + command: ["bash", "-c", "curl -o superset-assets.zip https://raw.githubusercontent.com/stackabletech/demos/main/demos/nifi-kafka-druid-earthquake-data/superset-assets.zip && python -u /tmp/script/script.py"] volumeMounts: - name: script mountPath: /tmp/script diff --git a/demos/nifi-kafka-druid-water-level-data/create-nifi-ingestion-job.yaml b/demos/nifi-kafka-druid-water-level-data/create-nifi-ingestion-job.yaml index f46f64dc..7f058d66 100644 --- a/demos/nifi-kafka-druid-water-level-data/create-nifi-ingestion-job.yaml +++ b/demos/nifi-kafka-druid-water-level-data/create-nifi-ingestion-job.yaml @@ -9,7 +9,7 @@ spec: containers: - name: create-nifi-ingestion-job image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1 - command: ["bash", "-c", "curl -O https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/nifi-kafka-druid-water-level-data/IngestWaterLevelsToKafka.xml && python -u /tmp/script/script.py"] + command: ["bash", "-c", "curl -O https://raw.githubusercontent.com/stackabletech/demos/main/demos/nifi-kafka-druid-water-level-data/IngestWaterLevelsToKafka.xml && python -u /tmp/script/script.py"] volumeMounts: - name: script mountPath: /tmp/script diff --git a/demos/nifi-kafka-druid-water-level-data/setup-superset.yaml b/demos/nifi-kafka-druid-water-level-data/setup-superset.yaml index a6195157..8a197673 100644 --- a/demos/nifi-kafka-druid-water-level-data/setup-superset.yaml +++ b/demos/nifi-kafka-druid-water-level-data/setup-superset.yaml @@ -9,7 +9,7 @@ spec: containers: - name: setup-superset image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1 - command: ["bash", "-c", "curl -o superset-assets.zip https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/nifi-kafka-druid-water-level-data/superset-assets.zip && python -u /tmp/script/script.py"] + command: ["bash", "-c", "curl -o superset-assets.zip https://raw.githubusercontent.com/stackabletech/demos/main/demos/nifi-kafka-druid-water-level-data/superset-assets.zip && python -u /tmp/script/script.py"] volumeMounts: - name: script mountPath: /tmp/script diff --git a/demos/signal-processing/create-nifi-ingestion-job.yaml b/demos/signal-processing/create-nifi-ingestion-job.yaml index 866023b0..d58a8de7 100644 --- a/demos/signal-processing/create-nifi-ingestion-job.yaml +++ b/demos/signal-processing/create-nifi-ingestion-job.yaml @@ -17,7 +17,7 @@ spec: - name: create-nifi-ingestion-job image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1 command: ["bash", "-c", "export PGPASSWORD=$(cat /timescale-admin-credentials/password) && \ - curl -O https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/signal-processing/DownloadAndWriteToDB.xml && \ + curl -O https://raw.githubusercontent.com/stackabletech/demos/main/demos/signal-processing/DownloadAndWriteToDB.xml && \ sed -i \"s/PLACEHOLDERPGPASSWORD/$PGPASSWORD/g\" DownloadAndWriteToDB.xml && \ python -u /tmp/script/script.py"] volumeMounts: diff --git a/demos/spark-k8s-anomaly-detection-taxi-data/setup-superset.yaml b/demos/spark-k8s-anomaly-detection-taxi-data/setup-superset.yaml index 790a1f82..70e25348 100644 --- a/demos/spark-k8s-anomaly-detection-taxi-data/setup-superset.yaml +++ b/demos/spark-k8s-anomaly-detection-taxi-data/setup-superset.yaml @@ -9,7 +9,7 @@ spec: containers: - name: setup-superset image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1 - command: ["bash", "-c", "curl -o superset-assets.zip https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/spark-k8s-anomaly-detection-taxi-data/superset-assets.zip && python -u /tmp/script/script.py"] + command: ["bash", "-c", "curl -o superset-assets.zip https://raw.githubusercontent.com/stackabletech/demos/main/demos/spark-k8s-anomaly-detection-taxi-data/superset-assets.zip && python -u /tmp/script/script.py"] volumeMounts: - name: script mountPath: /tmp/script diff --git a/demos/trino-taxi-data/setup-superset.yaml b/demos/trino-taxi-data/setup-superset.yaml index 9a2933af..f815e5d4 100644 --- a/demos/trino-taxi-data/setup-superset.yaml +++ b/demos/trino-taxi-data/setup-superset.yaml @@ -9,7 +9,7 @@ spec: containers: - name: setup-superset image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1 - command: ["bash", "-c", "curl -o superset-assets.zip https://raw.githubusercontent.com/stackabletech/demos/release-24.11/demos/trino-taxi-data/superset-assets.zip && python -u /tmp/script/script.py"] + command: ["bash", "-c", "curl -o superset-assets.zip https://raw.githubusercontent.com/stackabletech/demos/main/demos/trino-taxi-data/superset-assets.zip && python -u /tmp/script/script.py"] volumeMounts: - name: script mountPath: /tmp/script diff --git a/docs/antora.yml b/docs/antora.yml index 2d6b5aae..189f7d75 100644 --- a/docs/antora.yml +++ b/docs/antora.yml @@ -2,4 +2,4 @@ # Use 'home' here so that the versioning is picked up automatically based # on SDP releases name: home -version: "24.11" +version: "nightly" diff --git a/stacks/_templates/jupyterhub.yaml b/stacks/_templates/jupyterhub.yaml index 05d56a3f..86f49376 100644 --- a/stacks/_templates/jupyterhub.yaml +++ b/stacks/_templates/jupyterhub.yaml @@ -51,7 +51,7 @@ options: initContainers: - name: download-notebook image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.11.1 - command: ['sh', '-c', 'curl https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/jupyterhub-pyspark-hdfs/notebook.ipynb -o /notebook/notebook.ipynb'] + command: ['sh', '-c', 'curl https://raw.githubusercontent.com/stackabletech/demos/main/stacks/jupyterhub-pyspark-hdfs/notebook.ipynb -o /notebook/notebook.ipynb'] volumeMounts: - mountPath: /notebook name: notebook diff --git a/stacks/end-to-end-security/superset.yaml b/stacks/end-to-end-security/superset.yaml index e1bb378f..56c12d20 100644 --- a/stacks/end-to-end-security/superset.yaml +++ b/stacks/end-to-end-security/superset.yaml @@ -31,7 +31,7 @@ spec: - -c - | cd /tmp - curl --fail -O https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/end-to-end-security/postgres_superset_dump.sql.gz + curl --fail -O https://raw.githubusercontent.com/stackabletech/demos/main/stacks/end-to-end-security/postgres_superset_dump.sql.gz gunzip postgres_superset_dump.sql.gz # We need to omit changing the users password, as otherwise the content in the Secrets does not match diff --git a/stacks/signal-processing/jupyterhub.yaml b/stacks/signal-processing/jupyterhub.yaml index 5dc54a7f..edd7b4ea 100644 --- a/stacks/signal-processing/jupyterhub.yaml +++ b/stacks/signal-processing/jupyterhub.yaml @@ -42,7 +42,7 @@ options: initContainers: - name: download-notebook image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.11.1 - command: ['sh', '-c', 'curl https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/signal-processing/tsdb.ipynb -o /notebook/tsdb.ipynb'] + command: ['sh', '-c', 'curl https://raw.githubusercontent.com/stackabletech/demos/main/stacks/signal-processing/tsdb.ipynb -o /notebook/tsdb.ipynb'] volumeMounts: - mountPath: /notebook name: notebook diff --git a/stacks/stacks-v2.yaml b/stacks/stacks-v2.yaml index 21224a60..95f5ca29 100644 --- a/stacks/stacks-v2.yaml +++ b/stacks/stacks-v2.yaml @@ -2,7 +2,7 @@ stacks: monitoring: description: Stack containing Prometheus and Grafana - stackableRelease: 24.11 + stackableRelease: dev stackableOperators: - commons - listener @@ -11,9 +11,9 @@ stacks: - prometheus - grafana manifests: - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/monitoring/grafana-dashboards.yaml - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/prometheus.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/prometheus-service-monitor.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/monitoring/grafana-dashboards.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/prometheus.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/prometheus-service-monitor.yaml supportedNamespaces: [] resourceRequests: cpu: 1750m @@ -25,7 +25,7 @@ stacks: default: adminadmin logging: description: Stack containing OpenSearch, OpenSearch Dashboards (Kibana) and Vector aggregator - stackableRelease: 24.11 + stackableRelease: dev stackableOperators: - commons - listener @@ -37,11 +37,11 @@ stacks: - opensearch-dashboards - vector manifests: - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/opensearch.yaml - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/opensearch-dashboards.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/logging/setup-opensearch-dashboards.yaml - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/vector-aggregator.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/vector-aggregator-discovery.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/opensearch.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/opensearch-dashboards.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/logging/setup-opensearch-dashboards.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/vector-aggregator.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/vector-aggregator-discovery.yaml supportedNamespaces: [] resourceRequests: cpu: 5150m @@ -60,7 +60,7 @@ stacks: observability: description: >- An observability stack with auto-injection of the opentelemetry-collector sidecar to receive traces/logs/metrics via OTLP, and send them to Jaeger/Tempo/Loki. - stackableRelease: 24.11 + stackableRelease: dev stackableOperators: - commons - listener @@ -71,21 +71,21 @@ stacks: - observability - tracing manifests: - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/observability/jaeger.yaml - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/observability/opentelemetry-operator.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/observability/grafana-admin-credentials.yaml - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/observability/grafana.yaml - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/observability/grafana-tempo.yaml - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/observability/grafana-loki.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/observability/opentelemetry-collector-sidecar.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/observability/opentelemetry-collector-deployment.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/observability/jaeger.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/observability/opentelemetry-operator.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/observability/grafana-admin-credentials.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/observability/grafana.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/observability/grafana-tempo.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/observability/grafana-loki.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/observability/opentelemetry-collector-sidecar.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/observability/opentelemetry-collector-deployment.yaml parameters: - name: grafanaAdminPassword description: Password of the Grafana admin user default: adminadmin airflow: description: Stack containing Airflow scheduling platform - stackableRelease: 24.11 + stackableRelease: dev stackableOperators: - commons - listener @@ -95,9 +95,9 @@ stacks: labels: - airflow manifests: - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/postgresql-airflow.yaml - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/redis-airflow.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/airflow/airflow.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/postgresql-airflow.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/redis-airflow.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/airflow/airflow.yaml supportedNamespaces: [] resourceRequests: cpu: 3400m @@ -112,7 +112,7 @@ stacks: default: airflowSecretKey data-lakehouse-iceberg-trino-spark: description: Data lakehouse using Iceberg lakehouse on S3, Trino as query engine, Spark for streaming ingest and Superset for data visualization - stackableRelease: 24.11 + stackableRelease: dev stackableOperators: - commons - listener @@ -135,17 +135,17 @@ stacks: - minio - s3 manifests: - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/minio-distributed.yaml - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/postgresql-hive.yaml - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/postgresql-hive-iceberg.yaml - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/postgresql-superset.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/data-lakehouse-iceberg-trino-spark/s3-connection.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/data-lakehouse-iceberg-trino-spark/hive-metastores.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/data-lakehouse-iceberg-trino-spark/trino.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/data-lakehouse-iceberg-trino-spark/zookeeper.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/data-lakehouse-iceberg-trino-spark/kafka.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/data-lakehouse-iceberg-trino-spark/nifi.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/nifi-kafka-druid-superset-s3/superset.yaml # Reuse + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/minio-distributed.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/postgresql-hive.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/postgresql-hive-iceberg.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/postgresql-superset.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/data-lakehouse-iceberg-trino-spark/s3-connection.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/data-lakehouse-iceberg-trino-spark/hive-metastores.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/data-lakehouse-iceberg-trino-spark/trino.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/data-lakehouse-iceberg-trino-spark/zookeeper.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/data-lakehouse-iceberg-trino-spark/kafka.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/data-lakehouse-iceberg-trino-spark/nifi.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/nifi-kafka-druid-superset-s3/superset.yaml # Reuse supportedNamespaces: [] resourceRequests: cpu: "71" @@ -169,7 +169,7 @@ stacks: default: supersetSecretKey hdfs-hbase: description: HBase cluster using HDFS as underlying storage - stackableRelease: 24.11 + stackableRelease: dev stackableOperators: - commons - listener @@ -181,9 +181,9 @@ stacks: - hbase - hdfs manifests: - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/hdfs-hbase/zookeeper.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/hdfs-hbase/hdfs.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/hdfs-hbase/hbase.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/hdfs-hbase/zookeeper.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/hdfs-hbase/hdfs.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/hdfs-hbase/hbase.yaml supportedNamespaces: [] resourceRequests: cpu: 4200m @@ -192,7 +192,7 @@ stacks: parameters: [] nifi-kafka-druid-superset-s3: description: Stack containing NiFi, Kafka, Druid, MinIO and Superset for data visualization - stackableRelease: 24.11 + stackableRelease: dev stackableOperators: - commons - listener @@ -210,14 +210,14 @@ stacks: - minio - s3 manifests: - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/minio.yaml - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/postgresql-druid.yaml - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/postgresql-superset.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/nifi-kafka-druid-superset-s3/zookeeper.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/nifi-kafka-druid-superset-s3/kafka.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/nifi-kafka-druid-superset-s3/druid.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/nifi-kafka-druid-superset-s3/superset.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/nifi-kafka-druid-superset-s3/nifi.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/minio.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/postgresql-druid.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/postgresql-superset.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/nifi-kafka-druid-superset-s3/zookeeper.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/nifi-kafka-druid-superset-s3/kafka.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/nifi-kafka-druid-superset-s3/druid.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/nifi-kafka-druid-superset-s3/superset.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/nifi-kafka-druid-superset-s3/nifi.yaml supportedNamespaces: [] resourceRequests: cpu: 8900m @@ -238,7 +238,7 @@ stacks: default: adminadmin spark-trino-superset-s3: description: Stack containing MinIO, Trino and Superset for data visualization - stackableRelease: 24.11 + stackableRelease: dev stackableOperators: - commons - listener @@ -254,15 +254,15 @@ stacks: - minio - s3 manifests: - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/minio.yaml - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/postgresql-hive.yaml - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/postgresql-hive-iceberg.yaml - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/postgresql-superset.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/trino-superset-s3/s3-connection.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/spark-trino-superset-s3/hive-metastore.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/trino-superset-s3/trino.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/spark-trino-superset-s3/trino-prediction-catalog.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/trino-superset-s3/superset.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/minio.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/postgresql-hive.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/postgresql-hive-iceberg.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/postgresql-superset.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/trino-superset-s3/s3-connection.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/spark-trino-superset-s3/hive-metastore.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/trino-superset-s3/trino.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/spark-trino-superset-s3/trino-prediction-catalog.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/trino-superset-s3/superset.yaml supportedNamespaces: [] resourceRequests: cpu: 7100m @@ -283,7 +283,7 @@ stacks: default: supersetSecretKey trino-superset-s3: description: Stack containing MinIO, Trino and Superset for data visualization - stackableRelease: 24.11 + stackableRelease: dev stackableOperators: - commons - listener @@ -298,13 +298,13 @@ stacks: - minio - s3 manifests: - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/minio.yaml - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/postgresql-hive.yaml - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/postgresql-superset.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/trino-superset-s3/s3-connection.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/trino-superset-s3/hive-metastore.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/trino-superset-s3/trino.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/trino-superset-s3/superset.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/minio.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/postgresql-hive.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/postgresql-superset.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/trino-superset-s3/s3-connection.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/trino-superset-s3/hive-metastore.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/trino-superset-s3/trino.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/trino-superset-s3/superset.yaml supportedNamespaces: [] resourceRequests: cpu: 6800m @@ -325,7 +325,7 @@ stacks: default: supersetSecretKey trino-iceberg: description: Stack containing Trino using Apache Iceberg as a S3 data lakehouse - stackableRelease: 24.11 + stackableRelease: dev stackableOperators: - commons - listener @@ -340,11 +340,11 @@ stacks: - minio - s3 manifests: - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/minio-distributed-small.yaml - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/postgresql-hive-iceberg.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/trino-iceberg/s3-connection.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/trino-iceberg/hive-metastores.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/trino-iceberg/trino.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/minio-distributed-small.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/postgresql-hive-iceberg.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/trino-iceberg/s3-connection.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/trino-iceberg/hive-metastores.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/trino-iceberg/trino.yaml supportedNamespaces: [] resourceRequests: cpu: 6000m # Measured 5600m @@ -359,7 +359,7 @@ stacks: default: adminadmin jupyterhub-pyspark-hdfs: description: Jupyterhub with PySpark and HDFS integration - stackableRelease: 24.11 + stackableRelease: dev stackableOperators: - commons - listener @@ -372,12 +372,12 @@ stacks: - hdfs - pyspark manifests: - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/jupyterhub.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/jupyterhub-pyspark-hdfs/zookeeper.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/jupyterhub-pyspark-hdfs/hdfs.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/jupyterhub-pyspark-hdfs/serviceaccount.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/jupyterhub.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/jupyterhub-pyspark-hdfs/zookeeper.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/jupyterhub-pyspark-hdfs/hdfs.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/jupyterhub-pyspark-hdfs/serviceaccount.yaml # TODO Use patched JHub that created service for us from customer setup (ask Sebastian) - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/jupyterhub-pyspark-hdfs/spark_driver_service.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/jupyterhub-pyspark-hdfs/spark_driver_service.yaml supportedNamespaces: [] resourceRequests: cpu: 3350m @@ -389,7 +389,7 @@ stacks: default: adminadmin dual-hive-hdfs-s3: description: Dual stack Hive on HDFS and S3 for Hadoop/Hive to Trino migration - stackableRelease: 24.11 + stackableRelease: dev stackableOperators: - commons - listener @@ -405,12 +405,12 @@ stacks: - hdfs - s3 manifests: - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/postgresql-hivehdfs.yaml - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/postgresql-hives3.yaml - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/minio.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/dual-hive-hdfs-s3/hdfs.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/dual-hive-hdfs-s3/hive.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/dual-hive-hdfs-s3/trino.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/postgresql-hivehdfs.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/postgresql-hives3.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/minio.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/dual-hive-hdfs-s3/hdfs.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/dual-hive-hdfs-s3/hive.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/dual-hive-hdfs-s3/trino.yaml supportedNamespaces: [] resourceRequests: cpu: 7750m @@ -426,7 +426,7 @@ stacks: The bind user credentials are: ldapadmin:ldapadminpassword. No AuthenticationClass is configured, The AuthenticationClass is created manually in the tutorial. Use the 'openldap' Stack for an OpenLDAD with an AuthenticationClass already installed. - stackableRelease: 24.11 + stackableRelease: dev stackableOperators: - commons - listener @@ -435,7 +435,7 @@ stacks: - authentication - ldap manifests: - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/authentication/openldap-tls.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/authentication/openldap-tls.yaml supportedNamespaces: ["default"] resourceRequests: cpu: 1950m @@ -449,7 +449,7 @@ stacks: The bind user credentials are: ldapadmin:ldapadminpassword. The LDAP AuthenticationClass is called 'ldap' and the SecretClass for the bind credentials is called 'ldap-bind-credentials'. The stack already creates an appropriate Secret, so referring to the 'ldap' AuthenticationClass in your ProductCluster should be enough. - stackableRelease: 24.11 + stackableRelease: dev stackableOperators: - commons - listener @@ -458,8 +458,8 @@ stacks: - authentication - ldap manifests: - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/authentication/openldap-tls.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/authentication/openldap-tls-authenticationclass.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/authentication/openldap-tls.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/authentication/openldap-tls-authenticationclass.yaml supportedNamespaces: [] resourceRequests: cpu: 1950m @@ -475,7 +475,7 @@ stacks: 3 users are created in Keycloak: admin:adminadmin, alice:alicealice, bob:bobbob. admin and alice are admins with full authorization in Druid and Trino, bob is not authorized. This is a proof-of-concept and the mechanisms used here are subject to change. - stackableRelease: 24.11 + stackableRelease: dev stackableOperators: - commons - listener @@ -490,18 +490,18 @@ stacks: - authentication - sso manifests: - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/postgresql-superset.yaml - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/postgresql-druid.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/keycloak-opa-poc/serviceaccount.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/keycloak-opa-poc/keycloak.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/keycloak-opa-poc/setup-keycloak.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/keycloak-opa-poc/opa.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/keycloak-opa-poc/policies.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/keycloak-opa-poc/zookeeper.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/keycloak-opa-poc/hdfs.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/keycloak-opa-poc/druid.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/keycloak-opa-poc/trino.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/keycloak-opa-poc/superset.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/postgresql-superset.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/postgresql-druid.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/keycloak-opa-poc/serviceaccount.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/keycloak-opa-poc/keycloak.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/keycloak-opa-poc/setup-keycloak.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/keycloak-opa-poc/opa.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/keycloak-opa-poc/policies.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/keycloak-opa-poc/zookeeper.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/keycloak-opa-poc/hdfs.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/keycloak-opa-poc/druid.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/keycloak-opa-poc/trino.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/keycloak-opa-poc/superset.yaml supportedNamespaces: ["default"] # ClusterRoleBinding needs explicit namespace resourceRequests: cpu: 7850m @@ -541,7 +541,7 @@ stacks: Note that this stack is tightly coupled with the demo. So if you install the stack you will get demo-specific parts (such as Keycloak users or regorules). - stackableRelease: 24.11 + stackableRelease: dev stackableOperators: - commons - listener @@ -567,22 +567,22 @@ stacks: memory: 19586Mi pvc: 40Gi manifests: - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/keycloak-serviceaccount.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/end-to-end-security/keycloak-realm-config.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/keycloak.yaml - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/postgresql-hive-iceberg.yaml - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/postgresql-superset.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/end-to-end-security/krb5.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/end-to-end-security/kerberos-secretclass.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/end-to-end-security/opa.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/end-to-end-security/zookeeper.yaml # TODO: Add authentication - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/end-to-end-security/hdfs.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/end-to-end-security/hdfs-regorules.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/end-to-end-security/hive-metastore.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/end-to-end-security/trino.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/end-to-end-security/trino-regorules.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/end-to-end-security/trino-policies.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/end-to-end-security/superset.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/keycloak-serviceaccount.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/end-to-end-security/keycloak-realm-config.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/keycloak.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/postgresql-hive-iceberg.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/postgresql-superset.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/end-to-end-security/krb5.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/end-to-end-security/kerberos-secretclass.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/end-to-end-security/opa.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/end-to-end-security/zookeeper.yaml # TODO: Add authentication + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/end-to-end-security/hdfs.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/end-to-end-security/hdfs-regorules.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/end-to-end-security/hive-metastore.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/end-to-end-security/trino.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/end-to-end-security/trino-regorules.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/end-to-end-security/trino-policies.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/end-to-end-security/superset.yaml parameters: - name: keycloakAdminPassword description: Password of the Keycloak admin user @@ -611,7 +611,7 @@ stacks: signal-processing: description: >- A stack used for creating, streaming and processing in-flight data and persisting it to TimescaleDB before it is displayed in Grafana - stackableRelease: 24.11 + stackableRelease: dev stackableOperators: - commons - listener @@ -624,15 +624,15 @@ stacks: - jupyterhub - grafana manifests: - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/signal-processing/secrets.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/signal-processing/grafana-dashboards.yaml - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/signal-processing/grafana.yaml - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/_templates/postgresql-timescaledb.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/nifi-kafka-druid-superset-s3/zookeeper.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/signal-processing/nifi.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/jupyterhub-pyspark-hdfs/serviceaccount.yaml - - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/jupyterhub-pyspark-hdfs/spark_driver_service.yaml - - helmChart: https://raw.githubusercontent.com/stackabletech/demos/release-24.11/stacks/signal-processing/jupyterhub.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/signal-processing/secrets.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/signal-processing/grafana-dashboards.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/signal-processing/grafana.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/_templates/postgresql-timescaledb.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/nifi-kafka-druid-superset-s3/zookeeper.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/signal-processing/nifi.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/jupyterhub-pyspark-hdfs/serviceaccount.yaml + - plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/jupyterhub-pyspark-hdfs/spark_driver_service.yaml + - helmChart: https://raw.githubusercontent.com/stackabletech/demos/main/stacks/signal-processing/jupyterhub.yaml parameters: - name: nifiAdminPassword description: Password of the NiFI admin user From b3b7e821794163cf20da3dd78bc4c1febaa67c06 Mon Sep 17 00:00:00 2001 From: Andrew Kenworthy Date: Thu, 30 Jan 2025 16:21:21 +0100 Subject: [PATCH 10/14] prepare update_refs for next release --- .scripts/update_refs.sh | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/.scripts/update_refs.sh b/.scripts/update_refs.sh index 32cd97ed..60d3a2e6 100755 --- a/.scripts/update_refs.sh +++ b/.scripts/update_refs.sh @@ -55,8 +55,8 @@ if [[ "$CURRENT_BRANCH" == release-* ]]; then # Replace 0.0.0-dev refs with ${STACKABLE_RELEASE}.0 # TODO (@NickLarsenNZ): handle patches later, and what about release-candidates? - SEARCH='stackable(0\.0\.0-dev|24\.7\.[0-9]+|24\.11\.0)' # TODO (@NickLarsenNZ): After https://github.com/stackabletech/stackable-cockpit/issues/310, only search for 0.0.0-dev - REPLACEMENT="stackable${STACKABLE_RELEASE}.1" # TODO (@NickLarsenNZ): Be a bit smarter about patch releases. + SEARCH='stackable(0\.0\.0-dev|24\.11\.[0-9]+)' # TODO (@NickLarsenNZ): After https://github.com/stackabletech/stackable-cockpit/issues/310, only search for 0.0.0-dev + REPLACEMENT="stackable${STACKABLE_RELEASE}.0" # TODO (@NickLarsenNZ): Be a bit smarter about patch releases. MESSAGE="Update image references with $REPLACEMENT" echo "$MESSAGE" find demos stacks -type f \ From b946ccc6bd8f705efa608b4a8b72bd56e8488a0f Mon Sep 17 00:00:00 2001 From: Andrew Kenworthy Date: Thu, 30 Jan 2025 16:24:19 +0100 Subject: [PATCH 11/14] typo --- stacks/end-to-end-security/superset.yaml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/stacks/end-to-end-security/superset.yaml b/stacks/end-to-end-security/superset.yaml index 56c12d20..5431ac3d 100644 --- a/stacks/end-to-end-security/superset.yaml +++ b/stacks/end-to-end-security/superset.yaml @@ -47,7 +47,7 @@ spec: - -c - | if psql --host postgresql-superset --user postgres --csv -c "SELECT datname FROM pg_database where datname = 'superset' limit 1" | grep -q superset; then - # The flask app will do any necesary migrations. + # The flask app will do any necessary migrations. echo "Skip restoring the DB as it already exists" exit 0 fi From a35fd39c0c03f74c795cb87e734bf5248135cf9f Mon Sep 17 00:00:00 2001 From: Andrew Kenworthy Date: Thu, 30 Jan 2025 16:41:32 +0100 Subject: [PATCH 12/14] revert changes for products --- .../create-hfile-and-import-to-hbase.yaml | 2 +- .../load-test-data.yaml | 2 +- stacks/signal-processing/jupyterhub.yaml | 2 +- stacks/signal-processing/nifi.yaml | 2 +- 4 files changed, 4 insertions(+), 4 deletions(-) diff --git a/demos/hbase-hdfs-load-cycling-data/create-hfile-and-import-to-hbase.yaml b/demos/hbase-hdfs-load-cycling-data/create-hfile-and-import-to-hbase.yaml index b822d886..d250e0db 100644 --- a/demos/hbase-hdfs-load-cycling-data/create-hfile-and-import-to-hbase.yaml +++ b/demos/hbase-hdfs-load-cycling-data/create-hfile-and-import-to-hbase.yaml @@ -9,7 +9,7 @@ spec: spec: containers: - name: create-hfile-and-import-to-hbase - image: docker.stackable.tech/stackable/hbase:2.4.18-stackable24.11.1 + image: docker.stackable.tech/stackable/hbase:2.4.18-stackable0.0.0-dev env: - name: HADOOP_USER_NAME value: stackable diff --git a/demos/jupyterhub-pyspark-hdfs-anomaly-detection-taxi-data/load-test-data.yaml b/demos/jupyterhub-pyspark-hdfs-anomaly-detection-taxi-data/load-test-data.yaml index 22abb7b1..d02c5088 100644 --- a/demos/jupyterhub-pyspark-hdfs-anomaly-detection-taxi-data/load-test-data.yaml +++ b/demos/jupyterhub-pyspark-hdfs-anomaly-detection-taxi-data/load-test-data.yaml @@ -8,7 +8,7 @@ spec: spec: containers: - name: load-ny-taxi-data - image: docker.stackable.tech/stackable/hadoop:3.4.0-stackable24.11.1 + image: docker.stackable.tech/stackable/hadoop:3.4.0-stackable0.0.0-dev # yamllint disable rule:line-length command: ["bash", "-c", "/stackable/hadoop/bin/hdfs dfs -mkdir -p /ny-taxi-data/raw \ && cd /tmp \ diff --git a/stacks/signal-processing/jupyterhub.yaml b/stacks/signal-processing/jupyterhub.yaml index edd7b4ea..1841e6c6 100644 --- a/stacks/signal-processing/jupyterhub.yaml +++ b/stacks/signal-processing/jupyterhub.yaml @@ -30,7 +30,7 @@ options: singleuser: cmd: null image: - # TODO (@NickLarsenNZ): Use a versioned image with stackable24.11.1 or stackableXX.X.X so that + # TODO (@NickLarsenNZ): Use a versioned image with stackable0.0.0-dev or stackableXX.X.X so that # the demo is reproducable for the release and it will be automatically replaced for the release branch. name: docker.stackable.tech/demos/jupyter-pyspark-with-alibi-detect tag: python-3.9 diff --git a/stacks/signal-processing/nifi.yaml b/stacks/signal-processing/nifi.yaml index 3bb28b09..48d7c39d 100644 --- a/stacks/signal-processing/nifi.yaml +++ b/stacks/signal-processing/nifi.yaml @@ -6,7 +6,7 @@ metadata: spec: image: productVersion: 1.27.0 - # TODO (@NickLarsenNZ): Use a versioned image with stackable24.11.1 or stackableXX.X.X so that + # TODO (@NickLarsenNZ): Use a versioned image with stackable0.0.0-dev or stackableXX.X.X so that # the demo is reproducable for the release and it will be automatically replaced for the release branch. custom: docker.stackable.tech/demos/nifi:1.27.0-postgresql clusterConfig: From 71317c83e5d99325480c2b96a14e448e5cae36b9 Mon Sep 17 00:00:00 2001 From: Andrew Kenworthy Date: Thu, 30 Jan 2025 17:06:30 +0100 Subject: [PATCH 13/14] set images back to dev except 24.3.0 exceptions --- .../03-enable-and-run-spark-dag.yaml | 2 +- .../airflow-scheduled-job/04-enable-and-run-date-dag.yaml | 2 +- .../create-nifi-ingestion-job.yaml | 4 ++-- .../create-spark-ingestion-job.yaml | 4 ++-- .../create-trino-tables.yaml | 4 ++-- .../setup-superset.yaml | 2 +- demos/end-to-end-security/create-spark-report.yaml | 4 ++-- demos/end-to-end-security/create-trino-tables.yaml | 2 +- .../create-druid-ingestion-job.yaml | 2 +- .../create-nifi-ingestion-job.yaml | 2 +- .../nifi-kafka-druid-earthquake-data/setup-superset.yaml | 2 +- .../create-druid-ingestion-job.yaml | 2 +- .../create-nifi-ingestion-job.yaml | 2 +- .../nifi-kafka-druid-water-level-data/setup-superset.yaml | 2 +- demos/signal-processing/Dockerfile-nifi | 2 +- demos/signal-processing/create-nifi-ingestion-job.yaml | 4 ++-- demos/signal-processing/create-timescale-tables.yaml | 2 +- .../create-spark-anomaly-detection-job.yaml | 4 ++-- .../setup-superset.yaml | 2 +- demos/trino-taxi-data/create-table-in-trino.yaml | 2 +- demos/trino-taxi-data/setup-superset.yaml | 2 +- stacks/_templates/jupyterhub.yaml | 2 +- stacks/_templates/keycloak.yaml | 2 +- stacks/end-to-end-security/krb5.yaml | 8 ++++---- stacks/end-to-end-security/superset.yaml | 2 +- stacks/keycloak-opa-poc/keycloak.yaml | 2 +- stacks/keycloak-opa-poc/setup-keycloak.yaml | 2 +- stacks/logging/setup-opensearch-dashboards.yaml | 2 +- stacks/signal-processing/jupyterhub.yaml | 2 +- 29 files changed, 38 insertions(+), 38 deletions(-) diff --git a/demos/airflow-scheduled-job/03-enable-and-run-spark-dag.yaml b/demos/airflow-scheduled-job/03-enable-and-run-spark-dag.yaml index 46e2760f..efa281b2 100644 --- a/demos/airflow-scheduled-job/03-enable-and-run-spark-dag.yaml +++ b/demos/airflow-scheduled-job/03-enable-and-run-spark-dag.yaml @@ -8,7 +8,7 @@ spec: spec: containers: - name: start-pyspark-job - image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.11.1 + image: docker.stackable.tech/stackable/tools:1.0.0-stackable0.0.0-dev # N.B. it is possible for the scheduler to report that a DAG exists, only for the worker task to fail if a pod is unexpectedly # restarted. Additionally, the db-init job takes a few minutes to complete before the cluster is deployed. The wait/watch steps # below are not "water-tight" but add a layer of stability by at least ensuring that the db is initialized and ready and that diff --git a/demos/airflow-scheduled-job/04-enable-and-run-date-dag.yaml b/demos/airflow-scheduled-job/04-enable-and-run-date-dag.yaml index 84c42550..f74bab50 100644 --- a/demos/airflow-scheduled-job/04-enable-and-run-date-dag.yaml +++ b/demos/airflow-scheduled-job/04-enable-and-run-date-dag.yaml @@ -8,7 +8,7 @@ spec: spec: containers: - name: start-date-job - image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.11.1 + image: docker.stackable.tech/stackable/tools:1.0.0-stackable0.0.0-dev # N.B. it is possible for the scheduler to report that a DAG exists, only for the worker task to fail if a pod is unexpectedly # restarted. Additionally, the db-init job takes a few minutes to complete before the cluster is deployed. The wait/watch steps # below are not "water-tight" but add a layer of stability by at least ensuring that the db is initialized and ready and that diff --git a/demos/data-lakehouse-iceberg-trino-spark/create-nifi-ingestion-job.yaml b/demos/data-lakehouse-iceberg-trino-spark/create-nifi-ingestion-job.yaml index 173a4881..6bcc924d 100644 --- a/demos/data-lakehouse-iceberg-trino-spark/create-nifi-ingestion-job.yaml +++ b/demos/data-lakehouse-iceberg-trino-spark/create-nifi-ingestion-job.yaml @@ -9,11 +9,11 @@ spec: serviceAccountName: demo-serviceaccount initContainers: - name: wait-for-kafka - image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.11.1 + image: docker.stackable.tech/stackable/tools:1.0.0-stackable0.0.0-dev command: ["bash", "-c", "echo 'Waiting for all kafka brokers to be ready' && kubectl wait --for=condition=ready --timeout=30m pod -l app.kubernetes.io/instance=kafka -l app.kubernetes.io/name=kafka"] containers: - name: create-nifi-ingestion-job - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable0.0.0-dev command: ["bash", "-c", "curl -O https://raw.githubusercontent.com/stackabletech/demos/main/demos/data-lakehouse-iceberg-trino-spark/LakehouseKafkaIngest.xml && python -u /tmp/script/script.py"] volumeMounts: - name: script diff --git a/demos/data-lakehouse-iceberg-trino-spark/create-spark-ingestion-job.yaml b/demos/data-lakehouse-iceberg-trino-spark/create-spark-ingestion-job.yaml index 526a8867..db28afc8 100644 --- a/demos/data-lakehouse-iceberg-trino-spark/create-spark-ingestion-job.yaml +++ b/demos/data-lakehouse-iceberg-trino-spark/create-spark-ingestion-job.yaml @@ -12,11 +12,11 @@ spec: serviceAccountName: demo-serviceaccount initContainers: - name: wait-for-kafka - image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.11.1 + image: docker.stackable.tech/stackable/tools:1.0.0-stackable0.0.0-dev command: ["bash", "-c", "echo 'Waiting for all kafka brokers to be ready' && kubectl wait --for=condition=ready --timeout=30m pod -l app.kubernetes.io/name=kafka -l app.kubernetes.io/instance=kafka"] containers: - name: create-spark-ingestion-job - image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.11.1 + image: docker.stackable.tech/stackable/tools:1.0.0-stackable0.0.0-dev command: ["bash", "-c", "echo 'Submitting Spark job' && kubectl apply -f /tmp/manifest/spark-ingestion-job.yaml"] volumeMounts: - name: manifest diff --git a/demos/data-lakehouse-iceberg-trino-spark/create-trino-tables.yaml b/demos/data-lakehouse-iceberg-trino-spark/create-trino-tables.yaml index 4a8a2e5d..6c0317bb 100644 --- a/demos/data-lakehouse-iceberg-trino-spark/create-trino-tables.yaml +++ b/demos/data-lakehouse-iceberg-trino-spark/create-trino-tables.yaml @@ -9,11 +9,11 @@ spec: serviceAccountName: demo-serviceaccount initContainers: - name: wait-for-testdata - image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.11.1 + image: docker.stackable.tech/stackable/tools:1.0.0-stackable0.0.0-dev command: ["bash", "-c", "echo 'Waiting for job load-test-data to finish' && kubectl wait --for=condition=complete --timeout=30m job/load-test-data"] containers: - name: create-tables-in-trino - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable0.0.0-dev command: ["bash", "-c", "python -u /tmp/script/script.py"] volumeMounts: - name: script diff --git a/demos/data-lakehouse-iceberg-trino-spark/setup-superset.yaml b/demos/data-lakehouse-iceberg-trino-spark/setup-superset.yaml index f85d98f6..1f70f900 100644 --- a/demos/data-lakehouse-iceberg-trino-spark/setup-superset.yaml +++ b/demos/data-lakehouse-iceberg-trino-spark/setup-superset.yaml @@ -8,7 +8,7 @@ spec: spec: containers: - name: setup-superset - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable0.0.0-dev command: ["bash", "-c", "curl -o superset-assets.zip https://raw.githubusercontent.com/stackabletech/demos/main/demos/data-lakehouse-iceberg-trino-spark/superset-assets.zip && python -u /tmp/script/script.py"] volumeMounts: - name: script diff --git a/demos/end-to-end-security/create-spark-report.yaml b/demos/end-to-end-security/create-spark-report.yaml index ded96dea..2e1bff2d 100644 --- a/demos/end-to-end-security/create-spark-report.yaml +++ b/demos/end-to-end-security/create-spark-report.yaml @@ -12,7 +12,7 @@ spec: serviceAccountName: demo-serviceaccount initContainers: - name: wait-for-trino-tables - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable0.0.0-dev command: - bash - -euo @@ -23,7 +23,7 @@ spec: kubectl wait --timeout=30m --for=condition=complete job/create-tables-in-trino containers: - name: create-spark-report - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable0.0.0-dev command: - bash - -euo diff --git a/demos/end-to-end-security/create-trino-tables.yaml b/demos/end-to-end-security/create-trino-tables.yaml index 469d8925..16b0e6f1 100644 --- a/demos/end-to-end-security/create-trino-tables.yaml +++ b/demos/end-to-end-security/create-trino-tables.yaml @@ -8,7 +8,7 @@ spec: spec: containers: - name: create-tables-in-trino - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable0.0.0-dev command: ["bash", "-c", "python -u /tmp/script/script.py"] volumeMounts: - name: script diff --git a/demos/nifi-kafka-druid-earthquake-data/create-druid-ingestion-job.yaml b/demos/nifi-kafka-druid-earthquake-data/create-druid-ingestion-job.yaml index 82dd92b8..93bedf98 100644 --- a/demos/nifi-kafka-druid-earthquake-data/create-druid-ingestion-job.yaml +++ b/demos/nifi-kafka-druid-earthquake-data/create-druid-ingestion-job.yaml @@ -8,7 +8,7 @@ spec: spec: containers: - name: create-druid-ingestion-job - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable0.0.0-dev command: ["bash", "-c", "curl -X POST --insecure -H 'Content-Type: application/json' -d @/tmp/ingestion-job-spec/ingestion-job-spec.json https://druid-coordinator:8281/druid/indexer/v1/supervisor"] volumeMounts: - name: ingestion-job-spec diff --git a/demos/nifi-kafka-druid-earthquake-data/create-nifi-ingestion-job.yaml b/demos/nifi-kafka-druid-earthquake-data/create-nifi-ingestion-job.yaml index d6098aad..0d1aa1cc 100644 --- a/demos/nifi-kafka-druid-earthquake-data/create-nifi-ingestion-job.yaml +++ b/demos/nifi-kafka-druid-earthquake-data/create-nifi-ingestion-job.yaml @@ -8,7 +8,7 @@ spec: spec: containers: - name: create-nifi-ingestion-job - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable0.0.0-dev command: ["bash", "-c", "curl -O https://raw.githubusercontent.com/stackabletech/demos/main/demos/nifi-kafka-druid-earthquake-data/IngestEarthquakesToKafka.xml && python -u /tmp/script/script.py"] volumeMounts: - name: script diff --git a/demos/nifi-kafka-druid-earthquake-data/setup-superset.yaml b/demos/nifi-kafka-druid-earthquake-data/setup-superset.yaml index f0b1714b..36bc2110 100644 --- a/demos/nifi-kafka-druid-earthquake-data/setup-superset.yaml +++ b/demos/nifi-kafka-druid-earthquake-data/setup-superset.yaml @@ -8,7 +8,7 @@ spec: spec: containers: - name: setup-superset - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable0.0.0-dev command: ["bash", "-c", "curl -o superset-assets.zip https://raw.githubusercontent.com/stackabletech/demos/main/demos/nifi-kafka-druid-earthquake-data/superset-assets.zip && python -u /tmp/script/script.py"] volumeMounts: - name: script diff --git a/demos/nifi-kafka-druid-water-level-data/create-druid-ingestion-job.yaml b/demos/nifi-kafka-druid-water-level-data/create-druid-ingestion-job.yaml index ce8a0a28..88b8fea5 100644 --- a/demos/nifi-kafka-druid-water-level-data/create-druid-ingestion-job.yaml +++ b/demos/nifi-kafka-druid-water-level-data/create-druid-ingestion-job.yaml @@ -8,7 +8,7 @@ spec: spec: containers: - name: create-druid-ingestion-job - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable0.0.0-dev command: ["bash", "-c", "curl -X POST --insecure -H 'Content-Type: application/json' -d @/tmp/ingestion-job-spec/stations-ingestion-job-spec.json https://druid-coordinator:8281/druid/indexer/v1/supervisor && curl -X POST --insecure -H 'Content-Type: application/json' -d @/tmp/ingestion-job-spec/measurements-ingestion-job-spec.json https://druid-coordinator:8281/druid/indexer/v1/supervisor && curl -X POST --insecure -H 'Content-Type: application/json' -d @/tmp/ingestion-job-spec/measurements-compaction-job-spec.json https://druid-coordinator:8281/druid/coordinator/v1/config/compaction"] volumeMounts: - name: ingestion-job-spec diff --git a/demos/nifi-kafka-druid-water-level-data/create-nifi-ingestion-job.yaml b/demos/nifi-kafka-druid-water-level-data/create-nifi-ingestion-job.yaml index 7f058d66..ff809981 100644 --- a/demos/nifi-kafka-druid-water-level-data/create-nifi-ingestion-job.yaml +++ b/demos/nifi-kafka-druid-water-level-data/create-nifi-ingestion-job.yaml @@ -8,7 +8,7 @@ spec: spec: containers: - name: create-nifi-ingestion-job - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable0.0.0-dev command: ["bash", "-c", "curl -O https://raw.githubusercontent.com/stackabletech/demos/main/demos/nifi-kafka-druid-water-level-data/IngestWaterLevelsToKafka.xml && python -u /tmp/script/script.py"] volumeMounts: - name: script diff --git a/demos/nifi-kafka-druid-water-level-data/setup-superset.yaml b/demos/nifi-kafka-druid-water-level-data/setup-superset.yaml index 8a197673..66f3ef56 100644 --- a/demos/nifi-kafka-druid-water-level-data/setup-superset.yaml +++ b/demos/nifi-kafka-druid-water-level-data/setup-superset.yaml @@ -8,7 +8,7 @@ spec: spec: containers: - name: setup-superset - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable0.0.0-dev command: ["bash", "-c", "curl -o superset-assets.zip https://raw.githubusercontent.com/stackabletech/demos/main/demos/nifi-kafka-druid-water-level-data/superset-assets.zip && python -u /tmp/script/script.py"] volumeMounts: - name: script diff --git a/demos/signal-processing/Dockerfile-nifi b/demos/signal-processing/Dockerfile-nifi index 0e89c9c7..eb242ad3 100644 --- a/demos/signal-processing/Dockerfile-nifi +++ b/demos/signal-processing/Dockerfile-nifi @@ -1,3 +1,3 @@ -FROM docker.stackable.tech/stackable/nifi:1.27.0-stackable24.11.1 +FROM docker.stackable.tech/stackable/nifi:1.27.0-stackable0.0.0-dev RUN curl --fail -o /stackable/nifi/postgresql-42.6.0.jar "https://repo.stackable.tech/repository/misc/postgresql-timescaledb/postgresql-42.6.0.jar" diff --git a/demos/signal-processing/create-nifi-ingestion-job.yaml b/demos/signal-processing/create-nifi-ingestion-job.yaml index d58a8de7..f63712bc 100644 --- a/demos/signal-processing/create-nifi-ingestion-job.yaml +++ b/demos/signal-processing/create-nifi-ingestion-job.yaml @@ -9,13 +9,13 @@ spec: serviceAccountName: demo-serviceaccount initContainers: - name: wait-for-timescale-job - image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.11.1 + image: docker.stackable.tech/stackable/tools:1.0.0-stackable0.0.0-dev command: ["bash", "-c", "echo 'Waiting for timescaleDB tables to be ready' && kubectl wait --for=condition=complete job/create-timescale-tables-job" ] containers: - name: create-nifi-ingestion-job - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable0.0.0-dev command: ["bash", "-c", "export PGPASSWORD=$(cat /timescale-admin-credentials/password) && \ curl -O https://raw.githubusercontent.com/stackabletech/demos/main/demos/signal-processing/DownloadAndWriteToDB.xml && \ sed -i \"s/PLACEHOLDERPGPASSWORD/$PGPASSWORD/g\" DownloadAndWriteToDB.xml && \ diff --git a/demos/signal-processing/create-timescale-tables.yaml b/demos/signal-processing/create-timescale-tables.yaml index d3e2afd9..173639b0 100644 --- a/demos/signal-processing/create-timescale-tables.yaml +++ b/demos/signal-processing/create-timescale-tables.yaml @@ -9,7 +9,7 @@ spec: serviceAccountName: demo-serviceaccount initContainers: - name: wait-for-timescale - image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.11.1 + image: docker.stackable.tech/stackable/tools:1.0.0-stackable0.0.0-dev command: ["bash", "-c", "echo 'Waiting for timescaleDB to be ready' && kubectl wait --for=condition=ready --timeout=30m pod -l app.kubernetes.io/name=postgresql-timescaledb" ] diff --git a/demos/spark-k8s-anomaly-detection-taxi-data/create-spark-anomaly-detection-job.yaml b/demos/spark-k8s-anomaly-detection-taxi-data/create-spark-anomaly-detection-job.yaml index a8854e4c..a4b46e46 100644 --- a/demos/spark-k8s-anomaly-detection-taxi-data/create-spark-anomaly-detection-job.yaml +++ b/demos/spark-k8s-anomaly-detection-taxi-data/create-spark-anomaly-detection-job.yaml @@ -8,11 +8,11 @@ spec: spec: initContainers: - name: wait-for-testdata - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable0.0.0-dev command: ["bash", "-c", "echo 'Waiting for job load-ny-taxi-data to finish' && kubectl wait --for=condition=complete --timeout=30m job/load-ny-taxi-data"] containers: - name: create-spark-anomaly-detection-job - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable0.0.0-dev command: ["bash", "-c", "echo 'Submitting Spark job' && kubectl apply -f /tmp/manifest/spark-ad-job.yaml"] volumeMounts: - name: manifest diff --git a/demos/spark-k8s-anomaly-detection-taxi-data/setup-superset.yaml b/demos/spark-k8s-anomaly-detection-taxi-data/setup-superset.yaml index 70e25348..8a1a8cbc 100644 --- a/demos/spark-k8s-anomaly-detection-taxi-data/setup-superset.yaml +++ b/demos/spark-k8s-anomaly-detection-taxi-data/setup-superset.yaml @@ -8,7 +8,7 @@ spec: spec: containers: - name: setup-superset - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable0.0.0-dev command: ["bash", "-c", "curl -o superset-assets.zip https://raw.githubusercontent.com/stackabletech/demos/main/demos/spark-k8s-anomaly-detection-taxi-data/superset-assets.zip && python -u /tmp/script/script.py"] volumeMounts: - name: script diff --git a/demos/trino-taxi-data/create-table-in-trino.yaml b/demos/trino-taxi-data/create-table-in-trino.yaml index 64dafcdb..44e2500e 100644 --- a/demos/trino-taxi-data/create-table-in-trino.yaml +++ b/demos/trino-taxi-data/create-table-in-trino.yaml @@ -8,7 +8,7 @@ spec: spec: containers: - name: create-ny-taxi-data-table-in-trino - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable0.0.0-dev command: ["bash", "-c", "python -u /tmp/script/script.py"] volumeMounts: - name: script diff --git a/demos/trino-taxi-data/setup-superset.yaml b/demos/trino-taxi-data/setup-superset.yaml index f815e5d4..792b459b 100644 --- a/demos/trino-taxi-data/setup-superset.yaml +++ b/demos/trino-taxi-data/setup-superset.yaml @@ -8,7 +8,7 @@ spec: spec: containers: - name: setup-superset - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable0.0.0-dev command: ["bash", "-c", "curl -o superset-assets.zip https://raw.githubusercontent.com/stackabletech/demos/main/demos/trino-taxi-data/superset-assets.zip && python -u /tmp/script/script.py"] volumeMounts: - name: script diff --git a/stacks/_templates/jupyterhub.yaml b/stacks/_templates/jupyterhub.yaml index 86f49376..05a28654 100644 --- a/stacks/_templates/jupyterhub.yaml +++ b/stacks/_templates/jupyterhub.yaml @@ -50,7 +50,7 @@ options: HADOOP_CONF_DIR: "/home/jovyan/hdfs" initContainers: - name: download-notebook - image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.11.1 + image: docker.stackable.tech/stackable/tools:1.0.0-stackable0.0.0-dev command: ['sh', '-c', 'curl https://raw.githubusercontent.com/stackabletech/demos/main/stacks/jupyterhub-pyspark-hdfs/notebook.ipynb -o /notebook/notebook.ipynb'] volumeMounts: - mountPath: /notebook diff --git a/stacks/_templates/keycloak.yaml b/stacks/_templates/keycloak.yaml index 3f7a9cb8..345cf15a 100644 --- a/stacks/_templates/keycloak.yaml +++ b/stacks/_templates/keycloak.yaml @@ -48,7 +48,7 @@ spec: - name: tls mountPath: /tls/ - name: create-auth-class - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable0.0.0-dev command: ["/bin/bash", "-c"] args: - | diff --git a/stacks/end-to-end-security/krb5.yaml b/stacks/end-to-end-security/krb5.yaml index 76c4635a..3ce20bd4 100644 --- a/stacks/end-to-end-security/krb5.yaml +++ b/stacks/end-to-end-security/krb5.yaml @@ -14,7 +14,7 @@ spec: spec: initContainers: - name: init - image: docker.stackable.tech/stackable/krb5:1.21.1-stackable24.11.1 + image: docker.stackable.tech/stackable/krb5:1.21.1-stackable0.0.0-dev args: - sh - -euo @@ -35,7 +35,7 @@ spec: name: data containers: - name: kdc - image: docker.stackable.tech/stackable/krb5:1.21.1-stackable24.11.1 + image: docker.stackable.tech/stackable/krb5:1.21.1-stackable0.0.0-dev args: - krb5kdc - -n @@ -48,7 +48,7 @@ spec: - mountPath: /var/kerberos/krb5kdc name: data - name: kadmind - image: docker.stackable.tech/stackable/krb5:1.21.1-stackable24.11.1 + image: docker.stackable.tech/stackable/krb5:1.21.1-stackable0.0.0-dev args: - kadmind - -nofork @@ -61,7 +61,7 @@ spec: - mountPath: /var/kerberos/krb5kdc name: data - name: client - image: docker.stackable.tech/stackable/krb5:1.21.1-stackable24.11.1 + image: docker.stackable.tech/stackable/krb5:1.21.1-stackable0.0.0-dev tty: true stdin: true env: diff --git a/stacks/end-to-end-security/superset.yaml b/stacks/end-to-end-security/superset.yaml index 5431ac3d..c606075c 100644 --- a/stacks/end-to-end-security/superset.yaml +++ b/stacks/end-to-end-security/superset.yaml @@ -25,7 +25,7 @@ spec: initContainers: # The postgres image does not contain curl or wget... - name: download-dump - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable0.0.0-dev command: - bash - -c diff --git a/stacks/keycloak-opa-poc/keycloak.yaml b/stacks/keycloak-opa-poc/keycloak.yaml index 3da32e87..eaf5d60b 100644 --- a/stacks/keycloak-opa-poc/keycloak.yaml +++ b/stacks/keycloak-opa-poc/keycloak.yaml @@ -70,7 +70,7 @@ spec: spec: containers: - name: propagate-keycloak-address - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable0.0.0-dev command: - bash - -x diff --git a/stacks/keycloak-opa-poc/setup-keycloak.yaml b/stacks/keycloak-opa-poc/setup-keycloak.yaml index ab30c381..ceba42fe 100644 --- a/stacks/keycloak-opa-poc/setup-keycloak.yaml +++ b/stacks/keycloak-opa-poc/setup-keycloak.yaml @@ -29,7 +29,7 @@ spec: spec: containers: - name: setup-keycloak - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable0.0.0-dev env: - name: KEYCLOAK_ADMIN_PASSWORD valueFrom: diff --git a/stacks/logging/setup-opensearch-dashboards.yaml b/stacks/logging/setup-opensearch-dashboards.yaml index c3632f94..b7e00641 100644 --- a/stacks/logging/setup-opensearch-dashboards.yaml +++ b/stacks/logging/setup-opensearch-dashboards.yaml @@ -8,7 +8,7 @@ spec: spec: containers: - name: setup-opensearch-dashboards - image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable24.11.1 + image: docker.stackable.tech/stackable/testing-tools:0.2.0-stackable0.0.0-dev env: - name: OPEN_SEARCH_ADMIN_PASSWORD valueFrom: diff --git a/stacks/signal-processing/jupyterhub.yaml b/stacks/signal-processing/jupyterhub.yaml index 1841e6c6..f9782df5 100644 --- a/stacks/signal-processing/jupyterhub.yaml +++ b/stacks/signal-processing/jupyterhub.yaml @@ -41,7 +41,7 @@ options: stackable.tech/vendor: Stackable initContainers: - name: download-notebook - image: docker.stackable.tech/stackable/tools:1.0.0-stackable24.11.1 + image: docker.stackable.tech/stackable/tools:1.0.0-stackable0.0.0-dev command: ['sh', '-c', 'curl https://raw.githubusercontent.com/stackabletech/demos/main/stacks/signal-processing/tsdb.ipynb -o /notebook/tsdb.ipynb'] volumeMounts: - mountPath: /notebook From e4bbc00e32eb9321fad710156ccf7c1972aa4655 Mon Sep 17 00:00:00 2001 From: Andrew Kenworthy <1712947+adwk67@users.noreply.github.com> Date: Thu, 30 Jan 2025 17:12:16 +0100 Subject: [PATCH 14/14] Update .scripts/update_refs.sh Co-authored-by: Nick <10092581+NickLarsenNZ@users.noreply.github.com> --- .scripts/update_refs.sh | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/.scripts/update_refs.sh b/.scripts/update_refs.sh index 60d3a2e6..6b2f57b2 100755 --- a/.scripts/update_refs.sh +++ b/.scripts/update_refs.sh @@ -55,7 +55,7 @@ if [[ "$CURRENT_BRANCH" == release-* ]]; then # Replace 0.0.0-dev refs with ${STACKABLE_RELEASE}.0 # TODO (@NickLarsenNZ): handle patches later, and what about release-candidates? - SEARCH='stackable(0\.0\.0-dev|24\.11\.[0-9]+)' # TODO (@NickLarsenNZ): After https://github.com/stackabletech/stackable-cockpit/issues/310, only search for 0.0.0-dev + SEARCH='stackable(0\.0\.0-dev)' REPLACEMENT="stackable${STACKABLE_RELEASE}.0" # TODO (@NickLarsenNZ): Be a bit smarter about patch releases. MESSAGE="Update image references with $REPLACEMENT" echo "$MESSAGE"