Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

feat: Update demos for Release 24.11 #116

Merged
merged 35 commits into from
Nov 11, 2024
Merged
Show file tree
Hide file tree
Changes from 28 commits
Commits
Show all changes
35 commits
Select commit Hold shift + click to select a range
42c38c1
chore(opa): bump to 0.67.0
NickLarsenNZ Aug 5, 2024
28afbe8
chore(opa): bump 0.67.1
NickLarsenNZ Aug 6, 2024
d3bec37
chore(opa): bump to 0.67.0
NickLarsenNZ Aug 5, 2024
7bf3c33
chore: Bump Vector to 0.40.0
NickLarsenNZ Aug 8, 2024
d80bd75
Merge pull request #87 from stackabletech/vector-0.40.0
NickLarsenNZ Aug 8, 2024
09e1df6
Merge remote-tracking branch 'origin/next' into opa-0.67.0
NickLarsenNZ Aug 8, 2024
94aff00
Merge pull request #85 from stackabletech/opa-0.67.0
NickLarsenNZ Aug 8, 2024
1f6e2bf
Merge remote-tracking branch 'origin/main' into next
NickLarsenNZ Aug 26, 2024
890df16
chore(airflow): bump version to 2.9.3
NickLarsenNZ Aug 26, 2024
1a1377c
Merge pull request #90 from stackabletech/airflow-2.9.3
NickLarsenNZ Sep 4, 2024
e5da07d
chore(trino): bump version to 455
xeniape Sep 5, 2024
599ac41
Merge pull request #93 from stackabletech/trino-455
xeniape Sep 5, 2024
a53b25b
chore(spark): version 3.5.2
razvan Sep 11, 2024
972eafd
Update create-spark-ingestion-job.yaml
razvan Sep 12, 2024
62516a1
Merge pull request #96 from stackabletech/chore/spark-3.5.2
razvan Sep 12, 2024
af5654b
chore: bump hdfs version to 3.4.0
razvan Sep 13, 2024
d5b4b25
chore: bump nifi for 24.11
razvan Sep 18, 2024
7f4565b
chore: Upgrade the Vector aggregator
siegfriedweber Sep 18, 2024
a1e727a
Merge pull request #101 from stackabletech/chore/vector-24.11
siegfriedweber Sep 18, 2024
0c7d349
Merge remote-tracking branch 'origin/main' into next
NickLarsenNZ Sep 19, 2024
8fbf391
Merge remote-tracking branch 'origin/next' into chore/nifi-24.11
NickLarsenNZ Sep 19, 2024
e5b02d7
Merge pull request #100 from stackabletech/chore/nifi-24.11
razvan Sep 19, 2024
db8cd8a
Merge pull request #98 from stackabletech/chore/hdfs-version-for-24.11
razvan Sep 20, 2024
9b5f892
chore(druid): Bump to 30.0.0
NickLarsenNZ Sep 27, 2024
25cb8a3
chore(pre-commit): Ignore braces lint for templates yaml files
NickLarsenNZ Sep 27, 2024
958f174
Merge pull request #102 from stackabletech/lts-druid-30.0.0
NickLarsenNZ Sep 27, 2024
5bc2356
Merge branch 'main' into next
NickLarsenNZ Oct 24, 2024
b2aa661
chore: lint
NickLarsenNZ Oct 24, 2024
6dec8e5
Apply suggestions from code review
NickLarsenNZ Nov 11, 2024
de3ca41
Merge remote-tracking branch 'origin/main' into next
NickLarsenNZ Nov 11, 2024
4680499
Apply suggestions from code review
NickLarsenNZ Nov 11, 2024
ec104d5
chore: fix lint issue
NickLarsenNZ Nov 11, 2024
7649d89
Apply suggestions from code review
NickLarsenNZ Nov 11, 2024
cc89f67
chore: fix lint issue
NickLarsenNZ Nov 11, 2024
97a3f5e
chore: disable the brace lint
NickLarsenNZ Nov 11, 2024
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
1 change: 1 addition & 0 deletions .yamllint.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -7,3 +7,4 @@ rules:
check-keys: false
comments:
min-spaces-from-content: 1 # Needed due to https://github.com/adrienverge/yamllint/issues/443
braces: disable # because the yaml files are templates which can have {{ ... }}
Original file line number Diff line number Diff line change
Expand Up @@ -43,13 +43,13 @@ data:
stackable.tech/vendor: Stackable
spec:
sparkImage:
productVersion: 3.5.1
productVersion: 3.5.2
mode: cluster
mainApplicationFile: local:///stackable/spark/jobs/spark-ingest-into-lakehouse.py
deps:
packages:
- org.apache.iceberg:iceberg-spark-runtime-3.5_2.12:1.5.0
- org.apache.spark:spark-sql-kafka-0-10_2.12:3.5.1
- org.apache.iceberg:iceberg-spark-runtime-3.5_2.12:1.6.1
- org.apache.spark:spark-sql-kafka-0-10_2.12:3.5.2
s3connection:
reference: minio
sparkConf:
Expand Down
2 changes: 1 addition & 1 deletion demos/end-to-end-security/create-spark-report.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -55,7 +55,7 @@ data:
name: spark-report
spec:
sparkImage:
productVersion: 3.5.1
productVersion: 3.5.2
mode: cluster
mainApplicationFile: local:///stackable/spark/jobs/spark-report.py
deps:
Expand Down
3 changes: 2 additions & 1 deletion demos/hbase-hdfs-load-cycling-data/distcp-cycling-data.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -11,14 +11,15 @@ spec:
# We use 24.3.0 here which contains the distcp MapReduce components
# This is not included in the 24.7 images and will fail.
# See: https://github.com/stackabletech/docker-images/issues/793
image: docker.stackable.tech/stackable/hadoop:3.3.4-stackable24.3.0
image: docker.stackable.tech/stackable/hadoop:3.4.0-stackable24.11.0
NickLarsenNZ marked this conversation as resolved.
Show resolved Hide resolved
env:
- name: HADOOP_USER_NAME
value: stackable
- name: HADOOP_CONF_DIR
value: "/stackable/conf/hdfs"
- name: HADOOP_CLASSPATH
value: "/stackable/hadoop/share/hadoop/tools/lib/*.jar"
# yamllint disable-line rule:line-length
command: ["bash", "-c", "bin/hdfs dfs -mkdir -p /data/raw && bin/hadoop distcp -D fs.s3a.aws.credentials.provider=org.apache.hadoop.fs.s3a.AnonymousAWSCredentialsProvider s3a://public-backup-nyc-tlc/cycling-tripdata/demo-cycling-tripdata.csv.gz hdfs://hdfs/data/raw"]
volumeMounts:
- name: config-volume-hdfs
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -8,13 +8,15 @@ spec:
spec:
containers:
- name: load-ny-taxi-data
image: docker.stackable.tech/stackable/hadoop:3.3.4-stackable24.7.0
image: docker.stackable.tech/stackable/hadoop:3.4.0-stackable24.11.0
NickLarsenNZ marked this conversation as resolved.
Show resolved Hide resolved
# yamllint disable rule:line-length
command: ["bash", "-c", "/stackable/hadoop/bin/hdfs dfs -mkdir -p /ny-taxi-data/raw \
&& cd /tmp \
&& for month in 2020-09; do \
curl -O https://repo.stackable.tech/repository/misc/ny-taxi-data/fhvhv_tripdata_$month.parquet \
&& /stackable/hadoop/bin/hdfs dfs -put fhvhv_tripdata_$month.parquet /ny-taxi-data/raw/; \
done"]
# yamllint enable rule:line-length
volumeMounts:
- name: hdfs-discovery-configmap
mountPath: /hdfs
Expand Down
4 changes: 3 additions & 1 deletion demos/signal-processing/Dockerfile-nifi
Original file line number Diff line number Diff line change
@@ -1,3 +1,5 @@
FROM docker.stackable.tech/stackable/nifi:1.27.0-stackable24.7.0
# docker build -f ./Dockerfile-nifi -t docker.stackable.tech/demos/nifi:1.27.0-postgresql .

FROM docker.stackable.tech/stackable/nifi:1.27.0-stackable24.11.0
NickLarsenNZ marked this conversation as resolved.
Show resolved Hide resolved

NickLarsenNZ marked this conversation as resolved.
Show resolved Hide resolved
RUN curl --fail -o /stackable/nifi/postgresql-42.6.0.jar "https://repo.stackable.tech/repository/misc/postgresql-timescaledb/postgresql-42.6.0.jar"
12 changes: 6 additions & 6 deletions demos/signal-processing/DownloadAndWriteToDB.xml
Original file line number Diff line number Diff line change
Expand Up @@ -61,7 +61,7 @@
<bundle>
<artifact>nifi-dbcp-service-nar</artifact>
<group>org.apache.nifi</group>
<version>1.21.0</version>
<version>1.27.0</version>
</bundle>
<comments></comments>
<descriptors>
Expand Down Expand Up @@ -258,7 +258,7 @@
<bundle>
<artifact>nifi-record-serialization-services-nar</artifact>
<group>org.apache.nifi</group>
<version>1.21.0</version>
<version>1.27.0</version>
</bundle>
<comments></comments>
<descriptors>
Expand Down Expand Up @@ -561,7 +561,7 @@
</position>
<height>88.0</height>
<label>This flow downloads a dataset, writing it to a temporary table in TimescaleDB.
This data is then written to the target table with the time offsets preserved,
This data is then written to the target table with the time offsets preserved,
but re-based to the current time. This means that the data can be displayed
in Grafana as if it were being streamed, whereas in fact the dashboard moves
through "future" data that has already been persisted.</label>
Expand All @@ -584,7 +584,7 @@ through "future" data that has already been persisted.</label>
<bundle>
<artifact>nifi-standard-nar</artifact>
<group>org.apache.nifi</group>
<version>1.21.0</version>
<version>1.27.0</version>
</bundle>
<config>
<backoffMechanism>PENALIZE_FLOWFILE</backoffMechanism>
Expand Down Expand Up @@ -1069,7 +1069,7 @@ through "future" data that has already been persisted.</label>
<bundle>
<artifact>nifi-standard-nar</artifact>
<group>org.apache.nifi</group>
<version>1.21.0</version>
<version>1.27.0</version>
</bundle>
<config>
<backoffMechanism>PENALIZE_FLOWFILE</backoffMechanism>
Expand Down Expand Up @@ -1223,7 +1223,7 @@ from conditions_temp;</value>
<bundle>
<artifact>nifi-standard-nar</artifact>
<group>org.apache.nifi</group>
<version>1.21.0</version>
<version>1.27.0</version>
</bundle>
<config>
<backoffMechanism>PENALIZE_FLOWFILE</backoffMechanism>
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -37,7 +37,7 @@ data:
name: spark-ad
spec:
sparkImage:
productVersion: 3.5.1
productVersion: 3.5.2
mode: cluster
mainApplicationFile: local:///spark-scripts/spark-ad.py
deps:
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -9,6 +9,7 @@ spec:
containers:
- name: load-ny-taxi-data
image: "bitnami/minio:2022-debian-10"
# yamllint disable-line rule:line-length
command: ["bash", "-c", "cd /tmp && for month in 2020-09 2020-10 2020-11 2020-12; do curl -O https://repo.stackable.tech/repository/misc/ny-taxi-data/fhvhv_tripdata_$month.parquet && mc --insecure alias set minio http://minio:9000/ $(cat /minio-s3-credentials/accessKey) $(cat /minio-s3-credentials/secretKey) && mc cp fhvhv_tripdata_$month.parquet minio/demo/ny-taxi-data/raw/ && mc mb --ignore-existing minio/prediction; done"]
volumeMounts:
- name: minio-s3-credentials
Expand Down
1 change: 1 addition & 0 deletions demos/trino-taxi-data/load-test-data.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -9,6 +9,7 @@ spec:
containers:
- name: load-ny-taxi-data
image: "bitnami/minio:2024-debian-12"
# yamllint disable-line rule:line-length
command: ["bash", "-c", "cd /tmp && for month in 2020-01 2020-02 2020-03 2020-04 2020-05 2020-06 2020-07 2020-08 2020-09 2020-10 2020-11 2020-12 2021-01 2021-02 2021-03 2021-04 2021-05 2021-06 2021-07 2021-08 2021-09 2021-10 2021-11 2021-12 2022-01 2022-02 2022-03 2022-04; do curl -O https://repo.stackable.tech/repository/misc/ny-taxi-data/yellow_tripdata_$month.parquet && mc --insecure alias set minio http://minio:9000/ $(cat /minio-s3-credentials/accessKey) $(cat /minio-s3-credentials/secretKey) && mc cp yellow_tripdata_$month.parquet minio/demo/ny-taxi-data/raw/; done"]
volumeMounts:
- name: minio-s3-credentials
Expand Down
2 changes: 1 addition & 1 deletion docs/modules/demos/pages/trino-iceberg.adoc
Original file line number Diff line number Diff line change
Expand Up @@ -78,7 +78,7 @@ As an alternative, you can use https://trino.io/download.html[trino-cli] by runn

[source,console]
----
$ java -jar ~/Downloads/trino-cli-451-executable.jar --user admin --insecure --password --server https://172.18.0.2:30856
$ java -jar ~/Downloads/trino-cli-455-executable.jar --user admin --insecure --password --server https://172.18.0.2:30856
----

Make sure to replace the server endpoint with the endpoint listed in the `stackablectl stacklet list` output.
Expand Down
6 changes: 3 additions & 3 deletions stacks/_templates/vector-aggregator.yaml
Original file line number Diff line number Diff line change
@@ -1,13 +1,13 @@
releaseName: vector-aggregator

Check warning on line 1 in stacks/_templates/vector-aggregator.yaml

View workflow job for this annotation

GitHub Actions / pre-commit

1:1 [document-start] missing document start "---"
name: vector
repo:
name: vector
url: https://helm.vector.dev
version: 0.34.0
version: 0.36.1 # app version 0.41.1
options:
commonLabels:
stackable.tech/vendor: Stackable
podLabels: # Doesn't seem to work?
podLabels:
stackable.tech/vendor: Stackable
role: Aggregator
customConfig:
Expand All @@ -25,7 +25,7 @@
- https://opensearch-cluster-master.default.svc.cluster.local:9200
mode: bulk
# The auto-detection of the API version does not work in Vector
# 0.39.0 for OpenSearch, so the version must be set explicitly
# 0.41.1 for OpenSearch, so the version must be set explicitly
# (see https://github.com/vectordotdev/vector/issues/17690).
api_version: v8
tls:
Expand Down
12 changes: 6 additions & 6 deletions stacks/airflow/airflow.yaml
Original file line number Diff line number Diff line change
@@ -1,12 +1,12 @@
{% raw %}
# {% raw %}
---
apiVersion: airflow.stackable.tech/v1alpha1
kind: AirflowCluster
metadata:
name: airflow
spec:
image:
productVersion: 2.9.2
productVersion: 2.9.3
clusterConfig:
listenerClass: external-unstable
loadExamples: false
Expand Down Expand Up @@ -272,7 +272,7 @@ data:
spec:
version: "1.0"
sparkImage:
productVersion: 3.5.1
productVersion: 3.5.2
mode: cluster
mainApplicationFile: local:///stackable/spark/examples/src/main/python/pi.py
job:
Expand All @@ -285,22 +285,22 @@ data:
limit: 512Mi
driver:
config:
resources:
resources:
cpu:
min: 1000m
max: 1200m
memory:
limit: 1024Mi
executor:
config:
resources:
resources:
cpu:
min: 500m
max: 1000m
memory:
limit: 1024Mi
replicas: 3
{% endraw %}
# {% endraw %}
---
apiVersion: v1
kind: Secret
Expand Down
4 changes: 2 additions & 2 deletions stacks/data-lakehouse-iceberg-trino-spark/nifi.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@ metadata:
name: nifi
spec:
image:
productVersion: 1.25.0
productVersion: 1.27.0
clusterConfig:
authentication:
- authenticationClass: nifi-admin-credentials
Expand Down Expand Up @@ -52,7 +52,7 @@ kind: Secret
metadata:
name: nifi-admin-credentials-secret
stringData:
admin: {{ nifiAdminPassword }}
admin: {{nifiAdminPassword}}
NickLarsenNZ marked this conversation as resolved.
Show resolved Hide resolved
---
apiVersion: zookeeper.stackable.tech/v1alpha1
kind: ZookeeperZnode
Expand Down
4 changes: 2 additions & 2 deletions stacks/data-lakehouse-iceberg-trino-spark/trino.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@ metadata:
name: trino
spec:
image:
productVersion: "451"
productVersion: "455"
clusterConfig:
listenerClass: external-unstable
catalogLabelSelector:
Expand Down Expand Up @@ -114,7 +114,7 @@ metadata:
name: opa
spec:
image:
productVersion: 0.66.0
productVersion: 0.67.1
servers:
roleGroups:
default: {}
Expand Down
2 changes: 1 addition & 1 deletion stacks/dual-hive-hdfs-s3/hdfs.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -25,7 +25,7 @@ metadata:
name: hdfs
spec:
image:
productVersion: 3.3.4
productVersion: 3.4.0
clusterConfig:
listenerClass: external-unstable
dfsReplication: 1
Expand Down
2 changes: 1 addition & 1 deletion stacks/dual-hive-hdfs-s3/trino.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -61,7 +61,7 @@ metadata:
name: trino
spec:
image:
productVersion: "451"
productVersion: "455"
clusterConfig:
authorization:
opa:
Expand Down
8 changes: 5 additions & 3 deletions stacks/end-to-end-security/hdfs.yaml
Original file line number Diff line number Diff line change
@@ -1,10 +1,11 @@
---
apiVersion: hdfs.stackable.tech/v1alpha1
kind: HdfsCluster
metadata:
name: hdfs
spec:
image:
productVersion: 3.3.4
productVersion: 3.4.0
clusterConfig:
zookeeperConfigMapName: hdfs-znode
authentication:
Expand All @@ -29,8 +30,9 @@ spec:
level: DEBUG
configOverrides: &configOverrides
core-site.xml:
# The idea is that the user "hive" can't do anything in hdfs, *but* it can impersonate other users
# (such as trino), that have the needed permissions
# The idea is that the user "hive" can't do anything in hdfs,
# *but* it can impersonate other users (such as trino),
# that have the needed permissions
hadoop.proxyuser.hive.users: "*"
hadoop.proxyuser.hive.hosts: "*"
roleGroups:
Expand Down
2 changes: 1 addition & 1 deletion stacks/end-to-end-security/opa.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@ metadata:
name: opa
spec:
image:
productVersion: 0.66.0
productVersion: 0.67.1
clusterConfig:
userInfo:
backend:
Expand Down
8 changes: 4 additions & 4 deletions stacks/end-to-end-security/trino-regorules.yaml
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
{% raw %}

Check failure on line 1 in stacks/end-to-end-security/trino-regorules.yaml

View workflow job for this annotation

GitHub Actions / pre-commit

1:2 syntax error: found character '%' that cannot start any token (syntax)
---
apiVersion: v1
kind: ConfigMap
Expand Down Expand Up @@ -1460,7 +1460,7 @@
# "groups": ["group1", ...],
# "user": "username",
# },
# "softwareStack": {"trinoVersion": "442"},
# "softwareStack": {"trinoVersion": "455"},
# }
# }
#
Expand Down Expand Up @@ -1610,7 +1610,7 @@
# "groups": ["group1", ...],
# "user": "username",
# },
# "softwareStack": {"trinoVersion": "442"},
# "softwareStack": {"trinoVersion": "455"},
# }
# }
#
Expand Down Expand Up @@ -1671,7 +1671,7 @@
# "groups": ["group1", ...],
# "user": "username",
# },
# "softwareStack": {"trinoVersion": "442"},
# "softwareStack": {"trinoVersion": "455"},
# }
# }
#
Expand Down Expand Up @@ -1736,7 +1736,7 @@
# "groups": ["group1", ...],
# "user": "username",
# },
# "softwareStack": {"trinoVersion": "442"},
# "softwareStack": {"trinoVersion": "455"},
# }
# }
#
Expand Down
2 changes: 1 addition & 1 deletion stacks/end-to-end-security/trino.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@ metadata:
name: trino
spec:
image:
productVersion: "451"
productVersion: "455"
clusterConfig:
listenerClass: external-unstable
tls:
Expand Down
3 changes: 2 additions & 1 deletion stacks/hdfs-hbase/hdfs.yaml
Original file line number Diff line number Diff line change
@@ -1,10 +1,11 @@
---
apiVersion: hdfs.stackable.tech/v1alpha1
kind: HdfsCluster
metadata:
name: hdfs
spec:
image:
productVersion: 3.3.4
productVersion: 3.4.0
clusterConfig:
dfsReplication: 1
zookeeperConfigMapName: hdfs-znode
Expand Down
2 changes: 1 addition & 1 deletion stacks/jupyterhub-pyspark-hdfs/hdfs.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -13,7 +13,7 @@ metadata:
name: hdfs
spec:
image:
productVersion: 3.3.4
productVersion: 3.4.0
clusterConfig:
dfsReplication: 1
zookeeperConfigMapName: hdfs-znode
Expand Down
Loading
Loading