Skip to content

Commit

Permalink
Merge remote-tracking branch 'origin/next' into chore/nifi-24.11
Browse files Browse the repository at this point in the history
  • Loading branch information
NickLarsenNZ committed Sep 19, 2024
2 parents d5b4b25 + 0c7d349 commit 8fbf391
Show file tree
Hide file tree
Showing 26 changed files with 454 additions and 619 deletions.
19 changes: 19 additions & 0 deletions .github/workflows/pr_pre-commit.yml
Original file line number Diff line number Diff line change
@@ -0,0 +1,19 @@
---
name: pre-commit

on:
pull_request:

jobs:
pre-commit:
runs-on: ubuntu-latest
steps:
- uses: actions/checkout@692973e3d937129bcbf40652eb9f2f61becf3332 # v4.1.7
with:
fetch-depth: 0
- uses: actions/setup-python@82c7e631bb3cdc910f68e0081d67478d79c6982d # v5.1.0
with:
python-version: '3.12'
- uses: pre-commit/action@2c7b3805fd2a0fd8c1884dcaebf91fc102a13ecd # v3.0.1
with:
extra_args: "--from-ref ${{ github.event.pull_request.base.sha }} --to-ref ${{ github.event.pull_request.head.sha }}"
20 changes: 20 additions & 0 deletions .markdownlint.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,20 @@
---
# All defaults or options can be checked here:
# https://github.com/DavidAnson/markdownlint/blob/main/schema/.markdownlint.yaml

# Default state for all rules
default: true

# MD013/line-length - Line length
MD013:
# Number of characters
line_length: 9999
# Number of characters for headings
heading_line_length: 9999
# Number of characters for code blocks
code_block_line_length: 9999

# MD024/no-duplicate-heading/no-duplicate-header - Multiple headings with the same content
MD024:
# Only check sibling headings
siblings_only: true
34 changes: 34 additions & 0 deletions .pre-commit-config.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,34 @@
---
default_language_version:
node: system

repos:
- repo: https://github.com/pre-commit/pre-commit-hooks
rev: 2c9f875913ee60ca25ce70243dc24d5b6415598c # 4.6.0
hooks:
- id: trailing-whitespace
- id: end-of-file-fixer
- id: detect-aws-credentials
args: ["--allow-missing-credentials"]
- id: detect-private-key

- repo: https://github.com/adrienverge/yamllint
rev: 81e9f98ffd059efe8aa9c1b1a42e5cce61b640c6 # 1.35.1
hooks:
- id: yamllint

- repo: https://github.com/igorshubovych/markdownlint-cli
rev: f295829140d25717bc79368d3f966fc1f67a824f # 0.41.0
hooks:
- id: markdownlint

- repo: https://github.com/koalaman/shellcheck-precommit
rev: 2491238703a5d3415bb2b7ff11388bf775372f29 # 0.10.0
hooks:
- id: shellcheck
args: ["--severity=info"]

- repo: https://github.com/rhysd/actionlint
rev: 62dc61a45fc95efe8c800af7a557ab0b9165d63b # 1.7.1
hooks:
- id: actionlint
9 changes: 9 additions & 0 deletions .yamllint.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,9 @@
---
extends: default

rules:
line-length: disable
truthy:
check-keys: false
comments:
min-spaces-from-content: 1 # Needed due to https://github.com/adrienverge/yamllint/issues/443
19 changes: 0 additions & 19 deletions demos/demos-v2.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -147,25 +147,6 @@ demos:
cpu: 6800m
memory: 15822Mi
pvc: 28Gi
trino-subsea-data:
description: Demo loading ca. 600m^2 of ocean floor in a surface plot to visualize the irregularities of the ocean floor.
# documentation: -- Currently not documented
stackableStack: trino-superset-s3
labels:
- trino
- superset
- minio
- s3
- parquet
manifests:
- plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/trino-subsea-data/load-test-data.yaml
- plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/trino-subsea-data/create-table-in-trino.yaml
- plainYaml: https://raw.githubusercontent.com/stackabletech/demos/main/demos/trino-subsea-data/setup-superset.yaml
supportedNamespaces: []
resourceRequests:
cpu: 6800m
memory: 15822Mi
pvc: 28Gi
data-lakehouse-iceberg-trino-spark:
description: Data lakehouse using Iceberg lakehouse on S3, Trino as query engine, Spark for streaming ingest and Superset for data visualization. Multiple datasources like taxi data, water levels in Germany, earthquakes, e-charging stations and more are loaded.
documentation: https://docs.stackable.tech/stackablectl/stable/demos/data-lakehouse-iceberg-trino-spark.html
Expand Down
83 changes: 0 additions & 83 deletions demos/trino-subsea-data/create-table-in-trino.yaml

This file was deleted.

21 changes: 0 additions & 21 deletions demos/trino-subsea-data/load-test-data.yaml

This file was deleted.

95 changes: 0 additions & 95 deletions demos/trino-subsea-data/setup-superset.yaml

This file was deleted.

Binary file removed demos/trino-subsea-data/superset-assets.zip
Binary file not shown.
6 changes: 4 additions & 2 deletions docs/modules/demos/pages/airflow-scheduled-job.adoc
Original file line number Diff line number Diff line change
@@ -1,5 +1,6 @@
= airflow-scheduled-job
:page-aliases: stable@stackablectl::demos/airflow-scheduled-job.adoc
:description: This demo installs Airflow with Postgres and Redis on Kubernetes, showcasing DAG scheduling, job runs, and status verification via the Airflow UI.

Install this demo on an existing Kubernetes cluster:

Expand Down Expand Up @@ -102,9 +103,10 @@ Click on the `run_every_minute` box in the centre of the page and then select `L

[WARNING]
====
In this demo, the logs are not available when the KubernetesExecutor is deployed. See the https://airflow.apache.org/docs/apache-airflow/stable/core-concepts/executor/kubernetes.html#managing-dags-and-logs[Airflow Documentation] for more details.
In this demo, the logs are not available when the KubernetesExecutor is deployed.
See the https://airflow.apache.org/docs/apache-airflow/stable/core-concepts/executor/kubernetes.html#managing-dags-and-logs[Airflow Documentation] for more details.
If you are interested in persisting the logs, please take a look at the xref:logging.adoc[] demo.
If you are interested in persisting the logs, take a look at the xref:logging.adoc[] demo.
====

image::airflow-scheduled-job/airflow_9.png[]
Expand Down
Loading

0 comments on commit 8fbf391

Please sign in to comment.