From ee0d49365bc363a6c87566d425b4cc8ca8c64b64 Mon Sep 17 00:00:00 2001 From: Oliver Howell Date: Thu, 12 Sep 2024 12:17:30 +0100 Subject: [PATCH 01/11] HZX-232: add build variables to playbook --- antora-playbook.yml | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/antora-playbook.yml b/antora-playbook.yml index 6a48fd1..d02bc7e 100644 --- a/antora-playbook.yml +++ b/antora-playbook.yml @@ -1,8 +1,8 @@ site: - title: Flow Docs [title] - url: + title: Hazelcast Flow + url: https://silly-otter-82acc9.netlify.app/ robots: disallow - start_page: + start_page: hz-flow:index.adoc keys: google_analytics: 'GTM-MQ59XD6' docsearch_id: 'QK2EAH8GB0' From 6ab4d1f8ef74e3dc23940417699a137ace7fe3a0 Mon Sep 17 00:00:00 2001 From: Patrick McGleenon Date: Mon, 30 Sep 2024 14:52:46 +0100 Subject: [PATCH 02/11] HZX-241: renamed to hazelflix (#39) # Description of change https://hazelcast.atlassian.net/browse/HZX-241 Renamed project to `hazelflix` --- docs/modules/guides/pages/apis-db-kafka.adoc | 10 +++++----- 1 file changed, 5 insertions(+), 5 deletions(-) diff --git a/docs/modules/guides/pages/apis-db-kafka.adoc b/docs/modules/guides/pages/apis-db-kafka.adoc index 76562b9..4c8cb1a 100644 --- a/docs/modules/guides/pages/apis-db-kafka.adoc +++ b/docs/modules/guides/pages/apis-db-kafka.adoc @@ -85,7 +85,7 @@ The project is a https://taxilang.org[Taxi] project which gets edited locally, t - From the sidebar, click http://localhost:9021/projects[Projects] - Click *Add Project* and add a Local Disk project - - For the **Project Path**, enter `petflix` + - For the **Project Path**, enter `demo` - The server will tell you it can't find an existing project, and let you create one - click **Create new project** - Enter the package co-ordinates as follows (this is similar to npm, maven, etc.) @@ -94,7 +94,7 @@ The project is a https://taxilang.org[Taxi] project which gets edited locally, t | Field | Value | *Organisation* -| `com.petflix` +| `com.hazelflix` | *Name* | `demo` @@ -189,7 +189,7 @@ To import the schema: | `film` | Default namespace -| `com.petflix.filmsdatabase` +| `com.hazelflix.demo.filmsdatabase` |=== Namespaces are used to help us group related content together, like packages in Java or namespaces in C# and Typescript. @@ -262,7 +262,7 @@ Fill in the form with the following values: | `+http://films-api/v3/api-docs+` | Default namespace -| `com.petflix.listings` +| `com.hazelflix.listings` | Base url | Leave this blank @@ -574,7 +574,7 @@ Fill out the rest of the form with the following details: | `LATEST` | Namespace -| `com.petflix.announcements` +| `com.hazelflix.announcements` | Message Type | `NewFilmReleaseAnnouncement` From b20ff0186d81215bcd74ae52e94eb996d7ad8994 Mon Sep 17 00:00:00 2001 From: Michael Pollett Date: Tue, 1 Oct 2024 15:31:36 +0100 Subject: [PATCH 03/11] replace some of the docker references (#38) * Where we refer to Docker Compose, keep the reference * Where we're just referring to a container, call it a container, not Docker --- docs/modules/data-formats/pages/avro.adoc | 2 +- docs/modules/deploying/pages/configuring.adoc | 5 +- .../pages/distributing-work-on-a-cluster.adoc | 4 +- .../pages/production-deployments.adoc | 4 +- .../pages/configuring-connections.adoc | 4 +- .../pages/enable-ui-schema-editing.adoc | 3 - .../pages/hazelcast.adoc | 2 +- docs/modules/guides/pages/apis-db-kafka.adoc | 4 +- .../introduction/pages/get-support.adoc | 2 +- .../modules/querying/pages/observability.adoc | 4 +- docs/modules/querying/pages/streams.adoc | 90 ------------------- 11 files changed, 14 insertions(+), 110 deletions(-) delete mode 100644 docs/modules/querying/pages/streams.adoc diff --git a/docs/modules/data-formats/pages/avro.adoc b/docs/modules/data-formats/pages/avro.adoc index ff73cf0..e93d626 100644 --- a/docs/modules/data-formats/pages/avro.adoc +++ b/docs/modules/data-formats/pages/avro.adoc @@ -203,7 +203,7 @@ NOTE: 'Local' means local to the server. This workflow adds a reference to an Avro file that's *on the disk of the server*. + + -It's intended for developers who are running {short-product-name} in a Docker image on their local machine. + +It's intended for developers who are running {short-product-name} in a container on their local machine. + + This workflow isn't intended for uploading an Avro schema to a remote server. Instead, use a Git repository diff --git a/docs/modules/deploying/pages/configuring.adoc b/docs/modules/deploying/pages/configuring.adoc index bea4733..fe056d2 100644 --- a/docs/modules/deploying/pages/configuring.adoc +++ b/docs/modules/deploying/pages/configuring.adoc @@ -65,11 +65,10 @@ There are several configuration settings referenced throughout these docs, which All config settings can be passed in a variety of ways: -==== Docker +==== Container -In a Docker / Docker Compose file, pass variables using the `OPTIONS` environment variable: +In a container or Docker Compose file, pass variables using the `OPTIONS` environment variable: -//rebranded vyne to flow below and substituted the short name variable for 'flow' as this doesn't render in code snippets - check if correct ---- services: flow: diff --git a/docs/modules/deploying/pages/distributing-work-on-a-cluster.adoc b/docs/modules/deploying/pages/distributing-work-on-a-cluster.adoc index 74155c0..555fda6 100644 --- a/docs/modules/deploying/pages/distributing-work-on-a-cluster.adoc +++ b/docs/modules/deploying/pages/distributing-work-on-a-cluster.adoc @@ -40,8 +40,6 @@ Caches are not shared between worker nodes, so remote services may receive a hig As the cluster size grows, the work can be parallelized across a greater number of nodes. While this provides improved throughput, the work coordinator incurs a heavier workload in serialization and deserialization of work tasks and responses. -//rebranded vyne to flow below - check! - To account for this the `flow.projection.distributionRemoteBias` allows tuning at which point the work is preferentially distributed to remote nodes, versus the query coordinator. Once this value is exceeded, the query coordinator node will perform a lower proportion of projection work in a query. @@ -67,7 +65,7 @@ For projection work to be distributed across all {short-product-name} nodes in a == Advanced cluster configuration -This section may only be required if you're configuring the Docker orchestration yourself and not using one of our preconfigured templates. +This section may only be required if you're configuring the container orchestration yourself and not using one of our preconfigured templates. === Cluster discovery types diff --git a/docs/modules/deploying/pages/production-deployments.adoc b/docs/modules/deploying/pages/production-deployments.adoc index c45f901..c37b427 100644 --- a/docs/modules/deploying/pages/production-deployments.adoc +++ b/docs/modules/deploying/pages/production-deployments.adoc @@ -1,7 +1,7 @@ = Deploying {short-product-name} :description: How to deploy {short-product-name} to production -Our recommended deployment method is with `docker compose` or deploying the Docker container directly using the orchestrator of your choice. +Our recommended deployment method is with `docker compose` or deploying the container directly using the orchestrator of your choice. == Deploying using Docker Compose @@ -23,7 +23,7 @@ When deploying to production, there are common tasks you may wish to perform: * Enable xref:deploying:authentication.adoc[authentication] and xref:deploying:authorization.adoc[authorization] in {short-product-name} * Use a xref:workspace:overview.adoc#reading-workspace-conf-from-git[Git-backed workspace.conf file], to align with IAC * Understand how xref:deploying:managing-secrets.adoc[secrets are handled] in {short-product-name} -* Understand how to configure {short-product-name} via xref:deploying:configuring.adoc#docker[application properties] or xref:deploying:configuring.adoc#passing-{short-product-name}-application-configuration[environment variables] +* Understand how to configure {short-product-name} via xref:deploying:configuring.adoc#container[application properties] or xref:deploying:configuring.adoc#passing-{short-product-name}-application-configuration[environment variables] == Further deployment templates diff --git a/docs/modules/describing-data-sources/pages/configuring-connections.adoc b/docs/modules/describing-data-sources/pages/configuring-connections.adoc index 5d27f18..ace7b9e 100644 --- a/docs/modules/describing-data-sources/pages/configuring-connections.adoc +++ b/docs/modules/describing-data-sources/pages/configuring-connections.adoc @@ -194,8 +194,8 @@ Connection parameters are as follows: [,HOCON] ---- jdbc { - mysql-docker { - connectionName=mysql-docker + mysql-db { + connectionName=mysql-db connectionParameters { database=test host=localhost diff --git a/docs/modules/describing-data-sources/pages/enable-ui-schema-editing.adoc b/docs/modules/describing-data-sources/pages/enable-ui-schema-editing.adoc index a135dc0..2d98d29 100644 --- a/docs/modules/describing-data-sources/pages/enable-ui-schema-editing.adoc +++ b/docs/modules/describing-data-sources/pages/enable-ui-schema-editing.adoc @@ -25,9 +25,6 @@ launched the Schema Server from. You may have overridden this when you launched the Schema Server, by specifying `+--vyne.repositories.config-file=...+`. -If you're running one of our demo tutorials, the config file is at `vyne/schema-server/schema-server.conf`, relative -to the docker-compose file you used. - Find and open the schema-server config file in your editor of choice. === Specify a new file-based repository diff --git a/docs/modules/describing-data-sources/pages/hazelcast.adoc b/docs/modules/describing-data-sources/pages/hazelcast.adoc index f571a95..d9f14a6 100644 --- a/docs/modules/describing-data-sources/pages/hazelcast.adoc +++ b/docs/modules/describing-data-sources/pages/hazelcast.adoc @@ -17,7 +17,7 @@ hazelcast { } ``` -NOTE: Assuming your docker-compose.yml contains a Hazelcast container and the container is up and running, here is an example of how the Hazelcast container configuration might look: +NOTE: Assuming you are using Docker Compose to run {short-product-name}, here is an example of how to configure an external Hazelcast container: ```yaml hazelcast: image: "docker.io/hazelcast/hazelcast:latest" diff --git a/docs/modules/guides/pages/apis-db-kafka.adoc b/docs/modules/guides/pages/apis-db-kafka.adoc index 4c8cb1a..c41eb0e 100644 --- a/docs/modules/guides/pages/apis-db-kafka.adoc +++ b/docs/modules/guides/pages/apis-db-kafka.adoc @@ -55,7 +55,7 @@ which we'll use in our next steps. If you run `docker ps`, you should see a collection of Docker containers now running. |=== -| Docker Image | Part of {short-product-name} stack or Demo? | Description +| Container Name | Part of {short-product-name} stack or Demo? | Description | {code-product-name} | {short-product-name} @@ -69,7 +69,7 @@ If you run `docker ps`, you should see a collection of Docker containers now run | Demo | A Postgres DB, which contains the Postgres https://github.com/devrimgunduz/pagila[Pagila] demo database for a fake DVD rental store -| apache/kafka +| kafka | Demo (Kafka) | The Apache Kafka image |=== diff --git a/docs/modules/introduction/pages/get-support.adoc b/docs/modules/introduction/pages/get-support.adoc index ccb5175..0cb9b30 100644 --- a/docs/modules/introduction/pages/get-support.adoc +++ b/docs/modules/introduction/pages/get-support.adoc @@ -25,7 +25,7 @@ When submitting a ticket to the team, please provide as much information and dat * Add details of use case. This is crucial as it helps Support narrow down the features and functionality in play when the problem occurred * Attach any specific errors found * Attach screenshots of the errors your encountered -* Attach the complete logs files, i.e., docker container logs +* Attach the complete logs files, i.e., container logs * Attach details of your 3rd party connections * Attach details of the queries you are making * Specify the time of incident diff --git a/docs/modules/querying/pages/observability.adoc b/docs/modules/querying/pages/observability.adoc index 9b01801..6630c00 100644 --- a/docs/modules/querying/pages/observability.adoc +++ b/docs/modules/querying/pages/observability.adoc @@ -75,8 +75,8 @@ scrape_configs: metrics_path: '/api/actuator/prometheus' scrape_interval: 3s static_configs: - # 172.17.0.1 is the ip address of localhost - # from within docker + # 172.17.0.1 should be the ip address of localhost + # from within the container network - targets: ['172.17.0.1:9021'] labels: application: 'Flow' diff --git a/docs/modules/querying/pages/streams.adoc b/docs/modules/querying/pages/streams.adoc deleted file mode 100644 index c64c43a..0000000 --- a/docs/modules/querying/pages/streams.adoc +++ /dev/null @@ -1,90 +0,0 @@ -= Build ETL pipelines - -// Author's Note - this topic has been removed from the Flow docs and does not appear in the TOC/navdoc. The content is retained for future use. - -NOTE: Pipelines are gradually being replaced by xref:streams:streaming-data.adoc[streaming queries]. The pipeline syntax is still supported, but we aim to migrate functionality to streaming queries. For workloads that are oriented around streaming data (such as reading from Kafka), use a streaming query instead. - -## Overview - -Where queries in {short-product-name} are great for on-demand data transformation, pipelines provide continuously streaming data between source and destination, using {short-product-name} to transform data along the way. - -Pipelines work by reading content from a source, transforming it using {short-product-name}, then writing to a sink. - -Pipelines can be defined either in a config file, or using the UI. - -// Above, link to config file topic if/when available - -## Starting pipeline engine -The pipeline engine ships as a separate Docker image. - -A typical Docker Compose config looks as follows: - -```yaml -version: "3.3" -services: - ## Other services omitted - pipeline-engine: - image: flow/pipeline:latest -``` - -## Creating a pipeline -Pipelines are created and stored as a series of HOCON files, within a Taxi project. - - -```hocon taxi.conf - name: com.demo/pipelines - version: 0.1.0 - sourceRoot: src/ - additionalSources: { - "@flow/config" : "flow/config/*.conf", - "@flow/pipelines" : "pipelines/*.conf" - } -``` - -Pipelines are defined as an Input (transport + spec) and one or more Outputs (transport + spec), with {short-product-name} handling the transformation and enrichment. - -```hocon pipeline.conf -// An example pipeline, which watches a local directory for new files. -// For each file, an integration is performed using a transformation, -// and then written to a db -pipelines: [{ - "name": "Filewatcher to db", - "id" : "file-watch", - "input": { - "type": "fileWatcher", - "direction": "INPUT", - "typeName" : "Film", - "path" : ${config.pipelines.watchedPath} - }, - "transformation" : """ - find { Films } as { - title: FilmTitle - director: DirectorName - imdbId: ImdbId - duration: DurationMillis? = first(TrackObject[]) as DurationMillis - } - """, - "outputs": [{ - "type": "jdbc", - "direction": "OUTPUT", - "connection": "films", - "tableName": "Films" - }] -}] - -``` - -Pipelines listen on the input, and as messages arrive they are passed to {short-product-name} to transform to -the output models. - -The data is transformed, and {short-product-name}'s query engine is used to enrich and populate any missing data. - -Although all pipelines are persisted as JSON files, they can be created either through the UI, -or authored as JSON directly. - -### Creating a pipeline through configuration -Pipelines are authored and stored in JSON files, with a file per pipeline. - -Files should adhere to the pipeline spec, which has its own dedicated reference documentation. - -// Above, link to reference doc if/when available: /docs/pipelines/reference[reference documentation]. From 932ad5d178ca470ade580e44042a2fdfa20b8950 Mon Sep 17 00:00:00 2001 From: Amanda Lindsay Date: Tue, 1 Oct 2024 15:33:09 +0100 Subject: [PATCH 04/11] Flow content v5 (#37) Minor docs changes to address issues with links and link text --- .../deploying/images/flow-components2.svg | 320 ++++++++++++++++++ .../deploying/pages/authentication.adoc | 2 +- docs/modules/deploying/pages/components.adoc | 4 +- .../pages/development-deployments.adoc | 3 +- .../pages/aws-services.adoc | 2 +- .../describing-data-sources/pages/kafka.adoc | 2 +- 6 files changed, 328 insertions(+), 5 deletions(-) create mode 100644 docs/modules/deploying/images/flow-components2.svg diff --git a/docs/modules/deploying/images/flow-components2.svg b/docs/modules/deploying/images/flow-components2.svg new file mode 100644 index 0000000..995320a --- /dev/null +++ b/docs/modules/deploying/images/flow-components2.svg @@ -0,0 +1,320 @@ + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + diff --git a/docs/modules/deploying/pages/authentication.adoc b/docs/modules/deploying/pages/authentication.adoc index 5e7aca2..1f75709 100644 --- a/docs/modules/deploying/pages/authentication.adoc +++ b/docs/modules/deploying/pages/authentication.adoc @@ -84,7 +84,7 @@ The presented JWT is expected to have the following attributes: Below you can find example Docker Compose files where MC is the identity provider and a more complex scenario where MC is configured to use an external identity provider. -NOTE: MC is preconfigured as the identity provider for {short-product-name} with the configuration prefixed `flow.security` and `FLOW_SECURITY` in the Docker Compose files below. Please update <> only when necessary; changing others may lead to unexpected results. +NOTE: MC is preconfigured as the identity provider for {short-product-name} with the configuration prefixed `flow.security` and `FLOW_SECURITY` in the Docker Compose files below. Please update specified properties only when necessary; changing others may lead to unexpected results. [#modification-of-sec-preconfig] === Potential modification of security pre-configuration diff --git a/docs/modules/deploying/pages/components.adoc b/docs/modules/deploying/pages/components.adoc index 96be13b..695cdc5 100644 --- a/docs/modules/deploying/pages/components.adoc +++ b/docs/modules/deploying/pages/components.adoc @@ -3,7 +3,9 @@ :description: A '{short-product-name} deployment consists of several components: -image:flow-components.png[] +// image:flow-components.png[] +image:flow-components2.svg[] + == {short-product-name} server {short-product-name} server is a combination of several components: diff --git a/docs/modules/deploying/pages/development-deployments.adoc b/docs/modules/deploying/pages/development-deployments.adoc index 38c3f1b..75efe07 100644 --- a/docs/modules/deploying/pages/development-deployments.adoc +++ b/docs/modules/deploying/pages/development-deployments.adoc @@ -6,7 +6,8 @@ To get up and running quickly, you can bring up a fully working {short-product-n == Development setup This guide will walk you through setting up {short-product-name} in a development environment. -This setup pre-configures Management Center in Dev Mode, without needing any security credentials for logging in or using the REST API. Dev Mode should not be used in production. For more information about Dev Mode, see https://docs.hazelcast.com/management-center/latest/deploy-manage/dev-mode +This setup pre-configures Management Center in Dev Mode, without needing any security credentials for logging in or using the REST API. Dev Mode should not be used in production. For more information about Dev Mode, see https://docs.hazelcast.com/management-center/latest/deploy-manage/dev-mode + === Prerequisites diff --git a/docs/modules/describing-data-sources/pages/aws-services.adoc b/docs/modules/describing-data-sources/pages/aws-services.adoc index e0aa50c..b239427 100644 --- a/docs/modules/describing-data-sources/pages/aws-services.adoc +++ b/docs/modules/describing-data-sources/pages/aws-services.adoc @@ -314,7 +314,7 @@ service AwsBucketService { ``` #### Filename patterns when writing to S3 -When writing to S3 filenames, filename patterns are not supported (unlike when xref:aws-services.adoc#filename-patterns-when-writing-to-s3.adoc[reading]). +When writing to S3 filenames, filename patterns are not supported (unlike when reading). If you declare a filename with a pattern, an error will be thrown. diff --git a/docs/modules/describing-data-sources/pages/kafka.adoc b/docs/modules/describing-data-sources/pages/kafka.adoc index 6411ec1..827b940 100644 --- a/docs/modules/describing-data-sources/pages/kafka.adoc +++ b/docs/modules/describing-data-sources/pages/kafka.adoc @@ -281,7 +281,7 @@ stream { StockPrice.filterEach( ( StockTicker ) -> StockTicker == 'AAPL' ) } ==== Streaming from Kafka to a database Streams from Kafka can be inserted into a database (or any other writable source, such as -xref:describing-data-sources:hazelcast.adoc#writing-data-to-hazelcast[Hazelcast] or xref:describing-data-sources:kafka.adoc[Dynamo]) using a mutating query. +xref:describing-data-sources:hazelcast.adoc#writing-data-to-hazelcast[Hazelcast] or xref:describing-data-sources:aws-services.adoc#DynamoDb[Dynamo]) using a mutating query. As with all mutating queries, it's not necessary for the data from Kafka to align with the format of the From 94ac7187aeef6792c669d795d3a957ffd4094d54 Mon Sep 17 00:00:00 2001 From: Patrick McGleenon Date: Wed, 2 Oct 2024 12:44:52 +0100 Subject: [PATCH 05/11] HZX-26: updates to Working with XML (#40) # Description of change Applying some updates to the XML guide after trying out the queries --- docs/modules/guides/pages/work-with-xml.adoc | 12 ++++++------ 1 file changed, 6 insertions(+), 6 deletions(-) diff --git a/docs/modules/guides/pages/work-with-xml.adoc b/docs/modules/guides/pages/work-with-xml.adoc index a9fa944..11cc7b9 100644 --- a/docs/modules/guides/pages/work-with-xml.adoc +++ b/docs/modules/guides/pages/work-with-xml.adoc @@ -103,7 +103,7 @@ model Actor { } service CastService { - @HttpOperation(url = "http://localhost:8044/film/\{filmId}/cast", method = "GET") + @HttpOperation(url = "http://localhost:8044/film/{filmId}/cast", method = "GET") operation fetchCastForFilm(@PathVariable("filmId") filmId : FilmId):ActorList } ---- @@ -226,11 +226,11 @@ To include data from our `CastService`, we just ask for the actor information: [,taxi] ---- - find { FilmList } as (Film[]) -> { - filmId : FilmId - nameOfFilm : FilmTitle - cast : Actor[] - } +find { FilmList } as (Film[]) -> { + filmId : FilmId + nameOfFilm : FilmTitle + cast : Actor[] +} ---- Which now gives us: From b989ac3447ee0867a6c3cf847c4200ad2d2515f4 Mon Sep 17 00:00:00 2001 From: Patrick McGleenon Date: Thu, 3 Oct 2024 13:13:39 +0100 Subject: [PATCH 06/11] HZX-261: Streaming tutorial rename (#42) # Description of change Renamed to `hazelflix` to be consistent with example tutorial --- docs/modules/guides/pages/streaming-data.adoc | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/docs/modules/guides/pages/streaming-data.adoc b/docs/modules/guides/pages/streaming-data.adoc index d6b0b56..44405a9 100644 --- a/docs/modules/guides/pages/streaming-data.adoc +++ b/docs/modules/guides/pages/streaming-data.adoc @@ -16,7 +16,7 @@ Our demo has a few services running, which we'll join together to create a bespo Our services share a taxonomy used to describe the elements we can fetch: ```taxi taxonomy.taxi -namespace petflix +namespace hazelflix type FilmId inherits String type FilmTitle inherits String @@ -52,8 +52,8 @@ syntax = "proto3"; import "taxi/dataType.proto"; message NewReviewPostedMessage { - int32 filmId = 1 [(taxi.dataType)="petflix.FilmId"]; - string reviewText = 2 [(taxi.dataType)="petflix.ReviewText"]; + int32 filmId = 1 [(taxi.dataType)="hazelflix.FilmId"]; + string reviewText = 2 [(taxi.dataType)="hazelflix.ReviewText"]; } ---- From ebbd551b9d477bdc542d0e70d62bc97418c7ea72 Mon Sep 17 00:00:00 2001 From: Patrick McGleenon Date: Thu, 3 Oct 2024 13:13:55 +0100 Subject: [PATCH 07/11] HZX-260: updated sample queries (#41) # Description of change Updated sample queries so that they work when published as HTTP endpoint --- .../describing-data-sources/pages/aws-services.adoc | 4 ++-- docs/modules/guides/pages/work-with-xml.adoc | 8 ++++---- 2 files changed, 6 insertions(+), 6 deletions(-) diff --git a/docs/modules/describing-data-sources/pages/aws-services.adoc b/docs/modules/describing-data-sources/pages/aws-services.adoc index b239427..852fc92 100644 --- a/docs/modules/describing-data-sources/pages/aws-services.adoc +++ b/docs/modules/describing-data-sources/pages/aws-services.adoc @@ -478,7 +478,7 @@ import flow.aws.sqs.SqsOperation @SqsService( connectionName = "moviesConnection" ) service MovieService { @SqsOperation( queue = "movies" ) - write operation publishMoveEvent(Movie):Movie + write operation publishMovieEvent(Movie):Movie } ---- @@ -498,7 +498,7 @@ service MovieService { operation newReleases():Stream @SqsOperation( queue = "moviesToReview" ) - write operation publishMoveEvent(Movie):Movie + write operation publishMovieEvent(Movie):Movie } // Query: consume from the new releases queue, and publish to diff --git a/docs/modules/guides/pages/work-with-xml.adoc b/docs/modules/guides/pages/work-with-xml.adoc index 11cc7b9..1fdfa32 100644 --- a/docs/modules/guides/pages/work-with-xml.adoc +++ b/docs/modules/guides/pages/work-with-xml.adoc @@ -217,7 +217,7 @@ with the data type indicating where the data is sourced from: find { FilmList } as (Film[]) -> { filmId : FilmId nameOfFilm : FilmTitle -} +} [] ---- === Linking our Actor Service @@ -230,7 +230,7 @@ find { FilmList } as (Film[]) -> { filmId : FilmId nameOfFilm : FilmTitle cast : Actor[] -} +} [] ---- Which now gives us: @@ -303,7 +303,7 @@ find { FilmList } as (Film[]) -> { nameOfFilm: FilmTitle cast: Actor[] awards: Award[] -} +} [] ---- Which gives us: @@ -347,7 +347,7 @@ query.taxi:: nameOfFilm : FilmTitle awards : Award[] cast : Actor[] - } + } [] } ---- ==== From 72e93a55c32f3e5562d98a2b9d52a37b07dfb93c Mon Sep 17 00:00:00 2001 From: Oliver Howell Date: Mon, 7 Oct 2024 10:17:40 +0100 Subject: [PATCH 08/11] Update local playbook to trigger new build --- antora-playbook-local.yml | 5 +++-- 1 file changed, 3 insertions(+), 2 deletions(-) diff --git a/antora-playbook-local.yml b/antora-playbook-local.yml index 8df4414..3bc1f67 100644 --- a/antora-playbook-local.yml +++ b/antora-playbook-local.yml @@ -1,5 +1,5 @@ site: - title: Documentation + title: Documentation (with Kapa) url: http:localhost:5000 start_page: hz-flow::index.adoc robots: disallow @@ -7,6 +7,7 @@ site: docsearch_id: 'QK2EAH8GB0' docsearch_api: 'ef7bd9485eafbd75d6e8425949eda1f5' docsearch_index: 'prod_hazelcast_docs' + ai_search_id: 'ad664bf0-07e2-42e7-9150-2e1b04b15cca' content: sources: - url: . @@ -14,7 +15,7 @@ content: start_path: docs ui: bundle: - url: https://github.com/hazelcast/hazelcast-docs-ui/releases/latest/download/ui-bundle.zip #../hazelcast-docs-ui/build/ui-bundle.zip + url: ../hazelcast-docs-ui/build/ui-bundle.zip snapshot: true asciidoc: attributes: From cae426dfb6449a376863b0060fef6b69523ac43e Mon Sep 17 00:00:00 2001 From: Michael Pollett Date: Tue, 8 Oct 2024 17:04:18 +0100 Subject: [PATCH 09/11] link to get a trial license (#44) Link the, you need a license, section, to get a trial license --- docs/modules/deploying/pages/development-deployments.adoc | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/docs/modules/deploying/pages/development-deployments.adoc b/docs/modules/deploying/pages/development-deployments.adoc index 75efe07..364d703 100644 --- a/docs/modules/deploying/pages/development-deployments.adoc +++ b/docs/modules/deploying/pages/development-deployments.adoc @@ -15,7 +15,7 @@ Before you get started, you'll need the following: * https://docs.docker.com/engine/install/[Docker], and Docker Compose (installed by default with Docker) * https://docs.hazelcast.com/clc/latest/install-clc[Hazelcast CLC]. CLC is the official Hazelcast commandline tool to interact with Hazelcast clusters and create Hazelcast projects. -* Hazelcast License Key +* https://hazelcast.com/get-started/[Hazelcast License Key] === Creating your project From f4abc5a95a342545ac01d90d53f4767d1268c81e Mon Sep 17 00:00:00 2001 From: Jack Green Date: Tue, 8 Oct 2024 17:42:01 +0100 Subject: [PATCH 10/11] Fix documentation typos (#45) --- .github/CONTRIBUTING.adoc | 4 ++-- .../connecting-data-sources/pages/connecting-a-database.adoc | 2 +- .../pages/schema-publication-methods.adoc | 2 +- docs/modules/deploying/pages/data-policies.adoc | 2 +- docs/modules/describing-data-sources/pages/databases.adoc | 2 +- docs/modules/querying/pages/writing-queries.adoc | 2 +- 6 files changed, 7 insertions(+), 7 deletions(-) diff --git a/.github/CONTRIBUTING.adoc b/.github/CONTRIBUTING.adoc index 36ddba3..ca06de9 100644 --- a/.github/CONTRIBUTING.adoc +++ b/.github/CONTRIBUTING.adoc @@ -27,7 +27,7 @@ If no version number is displayed in the output, you need to {url-node}[install === Step 2. Clone a documentation project -Documentation is hosted in xref:{url-readme}#documentation-content[separate GitHub repositorites]. To work on a particular documentation project, you must fork it, clone it, and configure the `antora-playbook-local.yml` file to process your local version. +Documentation is hosted in xref:{url-readme}#documentation-content[separate GitHub repositories]. To work on a particular documentation project, you must fork it, clone it, and configure the `antora-playbook-local.yml` file to process your local version. NOTE: You can find all content repositories in the `antora-playbook.yml` file under `content.sources`. @@ -140,7 +140,7 @@ modules/ <2> ---- <1> This file tells Antora that the contents of the `modules/` folder should be processed and added to the documentation site. This file is called the {url-antora-yml}[component version descriptor file]. <2> This folder contains the content that Antora will process -<3> This folder contains any content that can't be categorized under a specfic module name. Unlike other modules, the name of this module is never displayed in the URL of the site. +<3> This folder contains any content that can't be categorized under a specific module name. Unlike other modules, the name of this module is never displayed in the URL of the site. <4> In any module, this folder contains downloadable content such as ZIP files that a user can download through a link. <5> In any module, this folder contains examples such as source code that you can include in Asciidoc pages. <6> In any module, this folder contains images that you can include in Asciidoc pages. diff --git a/docs/modules/connecting-data-sources/pages/connecting-a-database.adoc b/docs/modules/connecting-data-sources/pages/connecting-a-database.adoc index 15e94e3..71667f6 100644 --- a/docs/modules/connecting-data-sources/pages/connecting-a-database.adoc +++ b/docs/modules/connecting-data-sources/pages/connecting-a-database.adoc @@ -100,7 +100,7 @@ jdbc { # The root element for database connections For the full specification, including supported database drivers and their connection details, see xref:describing-data-sources:configuring-connections.adoc[Configure connections] === Passing sensitive data -It may not always be desirable to specify sensistive connection information directly in the config file, especially +It may not always be desirable to specify sensitive connection information directly in the config file, especially if these are being checked into source control. Environment variables can be used anywhere in the config file, following the https://github.com/lightbend/config#uses-of-substitutions[HOCON standards]. diff --git a/docs/modules/connecting-data-sources/pages/schema-publication-methods.adoc b/docs/modules/connecting-data-sources/pages/schema-publication-methods.adoc index 0e72318..3c90bcd 100644 --- a/docs/modules/connecting-data-sources/pages/schema-publication-methods.adoc +++ b/docs/modules/connecting-data-sources/pages/schema-publication-methods.adoc @@ -170,7 +170,7 @@ Cons: * OpenAPI endpoints * HTTP servers -This is a strong option for scenarios where sytems can't publish their own schemas (eg., databases), +This is a strong option for scenarios where systems can't publish their own schemas (eg., databases), or for data sources that are otherwise structureless (eg., CSV files). Additionally, using a git-backed repository for a shared glossary / taxonomy is a great way to diff --git a/docs/modules/deploying/pages/data-policies.adoc b/docs/modules/deploying/pages/data-policies.adoc index 27a372b..903baae 100644 --- a/docs/modules/deploying/pages/data-policies.adoc +++ b/docs/modules/deploying/pages/data-policies.adoc @@ -57,7 +57,7 @@ Similar to how you can use claims from your IDP, to pass through to a data sourc Policies can be defined against any data that is discoverable using a {short-product-name} query - not just the values present on the inbound claim. - A seperate subquery is executed to discover data that is needed to evaluate the policy. + A separate subquery is executed to discover data that is needed to evaluate the policy. For example: diff --git a/docs/modules/describing-data-sources/pages/databases.adoc b/docs/modules/describing-data-sources/pages/databases.adoc index 252539a..d23a14d 100644 --- a/docs/modules/describing-data-sources/pages/databases.adoc +++ b/docs/modules/describing-data-sources/pages/databases.adoc @@ -369,7 +369,7 @@ No schema migrations are performed. === Example queries When writing data from one data source into a database, it's not -neccessary for the data to align with the format of the + necessary for the data to align with the format of the persisted value. {short-product-name} will automatically adapt the incoming data to the diff --git a/docs/modules/querying/pages/writing-queries.adoc b/docs/modules/querying/pages/writing-queries.adoc index 9dc1394..6a4afaf 100644 --- a/docs/modules/querying/pages/writing-queries.adoc +++ b/docs/modules/querying/pages/writing-queries.adoc @@ -401,7 +401,7 @@ TaxiQL queries with any of the following content types headers: * `Content-Type: application/taxiql` * `Content-Type: text/plain` -This is to allow broad compatability with clients. +This is to allow broad compatibility with clients. === Large queries with server-sent events From 278c87d2a0d2b96ed79b35785d731d3bd1d4bb11 Mon Sep 17 00:00:00 2001 From: Michael Pollett Date: Tue, 8 Oct 2024 21:02:47 +0100 Subject: [PATCH 11/11] couple more typos (#46) inspired by Jack, only a couple more minor ones --- README.adoc | 4 ++-- .../connecting-data-sources/pages/connecting-a-database.adoc | 2 +- docs/modules/introduction/pages/index.adoc | 2 +- docs/modules/workspace/pages/connecting-a-disk-repo.adoc | 2 +- docs/modules/workspace/pages/overview.adoc | 4 ++-- 5 files changed, 7 insertions(+), 7 deletions(-) diff --git a/README.adoc b/README.adoc index b3597f6..25880b9 100644 --- a/README.adoc +++ b/README.adoc @@ -1,4 +1,4 @@ -= {replace with your project} Documentation += Flow Documentation // Settings: ifdef::env-github[] :warning-caption: :warning: @@ -15,7 +15,7 @@ endif::[] image:https://img.shields.io/badge/Build-Staging-yellow[link="{url-staging}"] -This repository contains the Antora components for the {replace with your project} documentation. +This repository contains the Antora components for the Flow documentation. The documentation source files are marked up with AsciiDoc. diff --git a/docs/modules/connecting-data-sources/pages/connecting-a-database.adoc b/docs/modules/connecting-data-sources/pages/connecting-a-database.adoc index 71667f6..a27c59a 100644 --- a/docs/modules/connecting-data-sources/pages/connecting-a-database.adoc +++ b/docs/modules/connecting-data-sources/pages/connecting-a-database.adoc @@ -113,7 +113,7 @@ jdbc { connectionName = another-connection jdbcDriver = POSTGRES # Defines the driver to use. See below for the possible options connectionParameters { - # .. other params omitted for bevity .. + # .. other params omitted for brevity .. password = ${postgres_password} # Reads the environment variable "postgres_password" } } diff --git a/docs/modules/introduction/pages/index.adoc b/docs/modules/introduction/pages/index.adoc index b60d61f..25e38dc 100644 --- a/docs/modules/introduction/pages/index.adoc +++ b/docs/modules/introduction/pages/index.adoc @@ -74,7 +74,7 @@ Today, when we connect two systems together, an engineer has to build a mental m [,taxi] ---- -// The id attribute in the Customer model from the Accounts sytems +// The id attribute in the Customer model from the Accounts systems // is the same as the custId attribute in the Purchase model from sales system namespace accounts { model Customer { diff --git a/docs/modules/workspace/pages/connecting-a-disk-repo.adoc b/docs/modules/workspace/pages/connecting-a-disk-repo.adoc index 30766c3..c7f87d2 100644 --- a/docs/modules/workspace/pages/connecting-a-disk-repo.adoc +++ b/docs/modules/workspace/pages/connecting-a-disk-repo.adoc @@ -29,7 +29,7 @@ NOTE: File paths that are absolute are resolved against the file system. File pa file { projects=[ { - // Optional (defaults to false). Indciates if the project can receive edits through the UI + // Optional (defaults to false). Indicates if the project can receive edits through the UI isEditable=true // Path to the folder containing a taxi.conf file. path="/opt/var/flow/schemas/taxi" diff --git a/docs/modules/workspace/pages/overview.adoc b/docs/modules/workspace/pages/overview.adoc index f854254..fc97b86 100644 --- a/docs/modules/workspace/pages/overview.adoc +++ b/docs/modules/workspace/pages/overview.adoc @@ -91,7 +91,7 @@ Another option is authoring your API specs directly in xref:introduction:about-t ---- model Customer { id : CustomerId inherits Int - firstName : CustomerFirstName inherits Sring + firstName : CustomerFirstName inherits String } service CustomerService { @@ -231,4 +231,4 @@ git { == Continue reading -Continue learning about {short-product-name} by xref:describing-data-sources:configuring-connections.adoc[connecting your data sources]. \ No newline at end of file +Continue learning about {short-product-name} by xref:describing-data-sources:configuring-connections.adoc[connecting your data sources].