From 11e0bf837149ae6b6da0149b783a4914b3e6395b Mon Sep 17 00:00:00 2001 From: sathya0 Date: Tue, 6 Aug 2024 13:47:47 -0700 Subject: [PATCH] more --- ajax/0.3.9/modules/ROOT/pages/index.adoc | 8 ++-- .../amazon-dynamodb-connector-examples.adoc | 16 ++++---- .../amazon-dynamodb-connector-studio.adoc | 4 +- .../pages/amazon-ec2-connector-examples.adoc | 12 +++--- .../amazon-kinesis-connector-examples.adoc | 6 +-- .../amazon-lambda-connector-examples.adoc | 4 +- amazon-rds/1.3/modules/ROOT/pages/index.adoc | 12 +++--- .../amazon-redshift-connector-example.adoc | 10 ++--- .../pages/amazon-s3-connector-examples.adoc | 14 +++---- .../amazon-secrets-manager-examples.adoc | 6 +-- .../pages/amazon-sns-connector-examples.adoc | 20 +++++----- .../pages/amazon-sns-connector-studio.adoc | 4 +- .../amazon-sqs-connector-config-topics.adoc | 6 +-- .../pages/amazon-sqs-connector-examples.adoc | 38 +++++++++---------- .../pages/amazon-sqs-connector-studio.adoc | 6 +-- .../azure-cosmos-db-connector-examples.adoc | 36 +++++++++--------- .../azure-data-lake-connector-examples.adoc | 34 ++++++++--------- .../azure-data-lake-connector-studio.adoc | 4 +- .../azure-event-hubs-connector-examples.adoc | 6 +-- ...t-hubs-connector-studio-configuration.adoc | 6 +-- .../2.1/modules/ROOT/pages/index.adoc | 2 +- .../azure-key-vault-connector-examples.adoc | 6 +-- ...ice-bus-management-connector-examples.adoc | 4 +- ...rvice-bus-management-connector-studio.adoc | 2 +- 24 files changed, 133 insertions(+), 133 deletions(-) diff --git a/ajax/0.3.9/modules/ROOT/pages/index.adoc b/ajax/0.3.9/modules/ROOT/pages/index.adoc index 0baeb71545..3ee62f03ef 100644 --- a/ajax/0.3.9/modules/ROOT/pages/index.adoc +++ b/ajax/0.3.9/modules/ROOT/pages/index.adoc @@ -2,7 +2,7 @@ :keywords: anypoint, connectors, ajax :page-aliases: 3.9@mule-runtime::ajax-connector.adoc -image::ajax-icon-no-caption.png[ajax icon] +image::ajax-icon-no-caption.png["Ajax icon"] The Ajax Connector creates a transport channel to send messages asynchronously to and from an Ajax server, which communicates with external Web resources. *Ajax* is an acronym for *Asynchronous JavaScript and XML*, a cluster of related technologies that facilitate the creation of asynchronous Web applications. @@ -32,7 +32,7 @@ Configure the tabs as shown. == General Tab -image::ajax-connector-general.png[Studio-ajax-gen] +image::ajax-connector-general.png["General tab in Ajax Connector configuration panel"] [%header%autowidth.spread] |=== @@ -44,7 +44,7 @@ image::ajax-connector-general.png[Studio-ajax-gen] == Advanced Tab -image::ajax-connector-advanced.png[Studio-ajax-outbound-adv] +image::ajax-connector-advanced.png["Advanced tab in Ajax Connector configuration panel"] [%header%autowidth.spread] @@ -63,7 +63,7 @@ image::ajax-connector-advanced.png[Studio-ajax-outbound-adv] == Transformers Tab -image::ajax-connector-transformers.png[Studio-ajax-trans] +image::ajax-connector-transformers.png["Transformers tab in Ajax Connector configuration panel"] [%header%autowidth.spread] |=== diff --git a/amazon-dynamodb/1.5/modules/ROOT/pages/amazon-dynamodb-connector-examples.adoc b/amazon-dynamodb/1.5/modules/ROOT/pages/amazon-dynamodb-connector-examples.adoc index fda5cb84eb..0bfeb6c565 100644 --- a/amazon-dynamodb/1.5/modules/ROOT/pages/amazon-dynamodb-connector-examples.adoc +++ b/amazon-dynamodb/1.5/modules/ROOT/pages/amazon-dynamodb-connector-examples.adoc @@ -17,7 +17,7 @@ This example shows how to use Amazon DynamoDB connector to create an Amazon Dyna The following screenshot shows an Anypoint Studio flow for creating an Amazon DynamoDB table: .Use this sample flow to create an Amazon DynamoDB table. -image::amazon-dynamodb-create-table-flow.png[Studio flow for creating an Amazon DynamoDB table] +image::amazon-dynamodb-create-table-flow.png["Studio flow for creating an Amazon DynamoDB table"] [[preliminary-steps-create]] === Preliminary Steps @@ -46,12 +46,12 @@ Configure `HTTP Listener` to listen for HTTP requests on the `/createtable` path . In the *Host* field, select `localhost`: + .Select `localhost` in the *Host* field to listen for HTTP requests on your local computer. -image::amazon-dynamodb-http-global-element.png[HTTP Listener global element] +image::amazon-dynamodb-http-global-element.png["HTTP Listener global element"] . Click *OK*. . In the *Path* field on the *Listener* tab, enter `/createtable`: + .Enter `/createtable` in the *Path* field to listen for HTTP requests on the `/createtable` path. -image::amazon-dynamodb-http-prop.png[HTTP Listener tab] +image::amazon-dynamodb-http-prop.png["HTTP Listener tab"] === Configure the Create Table Operation @@ -81,7 +81,7 @@ Configure the `Create Table` operation to create a table named `Issues`: The following screenshot shows an example of configuring a global element for Amazon DynamoDB Connector: .Configure the credentials needed to access Amazon DynamoDB in the *Connection* section of the global element. -image::amazon-dynamodb-global-element.png[Amazon DynamoDB global element] +image::amazon-dynamodb-global-element.png["Amazon DynamoDB global element"] -- + . Click *Test Connection* to confirm that Mule can connect with the Amazon DynamoDB instance: @@ -113,7 +113,7 @@ image::amazon-dynamodb-global-element.png[Amazon DynamoDB global element] The following screenshot shows an example of configuring the `Create table` operation: + .Enter values for the `Create table` operation fields. -image::amazon-dynamodb-create-table-props.png[Create table operation fields] +image::amazon-dynamodb-create-table-props.png["Create table operation fields"] + In this screenshot: @@ -274,7 +274,7 @@ Configure a `Logger` component to print to the Mule console the response generat . Enter `#[payload]` in the *Message* field on the *Logger* tab. + .Enter `#[payload]` in the *Message* field to print the operation's response to the Mule console. -image::amazon-dynamodb-logger-props.png[Logger tab] +image::amazon-dynamodb-logger-props.png["Logger tab"] === XML for the Complete Create Table Example @@ -415,7 +415,7 @@ This example shows how to use Amazon DynamoDB connector to delete an Amazon Dyna The following screenshot shows an Anypoint Studio flow for deleting an Amazon DynamoDB table: .Use this sample flow to delete an Amazon DynamoDB table. -image::amazon-dynamodb-delete-table-flow.png[Studio flow for deleting an Amazon DynamoDB table] +image::amazon-dynamodb-delete-table-flow.png["Studio flow for deleting an Amazon DynamoDB table"] === Preliminary Steps @@ -547,7 +547,7 @@ When the operation has returned all pages of data, the `LastEvaluatedKey` value The following screenshot shows an Anypoint Studio flow for scanning an Amazon DynamoDB table: .Use this sample flow to scan an Amazon DynamoDB table. -image::amazon-dynamodb-scan-flow.png[Studio flow for scanning an Amazon DynamoDB table] +image::amazon-dynamodb-scan-flow.png["Studio flow for scanning an Amazon DynamoDB table"] === Preliminary Steps diff --git a/amazon-dynamodb/1.5/modules/ROOT/pages/amazon-dynamodb-connector-studio.adoc b/amazon-dynamodb/1.5/modules/ROOT/pages/amazon-dynamodb-connector-studio.adoc index 0012a9bc79..28cf1c4935 100644 --- a/amazon-dynamodb/1.5/modules/ROOT/pages/amazon-dynamodb-connector-studio.adoc +++ b/amazon-dynamodb/1.5/modules/ROOT/pages/amazon-dynamodb-connector-studio.adoc @@ -102,7 +102,7 @@ Enter the following information on the *General* tab of the *Global Element Prop The following screenshot shows an example of configuring Basic authentication: .Basic authentication fields -image::amazon-dynamodb-basic-authentication.png[To configure authentication, select `Basic` in the *Connection* field and then complete the fields on the *General* tab.] +image::amazon-dynamodb-basic-authentication.png["Configure authentication by selecting Basic in the Connection field and then complete the fields on the General tab"] The first item shows where to specify that the connector will use Basic authentication. The second item shows the *General tab*, which contains fields related to Basic authentication. @@ -126,7 +126,7 @@ Enter the following information on the *General* tab of the global element confi The following screenshot shows an example of configuring Role authentication: .Role authentication fields -image::amazon-dynamodb-role-authentication.png[To configure Role authentication, select `Role` in the *Connection* field and then complete the fields on the *General* tab.] +image::amazon-dynamodb-role-authentication.png["Configure Role authentication by selecting Role in the Connection field and then complete the fields on the General tab"] The first item shows where to specify that the connector will use Role authentication. The second item shows the *General tab*, which contains fields related to Role authentication. diff --git a/amazon-ec2/2.5/modules/ROOT/pages/amazon-ec2-connector-examples.adoc b/amazon-ec2/2.5/modules/ROOT/pages/amazon-ec2-connector-examples.adoc index 8a953a301d..652f6ac26f 100644 --- a/amazon-ec2/2.5/modules/ROOT/pages/amazon-ec2-connector-examples.adoc +++ b/amazon-ec2/2.5/modules/ROOT/pages/amazon-ec2-connector-examples.adoc @@ -7,7 +7,7 @@ The example includes configuring global elements for Anypoint Connector for Amaz The following image shows the Mule flow for this example: -image::amazon-ec2-create-volume-usecase-flow.png[Flow for creating an EBS volume] +image::amazon-ec2-create-volume-usecase-flow.png["Flow for creating an EBS volume"] This example uses variables for some field values. You can either: @@ -45,7 +45,7 @@ amazon.sessionToken= + The following image shows an example of the *HTTP Listener* global element configuration: + -image::http-listener-config.png[Global element configuration for HTTP Listener] +image::http-listener-config.png["Global element configuration for HTTP Listener"] + .. Click *OK*. . Drag the Amazon EC2 Connector *Create volume* operation next to the *Listener* component. @@ -67,7 +67,7 @@ Region to select from the drop-down menu for the Amazon EC2 client + The following image shows an example of what the configuration looks like: + -image::mule-amazon-ec2-global-config.png[EC2 global configuration] +image::mule-amazon-ec2-global-config.png["EC2 global configuration"] + The corresponding XML looks like this: + @@ -107,7 +107,7 @@ Size of the volume in gibibytes (GiBs). This example uses `5`. + The default volume type is *Standard*. + -image::amazon-ec2-create-volume-props.png[publish message connector props] +image::amazon-ec2-create-volume-props.png["Publish message connector props"] + . Verify that your XML looks like this: + @@ -130,7 +130,7 @@ This prints the volume ID of the EBS volume. + The following image shows the *Set Payload* configuration: + -image::amazon-ec2-create-volume-payload.png[EC2 Create Volume payload transformer] +image::amazon-ec2-create-volume-payload.png["EC2 Create Volume payload transformer"] == Add and Configure the Logger Component @@ -142,7 +142,7 @@ The *Logger* component displays the volume ID that is transformed by the *Set Pa * *Level* + `INFO` + -image::amazon-ec2-create-volume-logger-props.png[EC2 create volume logger] +image::amazon-ec2-create-volume-logger-props.png["EC2 create volume logger"] + . Save the project. diff --git a/amazon-kinesis/1.0/modules/ROOT/pages/amazon-kinesis-connector-examples.adoc b/amazon-kinesis/1.0/modules/ROOT/pages/amazon-kinesis-connector-examples.adoc index e442e90ef8..e3b459f211 100644 --- a/amazon-kinesis/1.0/modules/ROOT/pages/amazon-kinesis-connector-examples.adoc +++ b/amazon-kinesis/1.0/modules/ROOT/pages/amazon-kinesis-connector-examples.adoc @@ -10,15 +10,15 @@ The following screenshots show the Anypoint Studio app flows for this example: * This flow uses the *Put Records* operation to put data into the Amazon Kinesis data stream. In this flow, *HTTP > Listener* receives a query parameter named *data* that sets the payload for the *Put Record* operation. It uses a query parameter named *partitionKey* to set the partition key value for the *Put Record* operation. + -image::amazon-kinesis-example-put-record.png[Put Record operation flow] +image::amazon-kinesis-example-put-record.png["Put Record operation flow"] + * This flow uses the *Listener* source to listen for new data records: + -image::amazon-kinesis-example-listener.png[Listener source flow] +image::amazon-kinesis-example-listener.png["Listener source flow"] + * This flow uses the *Listener* source to listen for new data records. After receiving the new records, it calls the *Checkpoint* operation to set a manual checkpoint: + -image::amazon-kinesis-example-checkpoint.png[Checkpoint operation flow] +image::amazon-kinesis-example-checkpoint.png["Checkpoint operation flow"] == XML Code for This Example diff --git a/amazon-lambda/1.1/modules/ROOT/pages/amazon-lambda-connector-examples.adoc b/amazon-lambda/1.1/modules/ROOT/pages/amazon-lambda-connector-examples.adoc index f4d89d7d03..3dee38c8d7 100644 --- a/amazon-lambda/1.1/modules/ROOT/pages/amazon-lambda-connector-examples.adoc +++ b/amazon-lambda/1.1/modules/ROOT/pages/amazon-lambda-connector-examples.adoc @@ -34,11 +34,11 @@ Configure an HTTP Listener source to initiate the Amazon Lambda Connector *Invok . In *Mule Palette*, select *Amazon Lambda Connector > Invoke*. . Drag *Invoke* to the Studio canvas to the right of the *Listener* source: + -image::amazon-lambda-simple-invoke-example.png[Simple invoke example flow in Studio] +image::amazon-lambda-simple-invoke-example.png["Simple invoke example flow in Studio"] + . Configure the *Invoke* operation: + -image::amazon-lambda-simple-invoke-example-properties.png[Simple invoke operation General properties configuration] +image::amazon-lambda-simple-invoke-example-properties.png["Simple invoke operation General properties configuration"] + In this example, the function name is supplied in the `environment.properties` file as `$config.functionName`. + diff --git a/amazon-rds/1.3/modules/ROOT/pages/index.adoc b/amazon-rds/1.3/modules/ROOT/pages/index.adoc index dab4634e27..f77730ff99 100644 --- a/amazon-rds/1.3/modules/ROOT/pages/index.adoc +++ b/amazon-rds/1.3/modules/ROOT/pages/index.adoc @@ -83,12 +83,12 @@ To use Amazon RDS Connector in your Mule application, configure a global Amazon . On the Global Configuration Elements screen, click *Create*. + The following window displays: + -image::amazon-rds-config-global-wizard.png[Global Configuration Elements Wizard] +image::amazon-rds-config-global-wizard.png["Global Configuration Elements Wizard"] + . In the Choose Global Type wizard, expand *Connector Configuration*, select *Amazon RDS Configuration*, and click *Ok*. + The following window displays: + -image::amazon-rds-create-db-instance-config.png[Create DB Instance Configure] +image::amazon-rds-create-db-instance-config.png["Create DB Instance Configure"] + . Configure the parameters according to instructions below: + @@ -253,7 +253,7 @@ Replace `x.x.x` with the version that corresponds to the connector you are using This demo Mule application creates a DB instance. -image::amazon-rds-create-db-instance-use-case-flow.png[Creating a DB instance] +image::amazon-rds-create-db-instance-use-case-flow.png["Creating a DB instance"] . Create a new Mule Project in Anypoint Studio. . Add the following properties to the `mule-artifact.properties` file for your Amazon RDS credentials and place it in the project's `src/main/resources` directory. @@ -267,7 +267,7 @@ config.region= + . Drag an HTTP Listener operation onto the canvas and configure the following parameters: + -image::amazon-rds-http-props.png[rds http config props] +image::amazon-rds-http-props.png["RDS HTTP config props"] + [%header,cols="30s,70a"] |=== @@ -323,7 +323,7 @@ output application/java |=== .. Your configuration should look like this: + -image::amazon-rds-create-db-instance-config.png[rds use case config] +image::amazon-rds-create-db-instance-config.png["RDS use case config"] + .. The corresponding XML configuration should be as follows: + @@ -354,7 +354,7 @@ image::amazon-rds-create-db-instance-config.png[rds use case config] |Allocated storage| A number representing the storage in GB. |=== + -image::amazon-rds-create-db-instance-param-config.png[create db instance parameter props] +image::amazon-rds-create-db-instance-param-config.png["Create DB instance parameter props"] + . Verify that your XML looks like this: + diff --git a/amazon-redshift/1.0/modules/ROOT/pages/amazon-redshift-connector-example.adoc b/amazon-redshift/1.0/modules/ROOT/pages/amazon-redshift-connector-example.adoc index db3e8b4431..7c10085051 100644 --- a/amazon-redshift/1.0/modules/ROOT/pages/amazon-redshift-connector-example.adoc +++ b/amazon-redshift/1.0/modules/ROOT/pages/amazon-redshift-connector-example.adoc @@ -188,10 +188,10 @@ If this doesn't work, add the Amazon Maven repository to your `pom.xml` file and The following images show the Amazon S3 and Amazon Redshift configurations: .Amazon S3 Configuration -image::s3-configuration.png[Amazon S3 configuration] +image::s3-configuration.png["Amazon S3 configuration"] .Amazon Redshift Configuration -image::redshift-configuration.png[Amazon Redshift configuration] +image::redshift-configuration.png["Amazon Redshift configuration"] [[prepare-data-file]] == Prepare the Data File @@ -217,15 +217,15 @@ The following screenshots show the Anypoint Studio app flows for this example: * This flow creates the Amazon Redshift table, Amazon S3 bucket and object: + -image::create-flow.png[Create the Redshift table Amazon S3 bucket and Amazon S3 object flow] +image::create-flow.png["Create the Redshift table Amazon S3 bucket and Amazon S3 object flow"] + * This flow executes the `COPY` command, which leverages the Amazon Redshift massively parallel processing (MPP) architecture to load data in parallel from a file in an Amazon S3 bucket: + -image::execute-flow.png[Execute the Copy command flow] +image::execute-flow.png["Execute the Copy command flow"] + * This flow deletes the Amazon Redshift table and Amazon S3 bucket: + -image:delete-flow.png[Delete the Amazon Redshift table and Amazon S3 bucket flow] +image:delete-flow.png["Delete the Amazon Redshift table and Amazon S3 bucket flow"] == Run the Example diff --git a/amazon-s3/6.3/modules/ROOT/pages/amazon-s3-connector-examples.adoc b/amazon-s3/6.3/modules/ROOT/pages/amazon-s3-connector-examples.adoc index 5a8c93333b..f6319dbd8e 100644 --- a/amazon-s3/6.3/modules/ROOT/pages/amazon-s3-connector-examples.adoc +++ b/amazon-s3/6.3/modules/ROOT/pages/amazon-s3-connector-examples.adoc @@ -35,7 +35,7 @@ image::flow.png["Studio 7 Visual Studio Icon Flow. The flow shows HTTP Listener, . Configure the global element properties, and then click *OK*. . Configure the connector parameters: + -image::s3_config.png[Connector Properties - General tab] +image::s3_config.png["Connector Properties General tab"] + [%header%autowidth.spread] |=== @@ -49,7 +49,7 @@ image::s3_config.png[Connector Properties - General tab] + . Drag the *Create Bucket* operation into the flow. + -image::create_bucket.png[S3 Connector Create Bucket in S3 General Tab] +image::create_bucket.png["S3 Connector Create Bucket in S3 General Tab"] + [%header%autowidth.spread] |=== @@ -62,7 +62,7 @@ image::create_bucket.png[S3 Connector Create Bucket in S3 General Tab] + . Add an HTTP > Connector to request the MuleSoft logo from MuleSoft. + -image::request.png[HTTP > Connector - General Tab] +image::request.png["HTTP Connector General Tab"] + [%header%autowidth.spread] |=== @@ -74,7 +74,7 @@ image::request.png[HTTP > Connector - General Tab] + . Drag the *Put Object* operation into the flow to create the requested MuleSoft logo in the selected Amazon S3 bucket. + -image::put_object.png[S3 Connector Create Logo Object in S3 Bucket General Tab] +image::put_object.png["S3 Connector Create Logo Object in S3 Bucket General Tab"] + [%header%autowidth.spread] |=== @@ -89,7 +89,7 @@ image::put_object.png[S3 Connector Create Logo Object in S3 Bucket General Tab] + . Drag the *Get Object* operation into the flow to get the newly created MuleSoft logo image object from the bucket. + -image::get_object.png[S3 Connector Get Image General Tab] +image::get_object.png["S3 Connector Get Image General Tab"] + [%header%autowidth.spread] |=== @@ -102,7 +102,7 @@ image::get_object.png[S3 Connector Get Image General Tab] + . Drag the *Delete Object* operation into the flow to delete the object. Because the *Delete Object* operation’s return type is void, the payload contains the object returned by the *Get Object* operation. + -image::delete_object.png[S3 Connector Delete S3 Bucket General Tab] +image::delete_object.png["S3 Connector Delete S3 Bucket General Tab"] + [%header%autowidth.spread] |=== @@ -115,7 +115,7 @@ image::delete_object.png[S3 Connector Delete S3 Bucket General Tab] + . Drag the *Delete Bucket* operation into the flow to delete the bucket. Because the *Delete Bucket* operation’s return type is void, the payload contains the object returned by the *Get Object* operation. + -image::delete_bucket.png[S3 Connector Delete S3 Bucket General Tab] +image::delete_bucket.png["S3 Connector Delete S3 Bucket General Tab"] + [%header%autowidth.spread] |=== diff --git a/amazon-secrets-manager-properties-provider/1.0/modules/ROOT/pages/amazon-secrets-manager-examples.adoc b/amazon-secrets-manager-properties-provider/1.0/modules/ROOT/pages/amazon-secrets-manager-examples.adoc index ee9a6593a1..6e278f8381 100644 --- a/amazon-secrets-manager-properties-provider/1.0/modules/ROOT/pages/amazon-secrets-manager-examples.adoc +++ b/amazon-secrets-manager-properties-provider/1.0/modules/ROOT/pages/amazon-secrets-manager-examples.adoc @@ -23,7 +23,7 @@ Anypoint Connectors can use Amazon Secrets Manager Properties Provider in their |Role ARN a| String | Uniquely identifies a role to gain cross-account access. | | |=== -image::amazon-secrets-manager-properties.png[Example of Amazon Secrets Manager Properties Provider in Global Element Properties window of Anypoint Studio] +image::amazon-secrets-manager-properties.png["Example of Amazon Secrets Manager Properties Provider in Global Element Properties window of Anypoint Studio"] In the *Configuration XML* editor, the configuration looks like this: @@ -50,7 +50,7 @@ aws secretsmanager create-secret --name aws-credentials --secret-string '{"acc . Set *Access Key* to `${aws-secrets::aws-credentials}.accessKey`. . Set *Secret Key* to `${aws-secrets::aws-credentials}.secretKey`. -image::aws-sm-json-1.png[Example of Amazon Secrets Manager Properties Provider configuration for Amazon S3 Connector with JSON expression] +image::aws-sm-json-1.png["Example of Amazon Secrets Manager Properties Provider configuration for Amazon S3 Connector with JSON expression"] In the *Configuration XML* editor, the configuration looks like this: @@ -77,7 +77,7 @@ In the *Amazon S3 Configuration* global element properties window: . Set *Access Key* to `${aws-secrets::accessKey}`. . Set *Secret Key* to `${aws-secrets::secretKey}`. -image::aws-sm-json-2.png[Example of Amazon Secrets Manager Properties Provider configuration for Amazon S3 Connector with a string value] +image::aws-sm-json-2.png["Example of Amazon Secrets Manager Properties Provider configuration for Amazon S3 Connector with a string value"] In the *Configuration XML* editor, the configuration looks like this: diff --git a/amazon-sns/4.7/modules/ROOT/pages/amazon-sns-connector-examples.adoc b/amazon-sns/4.7/modules/ROOT/pages/amazon-sns-connector-examples.adoc index eb39f3734d..0a5e5dee01 100644 --- a/amazon-sns/4.7/modules/ROOT/pages/amazon-sns-connector-examples.adoc +++ b/amazon-sns/4.7/modules/ROOT/pages/amazon-sns-connector-examples.adoc @@ -26,7 +26,7 @@ You can use Amazon SNS to send notification messages to one or more HTTP or HTTP The following diagram shows an example flow for this use case: -image::amazon-sns-use-case-flow.png[send-messages-to-SQS,width=40%] +image::amazon-sns-use-case-flow.png["Send messages to SQS",width=40%] Before you try the example, access Anypoint Studio, and verify that the Mule Palette displays entries for *Amazon SNS* and *Amazon SQS*. If not, follow the instructions in xref:amazon-sns-connector-studio.adoc#install-connector[Install the Connector in Studio Using Exchange] to install both connectors. @@ -37,18 +37,18 @@ To configure a Mule app that uses a connector to send Amazon SNS messages to Ama + . In the Mule Palette, search for *HTTP*, and select the *Listener* operation: + -image::amazon-sns-select-listener.png[select-listener] +image::amazon-sns-select-listener.png["Select listener"] + . Drag the *Listener* operation to the canvas. . In the *Listener* configuration, click *+* next to the *Connector configuration* field to add a global element. . Accept the defaults. . In the *Listener* configuration *General* tab, set the *Display Name* field to `HTTP` and the *Path* field to`/''. + -image::amazon-sns-http-properties.png[http-properties] +image::amazon-sns-http-properties.png["HTTP properties"] + . In the Mule Palette, search for *SNS* and select the *Publish* operation: + -image::amazon-sns-publish-selection.png[publish-selection,width=50%] +image::amazon-sns-publish-selection.png["Publish selection",width=50%] + . Drag the *Publish* operation to the right of the *HTTP Listener* component. . In the *Publish* configuration, click *+* next to the *Connector configuration* field to add a global element. @@ -65,7 +65,7 @@ image::amazon-sns-publish-selection.png[publish-selection,width=50%] + The following image shows a sample global element for the Publish operation: + -image::amazon-sns-global-config.png[sns-global-configuration,width=70%] +image::amazon-sns-global-config.png["SNS global configuration",width=70%] + . Select the *Configuration XML* tab to view the corresponding XML: + @@ -103,7 +103,7 @@ image::amazon-sns-global-config.png[sns-global-configuration,width=70%] + The following image shows a sample configuration for the Publish Message operation (General Tab): + -image::amazon-sns-publish-message-general.png[sns-publish-message-props] +image::amazon-sns-publish-message-general.png["SNS publish message props"] . Configure parameters in Advanced Tab: + [%header%autowidth.spread] @@ -120,7 +120,7 @@ image::amazon-sns-publish-message-general.png[sns-publish-message-props] + The following image shows a sample configuration for the Publish Message operation (Advanced Tab): + -image::amazon-sns-publish-message-advanced.png[sns-publish-message-props] +image::amazon-sns-publish-message-advanced.png["SNS publish message props"] + . Select the *Configuration XML* tab to view the corresponding XML: + @@ -152,7 +152,7 @@ Using the same example, add a Logger component to display the data processed by |*Level* |`INFO` |=== + -image::amazon-sns-logger.png[sns-logger] +image::amazon-sns-logger.png["SNS logger"] [[receive-sns]] === Receive SNS Messages @@ -177,7 +177,7 @@ Using the same example, add a second flow that enables Amazon SQS to receive mes + The following image shows a sample global element for the Receive messages operation: + -image::amazon-sns-sqs-global-config.png[sns-sqs-config,width=70%] +image::amazon-sns-sqs-global-config.png["SNS SQS config",width=70%] + . Select the *Configuration XML* tab to view the corresponding XML: + @@ -212,7 +212,7 @@ Verify that the SQS queue in the configuration is **subscribed** to the SNS topi + For example: + -image::amazon-sns-sqs-config.png[sns-sqs-config] +image::amazon-sns-sqs-config.png["SNS SQS config"] + . Select the *Configuration XML* tab to view the corresponding XML: + diff --git a/amazon-sns/4.7/modules/ROOT/pages/amazon-sns-connector-studio.adoc b/amazon-sns/4.7/modules/ROOT/pages/amazon-sns-connector-studio.adoc index 6cbd542d52..60f3a6483e 100644 --- a/amazon-sns/4.7/modules/ROOT/pages/amazon-sns-connector-studio.adoc +++ b/amazon-sns/4.7/modules/ROOT/pages/amazon-sns-connector-studio.adoc @@ -39,12 +39,12 @@ To use Amazon SNS Connector in your Mule app, configure a global Amazon SNS elem . Click the *Global Elements* tab at the base of the canvas, and then click *Create*. . In the *Choose Global Type* window, expand *Connector Configuration*, and click *Amazon SNS configuration*: + -image::amazon-sns-config-global-wizard.png[Global Element Configuration Wizard] +image::amazon-sns-config-global-wizard.png["Global Element Configuration Wizard"] + . Click *OK*. . Specify the global element properties: + -image::amazon-sns-global-config.png[Global Element Properties] +image::amazon-sns-global-config.png["Global Element Properties"] + . Configure the parameters as follows: + diff --git a/amazon-sqs/5.11/modules/ROOT/pages/amazon-sqs-connector-config-topics.adoc b/amazon-sqs/5.11/modules/ROOT/pages/amazon-sqs-connector-config-topics.adoc index 3559f9b86e..72b2697bf3 100644 --- a/amazon-sqs/5.11/modules/ROOT/pages/amazon-sqs-connector-config-topics.adoc +++ b/amazon-sqs/5.11/modules/ROOT/pages/amazon-sqs-connector-config-topics.adoc @@ -25,7 +25,7 @@ To control polling behavior globally, use the Default Wait Time field on the Adv In the following screenshot, the Default Wait Time field is set to `20`. This value directs the connector to use long polling and to wait 20 seconds before returning an empty response. -image::amazon-sqs-config-wait-time.png[Default Wait Time Configuration] +image::amazon-sqs-config-wait-time.png["Default Wait Time Configuration"] === Control Polling Behavior Locally @@ -41,7 +41,7 @@ If you want the app to receive messages on the primary node only: . In Studio, select the *Receive Messages* source for your flow. . On the *Advanced* sub-tab of the *Receive messages* tab, select the *Primary node only* checkbox: + -image::amazon-sqs-primary-node-only.png[Read message configuration Advanced tab with Primary node only option selected] +image::amazon-sqs-primary-node-only.png["Read message configuration Advanced tab with Primary node only option selected"] The multi-node cluster behavior depends on the selected SQS queue type. When using the Standard queue, SQS guarantees that a message is delivered at least once. However, duplicate messages might be introduced in the queue. To avoid duplication of messages in the queue, you can use a FIFO (First-In-First-Out) queue to ensure that a message is delivered only once and that it remains available until a consumer processes and deletes it. @@ -71,7 +71,7 @@ To enable support for managing large message payloads in Amazon SQS Connector, f * *Expression or Bean Reference* * *Edit Inline* + -image::amazon-sqs-studio-large-payload-support.png[Configure Large Payload Support] +image::amazon-sqs-studio-large-payload-support.png["Configure Large Payload Support"] . Configure the parameters as follows: + [%header%autowidth.spread] diff --git a/amazon-sqs/5.11/modules/ROOT/pages/amazon-sqs-connector-examples.adoc b/amazon-sqs/5.11/modules/ROOT/pages/amazon-sqs-connector-examples.adoc index d76a96b4d2..47cf07ca2e 100644 --- a/amazon-sqs/5.11/modules/ROOT/pages/amazon-sqs-connector-examples.adoc +++ b/amazon-sqs/5.11/modules/ROOT/pages/amazon-sqs-connector-examples.adoc @@ -15,7 +15,7 @@ This example sends a message with metadata to an Amazon SQS queue and then recei * The first flow sends a message with metadata and gets the count of the messages in the queue to validate that the message was sent. * The second flow receives the message and logs the message body. -image::amazon-sqs-message-flow.png[Send and Receive Messages flows] +image::amazon-sqs-message-flow.png["Send and Receive Messages flows"] === Configure the First Flow @@ -89,7 +89,7 @@ Add the *Send message* operation to send the message to the specified queue: + The following image shows an example of the *Send Message* configuration of the global element: + -image::amazon-sqs-studio-global-config-new.png[Send Message global element configuration window] +image::amazon-sqs-studio-global-config-new.png["Send Message global element configuration window"] + . In the properties window, configure the following fields: + @@ -106,7 +106,7 @@ If provided, the value of this field takes precedence over the value of the *Def + The following image shows an example of the *Send Message* configuration in the properties window: + -image::amazon-sqs-send-message.png[Send Message properties window configuration] +image::amazon-sqs-send-message.png["Send Message properties window configuration"] ==== Add the First Logger Component @@ -125,7 +125,7 @@ Add the first *Logger* component to display the response in the Mule console: + The following image shows an example of the *Logger* configuration in the properties window: + -image::amazon-sqs-logger.png[Logger properties window configuration] +image::amazon-sqs-logger.png["Logger properties window configuration"] ==== Add the Get Approximate Number of Messages Operation @@ -145,7 +145,7 @@ Add the *Get approximate number of messages* operation to obtain the number of m + The following image shows an example of the *Get approximate number of messages* configuration in the properties window: + -image::amazon-sqs-get-message-count.png[Get Approximate Number of Messages properties window configuration] +image::amazon-sqs-get-message-count.png["Get Approximate Number of Messages properties window configuration"] ==== Add the Second Logger Component @@ -164,7 +164,7 @@ Add the second *Logger* component to display the message count in the Mule conso + The following image shows an example of the *Logger* configuration in the properties window: + -image::amazon-sqs-logger2.png[Logger properties window configuration] +image::amazon-sqs-logger2.png["Logger properties window configuration"] === Configure the Second Flow @@ -191,7 +191,7 @@ If provided, the value of this field takes precedence over the value of the *Def + The following image shows an example of the *Receive messages* configuration in the properties window: + -image::amazon-sqs-receive-message.png[Receive Messages properties window configuration] +image::amazon-sqs-receive-message.png["Receive Messages properties window configuration"] ==== Add the Logger Component @@ -210,7 +210,7 @@ Add the *Logger* component to display the message in the Mule console: + The following image shows an example of the *Logger* configuration in the properties window: + -image::amazon-sqs-logger-receive.png[Logger properties window configuration] +image::amazon-sqs-logger-receive.png["Logger properties window configuration"] === XML for This Example @@ -297,7 +297,7 @@ This example sends a large message with metadata to an Amazon SQS queue that ref * The first flow sends a large message with metadata and gets the count of the messages in the queue to validate that the large message was sent. * The second flow receives the large message and logs the message body. -image::amazon-sqs-large-message-flow.png[Send and Receive Large Messages flows] +image::amazon-sqs-large-message-flow.png["Send and Receive Large Messages flows"] === Before You Begin @@ -334,7 +334,7 @@ The default is your user home directory. + In this example, the file is read from `src/main/resources` of the Mule application: + -image::amazon-sqs-file-global-config.png[Working Directory configuration window for Read operation] +image::amazon-sqs-file-global-config.png["Working Directory configuration window for Read operation"] + . In the properties window, configure the following fields: + @@ -348,7 +348,7 @@ image::amazon-sqs-file-global-config.png[Working Directory configuration window + The following image shows an example of the *Read* configuration in the properties window: + -image::amazon-sqs-file-config.png[Read properties window configuration] +image::amazon-sqs-file-config.png["Read properties window configuration"] ==== Add the Transform Message Component @@ -408,7 +408,7 @@ Add the *Send message* operation to send the large message to the specified queu + The following image shows an example of the *Send Message* configuration of the global element: + -image::amazon-sqs-studio-global-config-new.png[Send Message global element configuration window] +image::amazon-sqs-studio-global-config-new.png["Send Message global element configuration window"] + . In the *Connection* section, click the *Advanced* tab. . In the *Large Payload Support* field, select one of the following: @@ -428,7 +428,7 @@ Configure the fields as follows: + The following image shows an example of the *Large Payload Support* fields configuration: + -image::amazon-sqs-studio-large-payload-support.png[Large Payload Support fields configuration window] +image::amazon-sqs-studio-large-payload-support.png["Large Payload Support fields configuration window"] + . In the properties window, configure the following fields: + @@ -445,7 +445,7 @@ If provided, the value of this field takes precedence over the value of the *Def + The following image shows an example of the *Send Message* configuration in the properties window: + -image::amazon-sqs-send-message.png[Send Message properties window configuration] +image::amazon-sqs-send-message.png["Send Message properties window configuration"] ==== Add the First Logger Component @@ -464,7 +464,7 @@ Add the first *Logger* component to display the response in the Mule console: + The following image shows an example of the *Logger* configuration in the properties window: + -image::amazon-sqs-logger.png[Logger properties window configuration] +image::amazon-sqs-logger.png["Logger properties window configuration"] ==== Add the Get Approximate Number of Messages Operation @@ -484,7 +484,7 @@ Add the *Get approximate number of messages* operation to obtain the number of m + The following image shows an example of the *Get approximate number of messages* configuration in the properties window: + -image::amazon-sqs-get-message-count.png[Get Approximate Number of Messages properties window configuration] +image::amazon-sqs-get-message-count.png["Get Approximate Number of Messages properties window configuration"] ==== Add the Second Logger Component @@ -503,7 +503,7 @@ Add the second *Logger* component to display the message count in the Mule conso + The following image shows an example of the *Logger* configuration in the properties window: + -image::amazon-sqs-logger2.png[Logger properties window configuration] +image::amazon-sqs-logger2.png["Logger properties window configuration"] === Configure the Second Flow @@ -530,7 +530,7 @@ If provided, the value of this field takes precedence over the value of the *Def + The following image shows an example of the *Receive messages* configuration in the properties window: + -image::amazon-sqs-receive-message.png[Receive Messages properties window configuration] +image::amazon-sqs-receive-message.png["Receive Messages properties window configuration"] ==== Add the Logger Component @@ -549,7 +549,7 @@ Add the *Logger* component to display the large message in the Mule console: + The following image shows an example of the *Logger* configuration in the properties window: + -image::amazon-sqs-logger-receive.png[Logger properties window configuration] +image::amazon-sqs-logger-receive.png["Logger properties window configuration"] === XML for This Example diff --git a/amazon-sqs/5.11/modules/ROOT/pages/amazon-sqs-connector-studio.adoc b/amazon-sqs/5.11/modules/ROOT/pages/amazon-sqs-connector-studio.adoc index a1bc479458..fca038a648 100644 --- a/amazon-sqs/5.11/modules/ROOT/pages/amazon-sqs-connector-studio.adoc +++ b/amazon-sqs/5.11/modules/ROOT/pages/amazon-sqs-connector-studio.adoc @@ -76,7 +76,7 @@ Number of messages to retrieve URL of the queue in which to receive messages . If you are using the *Receive messages* source in cluster mode and want to receive messages on the primary node only, select the *Primary node only* checkbox on the *Receive messages > Advanced* tab: + -image::amazon-sqs-primary-node-only.png[Read message configuration Advanced tab with Primary node only option selected] +image::amazon-sqs-primary-node-only.png["Read message configuration Advanced tab with Primary node only option selected"] + For more information, see xref:amazon-sqs-connector-config-topics.adoc#node-behavior[Receiving Messages in Cluster Mode]. @@ -92,12 +92,12 @@ You can either enter your credentials into the global configuration properties o . To create a global element for the connector, click the *Global Elements* tab at the base of the canvas, and then click *Create*. . In the *Choose Global Type* window, expand *Connector Configuration* and select *Amazon SQS Configuration*: + -image::amazon-sqs-studio-select-global.png[Select Connector for Global Element] +image::amazon-sqs-studio-select-global.png["Select Connector for Global Element"] + . Click *OK*. . Specify the global element properties: + -image::amazon-sqs-studio-global-config-new.png[Configure Global Element] +image::amazon-sqs-studio-global-config-new.png["Configure Global Element"] + [%header%autowidth.spread] |=== diff --git a/azure-cosmos-db/1.0/modules/ROOT/pages/azure-cosmos-db-connector-examples.adoc b/azure-cosmos-db/1.0/modules/ROOT/pages/azure-cosmos-db-connector-examples.adoc index f244644615..d4825b4212 100644 --- a/azure-cosmos-db/1.0/modules/ROOT/pages/azure-cosmos-db-connector-examples.adoc +++ b/azure-cosmos-db/1.0/modules/ROOT/pages/azure-cosmos-db-connector-examples.adoc @@ -48,7 +48,7 @@ Outputs the results of the *Create Document* operation in JSON format * *Logger* + Shows the HTTP response from the *Create Document* operation -image::create-document.png[Studio Flow for the Create Document operation] +image::create-document.png["Studio Flow for the Create Document operation"] === XML for This Example @@ -136,7 +136,7 @@ Outputs the results of the *Get Document* operation in JSON format * *Logger* + Shows the HTTP response from the *Get Document* operation -image::get-document.png[Studio Flow for the Get Document operation] +image::get-document.png["Studio Flow for the Get Document operation"] === XML for This Example @@ -218,7 +218,7 @@ Outputs the results of the *Replace Document* operation in JSON format * *Logger* + Shows the HTTP response from the *Replace Document* operation -image::replace-document.png[Studio Flow for the Replace Document operation] +image::replace-document.png["Studio Flow for the Replace Document operation"] === XML for This Example @@ -307,7 +307,7 @@ Outputs the results of the *List Documents* operation in JSON format * *Logger* + Shows the HTTP response from the *List Documents* operation -image::list-documents.png[Studio Flow for the List Documents operation] +image::list-documents.png["Studio Flow for the List Documents operation"] === XML for This Example @@ -388,7 +388,7 @@ Outputs the results of the *Query Documents* operation in JSON format * *Logger* + Shows the HTTP response from the *Query Documents* operation -image::query-documents.png[Studio Flow for the Query Documents operation] +image::query-documents.png["Studio Flow for the Query Documents operation"] === XML for This Example @@ -482,7 +482,7 @@ Deletes an existing document from a collection * *Logger* + Shows the HTTP response from the *Delete Document* operation -image::delete-document.png[Studio Flow for the Delete Document operation] +image::delete-document.png["Studio Flow for the Delete Document operation"] === XML for This Example @@ -554,7 +554,7 @@ Outputs the results of the *Create Stored Procedure* operation in JSON format * *Logger* + Shows the HTTP response from the *Create Stored Procedure* operation -image::create-stored-procedure.png[Studio Flow for the Create Stored Procedure operation] +image::create-stored-procedure.png["Studio Flow for the Create Stored Procedure operation"] === XML for This Example @@ -645,7 +645,7 @@ Outputs the results of the *Execute Stored Procedure* operation in JSON format * *Logger* + Shows the HTTP response from the *Execute Stored Procedure* operation -image::execute-stored-procedure.png[Studio Flow for the Execute Stored Procedure operation] +image::execute-stored-procedure.png["Studio Flow for the Execute Stored Procedure operation"] === XML for This Example @@ -734,7 +734,7 @@ Outputs the results of the *Replace Stored Procedure* operation in JSON format * *Logger* + Shows the HTTP response from the *Replace Stored Procedure* operation -image::replace-stored-procedure.png[Studio Flow for the Replace Stored Procedure operation] +image::replace-stored-procedure.png["Studio Flow for the Replace Stored Procedure operation"] === XML for This Example @@ -823,7 +823,7 @@ Outputs the results of the *List Stored Procedures* operation in JSON format * *Logger* + Shows the HTTP response from the *List Stored Procedures* operation -image::list-stored-procedures.png[Studio Flow for the List Stored Procedures operation] +image::list-stored-procedures.png["Studio Flow for the List Stored Procedures operation"] === XML for This Example @@ -897,7 +897,7 @@ Deletes an existing stored procedure from a collection * *Logger* + Shows the HTTP response from the *Delete Stored Procedure* operation -image::delete-stored-procedure.png[Studio Flow for the Delete Stored Procedure operation] +image::delete-stored-procedure.png["Studio Flow for the Delete Stored Procedure operation"] === XML for This Example @@ -966,7 +966,7 @@ Outputs the results of the *Create Permission* operation in JSON format * *Logger* + Shows the HTTP response from the *Create Permission* operation -image::create-permission.png[Studio Flow for the Create Permission operation] +image::create-permission.png["Studio Flow for the Create Permission operation"] === XML for This Example @@ -1051,7 +1051,7 @@ Outputs the results of the *Get Permission* operation in JSON format * *Logger* + Shows the HTTP response from the *Get Permission* operation -image::get-permission.png[Studio Flow for the Get Permission operation] +image::get-permission.png["Studio Flow for the Get Permission operation"] === XML for This Example @@ -1130,7 +1130,7 @@ Outputs the results of the *Replace Permission* operation in JSON format * *Logger* + Shows the HTTP response from the *Replace Permission* operation -image::replace-permission.png[Studio Flow for the Replace Permission operation] +image::replace-permission.png["Studio Flow for the Replace Permission operation"] === XML for This Example @@ -1215,7 +1215,7 @@ Outputs the results of the *List Permissions* operation in JSON format * *Logger* + Shows the HTTP response from the *List Permissions* operation -image::list-permissions.png[Studio Flow for the List Permissions operation] +image::list-permissions.png["Studio Flow for the List Permissions operation"] === XML for This Example @@ -1289,7 +1289,7 @@ Deletes a permission from a database * *Logger* + Shows the HTTP response from the *Delete Permission* operation -image::delete-permission.png[Studio Flow for the Delete Permission operation] +image::delete-permission.png["Studio Flow for the Delete Permission operation"] === XML for This Example @@ -1363,7 +1363,7 @@ Clears the permissions * *Logger* + Shows the HTTP response from the *List Documents* operation -image::list-documents-resource-token.png[Studio Flow for the List Documents Resource Token operation] +image::list-documents-resource-token.png["Studio Flow for the List Documents Resource Token operation"] === XML for This Example @@ -1447,7 +1447,7 @@ Clears the permissions * *Logger* + Shows the HTTP response from the *List Stored Procedures* operation -image::list-stored-procedures-resource-token.png[Studio Flow for the List Stored Procedures operation] +image::list-stored-procedures-resource-token.png["Studio Flow for the List Stored Procedures operation"] === XML for This Example diff --git a/azure-data-lake/1.0/modules/ROOT/pages/azure-data-lake-connector-examples.adoc b/azure-data-lake/1.0/modules/ROOT/pages/azure-data-lake-connector-examples.adoc index 334ce8ae13..a17d03967e 100644 --- a/azure-data-lake/1.0/modules/ROOT/pages/azure-data-lake-connector-examples.adoc +++ b/azure-data-lake/1.0/modules/ROOT/pages/azure-data-lake-connector-examples.adoc @@ -32,11 +32,11 @@ This example shows you how to create a file system in Azure Data Lake Storage, a . In the *Listener* configuration, click *+* next to the *Connector configuration* field to add a global element. . Accept the defaults and click *OK*. + -image::azure-data-lake-create-file-system-setting-up-the-flow-1.png[HTTP Listener with default values] +image::azure-data-lake-create-file-system-setting-up-the-flow-1.png["HTTP Listener with default values"] + . In the *General* properties tab, set the *Path* field to `create-fs`: -image::azure-data-lake-create-file-system-setting-up-the-flow-2.png[HTTP Listener General properties tab with Path value set to create-fs] +image::azure-data-lake-create-file-system-setting-up-the-flow-2.png["HTTP Listener General properties tab with Path value set to create fs"] [[configure-file-system-operation]] === Configure the Create File System Operation @@ -46,12 +46,12 @@ image::azure-data-lake-create-file-system-setting-up-the-flow-2.png[HTTP Listene . In the configuration screen for the operation, click the plus sign (*+*) next to the *Connector configuration* field to access the global element configuration fields. . On the *General* tab, in *Connection*, select the authentication method to configure and enter all the required values: + -image::azure-data-lake-create-file-system-filling-values-1.png[Create File System Config with Shared Access Signature connection values] +image::azure-data-lake-create-file-system-filling-values-1.png["Create File System Config with Shared Access Signature connection values"] + . Click *OK*. . On the *General* properties tab, in the *File System* field, enter a name for the file system: + -image::azure-data-lake-create-file-system-filling-values-2.png[Create File System operation General properties tab] +image::azure-data-lake-create-file-system-filling-values-2.png["Create File System operation General properties tab"] + . Save the project. . Open a browser and send a GET request to `localhost:8081/create-fs` to test your app. @@ -109,7 +109,7 @@ Follow these steps to create and configure an HTTP *Listener* source to initiate . Accept the defaults and click *OK*. . In the *General* properties tab, set the *Path* field to `create-path`: -image::azure-data-lake-create-path-setting-up-the-flow-2.png[HTTP Listener General properties tab] +image::azure-data-lake-create-path-setting-up-the-flow-2.png["HTTP Listener General properties tab"] [[configure-create-path]] === Configure the Create Path operation @@ -121,7 +121,7 @@ Follow these steps to create a file or directory on the provided Azure Data Lake . In the *Create/Rename Path* configuration, click *+* next to the *Connector Configuration* field to add a connection configuration. . On the *General* tab, in *Connection*, select the authentication method to configure and enter all the required values: + -image::azure-data-lake-create-file-system-filling-values-1.png[Create File System Config with Shared Access Signature connection values] +image::azure-data-lake-create-file-system-filling-values-1.png["Create File System Config with Shared Access Signature connection values"] + . Click *OK*. . In the *File System* field, enter the value for your target Azure Data Lake Storage file system. + @@ -129,7 +129,7 @@ If the file system does not exist, an error is thrown. . In *Path*, enter your desired path name. . In *Resource*, enter either `file` or `directory`: + -image::azure-data-lake-create-path-filling-values.png[General properties configuration for Create/Rename Path operation] +image::azure-data-lake-create-path-filling-values.png["General properties configuration for Create/Rename Path operation"] + . Save the project. . Test the app by sending a GET request to `localhost:8081/create-path`. @@ -187,7 +187,7 @@ Follow these steps to create and configure an HTTP *Listener* source to initiate . Accept the defaults and click *OK*. . In the *General* properties tab, set the *Path* field to `update-path`: -image::azure-data-lake-update-path-setting-up-the-flow-2.png[Listener General properties configuration] +image::azure-data-lake-update-path-setting-up-the-flow-2.png["Listener General properties configuration"] [[add-file-connector]] === Add and Configure File Connector @@ -200,14 +200,14 @@ Add File Connector to your project and configure the *Read* operation to read fi . Click *Add*. . Click *Finish*. + -image::azure-data-lake-update-path-adding-file-connector.png[Add File Connector from Exchange to your flow] +image::azure-data-lake-update-path-adding-file-connector.png["Add File Connector from Exchange to your flow"] + . In the *Mule Palette* view, select *File > Read* and drag the *Read* operation onto the flow. . In *General* properties tab for the *Read* operation, click *+* next to the *Connector Configuration* field to add a connection configuration. . Enter the values and click *OK*. . In the *General* properties for the *Read* operation, enter the absolute path to the file upload in *File Path*: + -image::azure-data-lake-update-path-using-the-file-connector-2.png[Read operation General properties configuration] +image::azure-data-lake-update-path-using-the-file-connector-2.png["Read operation General properties configuration"] [[configure-set-variable]] === Configure the Set Variable Component @@ -222,7 +222,7 @@ Configure a *Set Variable* component to store and use a variable later in the fl `#[attributes.size]`+ This saves the length of the file variable for the next operations. -image::azure-data-lake-update-path-using-the-file-connector-3.png[General properties configuration for the Set Variable component] +image::azure-data-lake-update-path-using-the-file-connector-3.png["General properties configuration for the Set Variable component"] [[update-path-append]] === Configure the Update Path Operation with Append @@ -246,7 +246,7 @@ Value for your target path * *Content* + `payload` + -image::azure-data-lake-update-path-using-update-append.png[General configuration properties with example values for the Update Path operation with `append`] +image::azure-data-lake-update-path-using-update-append.png["General configuration properties with example values for the Update Path operation with append"] + . Save the project. @@ -270,7 +270,7 @@ Your target path * *Content Length* + `0` + -image::azure-data-lake-update-path-using-update-flush.png[General configuration properties with example values for the Update Path operation with `flush`] +image::azure-data-lake-update-path-using-update-flush.png["General configuration properties with example values for the Update Path operation with flush"] + . Save the project. . Open a browser and send a GET request to `localhost:8081/update-path` to test the app. @@ -318,11 +318,11 @@ This section shows you how to implement simple error handling on a flow that use . In any flow, from the *Mule Palette* view, select *Core > On Error Continue* and drag it to the *Error Handling* section of the flow: + -image::azure-data-lake-error-handling-setting-up-1.png[Drag the On Error Continue component to the flow] +image::azure-data-lake-error-handling-setting-up-1.png["Drag the On Error Continue component to the flow"] + . From the *Mule Palette* view, drag *Core > Transform Message* to the flow: + -image::azure-data-lake-error-handling-setting-up-2.png[Drag the Transform Message component to the flow] +image::azure-data-lake-error-handling-setting-up-2.png["Drag the Transform Message component to the flow"] + . In the *Mule Palette* view, search for `logger` and drag the *Logger* component to the right of *Transform Message*. . In the *General* properties tab for the *Logger* component, enter the following values: @@ -336,12 +336,12 @@ image::azure-data-lake-error-handling-setting-up-2.png[Drag the Transform Messag To access the error code that comes from the Azure service, the expression is: + `error.muleMessage.typedValue.error.code` + -image::azure-data-lake-error-handling-setting-up-3.png[Expression to access the error code that comes from the Azure service] +image::azure-data-lake-error-handling-setting-up-3.png["Expression to access the error code that comes from the Azure service"] + To access the error message, the expression is: + `error.muleMessage.typedValue.error.message` -image::azure-data-lake-error-handling-setting-up-4.png[Expression to access the error message] +image::azure-data-lake-error-handling-setting-up-4.png["Expression to access the error message"] == See Also diff --git a/azure-data-lake/1.0/modules/ROOT/pages/azure-data-lake-connector-studio.adoc b/azure-data-lake/1.0/modules/ROOT/pages/azure-data-lake-connector-studio.adoc index 69a1b37995..436eb26564 100644 --- a/azure-data-lake/1.0/modules/ROOT/pages/azure-data-lake-connector-studio.adoc +++ b/azure-data-lake/1.0/modules/ROOT/pages/azure-data-lake-connector-studio.adoc @@ -113,7 +113,7 @@ Scopes to request during the OAuth dance The following image shows an example of the Azure Active Directory OAuth 2.0 connection configuration values: .OAuth 2.0 Connection Provider authentication fields -image::azure-data-lake-oauth2-connection-provider.png[Example values for the OAuth 2.0 Connection Provider connection type] +image::azure-data-lake-oauth2-connection-provider.png["Example values for the OAuth 2.0 Connection Provider connection type"] === Shared Access Signature Authentication @@ -129,7 +129,7 @@ SAS token for the Azure storage account The following image shows an example of the *Shared Access Signature* connection configuration values: .Shared Access Signature authentication fields -image::azure-data-lake-shared-access-config.png[Example values for the Shared Access Signature connection type] +image::azure-data-lake-shared-access-config.png["Example values for the Shared Access Signature connection type"] === Shared Key Connection Authentication diff --git a/azure-event-hubs/1.1/modules/ROOT/pages/azure-event-hubs-connector-examples.adoc b/azure-event-hubs/1.1/modules/ROOT/pages/azure-event-hubs-connector-examples.adoc index d4d9bb7aed..557dd04711 100644 --- a/azure-event-hubs/1.1/modules/ROOT/pages/azure-event-hubs-connector-examples.adoc +++ b/azure-event-hubs/1.1/modules/ROOT/pages/azure-event-hubs-connector-examples.adoc @@ -62,7 +62,7 @@ Reads events from an event hub in a namespace. * *Logger* + Logs that the event was listened. -image::aeh-example-1.png[Studio flow for the Send a Single Event flow] +image::aeh-example-1.png["Studio flow for the Send a Single Event flow"] === XML for This Example @@ -153,7 +153,7 @@ Reads events from a specific event hub partition in a namespace * *Logger* + Logs twice the events listened. -image::aeh-example-2.png[Studio flow for the Publish a batch of Events flow] +image::aeh-example-2.png["Studio flow for the Publish a batch of Events flow"] === XML for This Example @@ -225,7 +225,7 @@ http://www.mulesoft.org/schema/mule/scripting http://www.mulesoft.org/schema/mul == Send Single Event with Content Type This Mule flow sends a single event to an Event Hub with content type as JSON. -image::aeh-example-3.png[Send Single Event with Content Type configuration in Studio] +image::aeh-example-3.png["Send Single Event with Content Type configuration in Studio"] === XML for This Example diff --git a/azure-event-hubs/1.1/modules/ROOT/pages/azure-event-hubs-connector-studio-configuration.adoc b/azure-event-hubs/1.1/modules/ROOT/pages/azure-event-hubs-connector-studio-configuration.adoc index 490e1a1f2c..74616f1e1a 100644 --- a/azure-event-hubs/1.1/modules/ROOT/pages/azure-event-hubs-connector-studio-configuration.adoc +++ b/azure-event-hubs/1.1/modules/ROOT/pages/azure-event-hubs-connector-studio-configuration.adoc @@ -178,7 +178,7 @@ The following table shows the fields for configuring *Token credentials properti The following example shows the configuration for *Sas Authentication* with *Azure blob storage sas authentication* in Studio: -image::aeh-sas.png[Configuring azure blob storage SAS authentication] +image::aeh-sas.png["Configuring azure blob storage SAS authentication"] In the *Configuration XML* editor, the `azure-eventhubs:sas-authentication-connection` and `azure-eventhubs:azure-blob-storage-sas-authentication` configurations look like this: @@ -195,7 +195,7 @@ In the *Configuration XML* editor, the `azure-eventhubs:sas-authentication-conne The following example shows the configuration for *Active directory authentication* with *Azure blob storage sas authentication* and *Token credentials properties* in Studio: -image::aeh-active-directory-client.png[Configuring active directory authentication with client credentials] +image::aeh-active-directory-client.png["Configuring active directory authentication with client credentials"] In the *Configuration XML* editor, the `azure-eventhubs:active-directory-authentication-connection`, `azure-eventhubs:azure-blob-storage-sas-authentication`, and `azure-eventhubs:token-credential-properties` configurations look like this: @@ -213,7 +213,7 @@ In the *Configuration XML* editor, the `azure-eventhubs:active-directory-authent The following example shows the configuration for *Active directory authentication* with *Azure blob storage active directory authentication* and *Use azure default credentials* in Studio: -image::aeh-active-directory-default-client.png[Configuring active directory authentication with default client credentials] +image::aeh-active-directory-default-client.png["Configuring active directory authentication with default client credentials"] In the *Configuration XML* editor, the `azure-eventhubs:active-directory-authentication-connection`, `azure-eventhubs:azure-blob-storage-sas-authentication` , and `useAzureDefaultCredential` configurations look like this: diff --git a/azure-key-vault-properties-provider/2.1/modules/ROOT/pages/index.adoc b/azure-key-vault-properties-provider/2.1/modules/ROOT/pages/index.adoc index e08cf34aa5..bfe4cd93ba 100644 --- a/azure-key-vault-properties-provider/2.1/modules/ROOT/pages/index.adoc +++ b/azure-key-vault-properties-provider/2.1/modules/ROOT/pages/index.adoc @@ -29,7 +29,7 @@ With Azure Key Vault Properties Provider, secrets can be retrieved in the pre-ru Connectors can use the Azure Key Vault Properties Provider in their configurations, for example: -image::properties-provider-example.png[Example of MongoDB Connector using Azure Key Vault Properties Provider in configuration window] +image::properties-provider-example.png["Example of MongoDB Connector using Azure Key Vault Properties Provider in configuration window"] == See Also diff --git a/azure-key-vault/1.1/modules/ROOT/pages/azure-key-vault-connector-examples.adoc b/azure-key-vault/1.1/modules/ROOT/pages/azure-key-vault-connector-examples.adoc index 5aa8d8d65e..61193372bc 100644 --- a/azure-key-vault/1.1/modules/ROOT/pages/azure-key-vault-connector-examples.adoc +++ b/azure-key-vault/1.1/modules/ROOT/pages/azure-key-vault-connector-examples.adoc @@ -97,7 +97,7 @@ Transforms the message using the algorithm * *Logger* + Decrypts the log -image::encrypt-decrypt.png[Studio flow for the Encrypt and Decrypt a Secret flow] +image::encrypt-decrypt.png["Studio flow for the Encrypt and Decrypt a Secret flow"] === XML for This Example @@ -170,7 +170,7 @@ Writes the content into the file * *Transform Message* + Outputs the success message that says the certificate is saved into a file -image::retrieve-save.png[Studio flow for the Retrieve and Save a Certificate flow] +image::retrieve-save.png["Studio flow for the Retrieve and Save a Certificate flow"] === XML for This Example @@ -236,7 +236,7 @@ Transforms metadata to JSON * *Logger* + Logs the message -image::retrieve-secret.png[Studio flow for the Retrieve a Secret and Its Metadata flow] +image::retrieve-secret.png["Studio flow for the Retrieve a Secret and Its Metadata flow"] === XML for This Example diff --git a/azure-service-bus-management/1.0/modules/ROOT/pages/azure-service-bus-management-connector-examples.adoc b/azure-service-bus-management/1.0/modules/ROOT/pages/azure-service-bus-management-connector-examples.adoc index a9e8728e7b..aba9fd91a9 100644 --- a/azure-service-bus-management/1.0/modules/ROOT/pages/azure-service-bus-management-connector-examples.adoc +++ b/azure-service-bus-management/1.0/modules/ROOT/pages/azure-service-bus-management-connector-examples.adoc @@ -23,7 +23,7 @@ Ensure you have the following: This example shows how to get information about a particular queue in the Azure service bus instance. The following screenshot shows the Anypoint Studio app flow for this example: .App flow for the Get Queue Information example -image::azure-sb-mgmt-flow-get-queue.png[This app flow shows the components used in the get Get Queue Information example] +image::azure-sb-mgmt-flow-get-queue.png["App flow that shows the components used in the get Get Queue Information example"] . Create a new Mule project in Studio. . In the *Mule Palette* view, search for `http` and select the *Listener* operation: @@ -136,7 +136,7 @@ ns ns01 http://schemas.microsoft.com/netservices/2010/10/servicebus/connect This example shows how to create a new topic for the Azure Service Bus instance. The following screenshot shows the Anypoint Studio flow for this example: .App flow for the Create Topic example -image::azure-sb-mgmt-flow-create-topic.png[This app flow shows the components used in the Create Topic example] +image::azure-sb-mgmt-flow-create-topic.png["App flow that shows the components used in the Create Topic example"] . Create a new Mule project in Studio. . In the *Mule Palette* view, search for `http` and select the *Listener* operation: . Drag the *Listener* operation onto the canvas. diff --git a/azure-service-bus-management/1.0/modules/ROOT/pages/azure-service-bus-management-connector-studio.adoc b/azure-service-bus-management/1.0/modules/ROOT/pages/azure-service-bus-management-connector-studio.adoc index 08c9401e2b..d54f0e1d4c 100644 --- a/azure-service-bus-management/1.0/modules/ROOT/pages/azure-service-bus-management-connector-studio.adoc +++ b/azure-service-bus-management/1.0/modules/ROOT/pages/azure-service-bus-management-connector-studio.adoc @@ -90,7 +90,7 @@ To configure the global element for Azure Service Bus Management Connector, foll The following screenshot shows an example of configuring shared access signature (SAS) authentication. This screenshot uses property placeholders for field values. For information about using property placeholders, see xref:connectors::introduction/intro-connector-configuration-overview.adoc[Anypoint Connector Configuration]. + .SAS authentication fields -image::azure-sb-mgmt-authentication.png[To configure SAS authentication, complete the fields on the *General* tab.] +image::azure-sb-mgmt-authentication.png["To configure SAS authentication, complete the fields on the General tab"] + . On the *Advanced* tab, optionally specify reconnection information, including a reconnection strategy. . Click *Test Connection* to confirm that Mule can connect with the specified server.