Welcome to the Amazon MSK Connect Workshop. Amazon MSK Connect is a feature of Amazon MSK (Managed Streaming for Apache Kafka), which enables you to run fully managed Apache Kafka Connect workloads on AWS. Today you will learn about the Amazon MSK Connect feature and how to easily deploy, monitor, and automatically scale connectors that move data between Apache Kafka clusters and external systems.
The workshop is divided into multiple modules. Each module targets a specific connector.
Some of the labs in the workshop will require you to create resources in your AWS cloud account. This will incur costs that may not be covered in AWS free tier. We recommend that you clean up the resources once you are finished with the workshop to stop incurring any additional charges.
You may refer to the Clean Up section for the instructions to clean up the resources created within the workshop.
The workshop is supported in the regions where the Amazon MSK Connect feature is available. Workshop content has been tested in the following regions:
- US East (N. Virginia) (us-east-1)
- Asia Pacific (Singapore) (ap-southeast-1)
At the time of writing (15 Oct 2021), MSK Connect is available in the following AWS Regions: Asia Pacific (Mumbai), Asia Pacific (Seoul), Asia Pacific (Singapore), Asia Pacific (Sydney), Asia Pacific (Tokyo), Canada (Central), EU (Frankfurt), EU (Ireland), EU (London), EU (Paris), EU (Stockholm), South America (Sao Paulo), US East (N. Virginia), US East (Ohio), US West (N. California), US West (Oregon). For the lastest information, visit the AWS Regional Services List.
- Open the Amazon EC2 console at https://console.aws.amazon.com/ec2/
Choose a region where the Amazon MSK Connect feature is available (e.g. us-east-1), this will be the region that you are using for the workshop.
- Choose Key Pairs in the navigation bar on the left, and click on Create key pair.
- Provide a name for the key pair (e.g. mskconnect-workshop-keypair), and click on Create key pair.
- You should see the message Successfully created key pair, and confirm the download of the generated .pem file to your local machine.
-
Download the CloudFormation template provided in the repo: msk-connect-workshop.yml
-
Open the AWS CloudFormation console at https://console.aws.amazon.com/cloudformation/
-
Click on Create stack and select With new resources (standard).
- Choose Upload a template file option and upload the CloudFormation template from your local machine, then click Next.
- Enter mskconnect-workshop as the stack name, choose the mskconnect-workshop-keypair under KeyName, and then click Next.
-
Leave the default settings on the step Configure stack options and click Next.
-
Scroll down to the bottom of the Review mskconnect-workshop screen and check the box to acknowledge the creation of IAM resources. Click Create stack to trigger the creation of the CloudFormation stack.
- Wait until the stack creation is completed before proceeding to the next steps. It may take a while for the MSK cluster creation (~30 to 45mins).
The CloudFormation stack creates the following resources:
- 2 CloudWatch Log Groups for MSK Cluster and MSK Connect connectors respectively, each with 7 days log retention period
- 1 S3 bucket for storing required Apache Kafka Connect plugin files, and other testing results in some workshop modules
- 1 VPC with 1 Public subnet and 3 Private subnets, and the required networking components such as Route Tables, Internet Gateway, NAT Gateway & Elastic IP, VPC Gateway Endpoint for S3
- 1
m5.large
EC2 instance with Java 1.8.0 and Apache Kafka version 2.2.1 installed, and the required Security Group, IAM Role, EC2 Instance Profile - 1 MSK Cluster with 3
kafka.m5.large
broker nodes configured with Apache Kafka version 2.2.1, and the required Security Group
Note that the encryption options are disabled in order to reduce the chance of errors in the workshop, this is not a best practice and you should not use the provided CloudFormation template in any Production environments
Download the Salesforce Connector from Confluent: https://www.confluent.io/hub/confluentinc/kafka-connect-salesforce
Please note that this connector is a Confluent Commercial Connector and supported by Confluent. The requires purchase of a Confluent Platform subscription, including a license to this Commercial Connector. You can also use this connector for a 30-day trial without an enterprise license key - after 30 days, you need to purchase a subscription.
- Click and download the ZIP file:
- Upload the ZIP file to the S3 bucket created by CloudFormation (e.g. mskconnect-bucket-AccountId-Region-RandomGUID).
- Open the Amazon MSK console at https://console.aws.amazon.com/msk/
In the left pane expand MSK Connect, then choose Custom plugins.
-
Choose Create custom plugin, and choose the ZIP file uploaded in Step 2.
-
Enter salesforce-connector-plugin for the custom plugin name, then choose Create custom plugin.
- Wait for a few minutes for the creation process to complete. You should see the following message in a banner at the top of the browser window when completed.
- Open the Amazon MSK console at https://console.aws.amazon.com/msk/
Choose the MSK Cluster created by CloudFormation (e.g. MSKCluster-RandomGUID), then click on View client information.
- Copy the Bootstrap servers and Apache ZooKeeper connection strings, you will need to use them in the later steps.
- Open the Amazon EC2 console at https://console.aws.amazon.com/ec2/
Find the EC2 instance named KafkaClientInstance, right-click on it and choose Connect.
- Go to the SSH client tab and follow the instructions to ssh into the EC2 instance.
- Run the following command on the EC2 instance, replacing ZookeeperConnectString with the value that you saved when you viewed the cluster's client information.
kafka/kafka_2.12-2.2.1/bin/kafka-topics.sh --create --zookeeper <ZookeeperConnectString> --replication-factor 2 --partitions 1 --topic mskconnect-salesforce-topic
If the command succeeds, you should see the message: Created topic mskconnect-salesforce-topic.
- Go to https://developer.Salesforce.com/signup and register for a trial of the Developer Edition account in Salesforce.
- Create a new Connected App in Salesforce. Go to Setup, search for App Manager in the search box. Click the App Manager under Apps, and then click the New Connected App button to set up a new API client.
- Enable the OAuth settings. For testing purpose, provide full API access:
- Copy the Consumer Key and Consumer Secret, you will need to use them in the later steps.
- Click on the Profile icon and go to the Settings. Choose Reset My Security Token to generate a new security token, it will be sent to your email address.
- Open the Amazon MSK console at https://console.aws.amazon.com/msk/
In the left pane expand MSK Connect, then choose Connectors.
- Click Create connector, and choose the salesforce-connector-plugin, and then click Next.
- For the Connector name, enter mskconnect-salesforce-connector. Choose the MSK cluster type and select the MSK Cluster created by CloudFormation (e.g. MSKCluster-RandomGUID).
- For the Connector configuration, paste the following values and update the corresponding fields with
<--PLACEHOLDER-->
(e.g. salesforce.username=abc@test.com):
connector.class=io.confluent.salesforce.SalesforcePushTopicSourceConnector
salesforce.username=<--YOUR-SALESFORCE-USERNAME-->
confluent.topic.bootstrap.servers=<--YOUR-MSK-CLUSTER-BOOTSTRAP-SERVERS-->
tasks.max=1
salesforce.consumer.key=<--YOUR-SALESFORCE-CONSUMER-KEY-->
salesforce.push.topic.name=LeadsPushTopic
salesforce.password=<--YOUR-SALESFORCE-PASSWORD-->
salesforce.password.token=<--YOUR-SALESFORCE-SECURITY-TOKEN-->
salesforce.initial.start=all
confluent.topic.replication.factor=1
kafka.topic=mskconnect-salesforce-topic
salesforce.consumer.secret=<--YOUR-SALESFORCE-CONSUMER-SECRET-->
salesforce.object=Lead
For details on the configuration properties, refer to the Salesforce Connector documentation.
salesforce.initial.start
Specify the initial starting point for the connector for replaying events. Use all to send a replayId of -2 to Salesforce that replays all events from last 24 hours, or use latest to send a replayId of -1 to Salesforce that plays only new incoming events that arrive after the connector has started.
The default is latest in case there are more enqueued events than might be allowed by API limits.
- For the Access permissions, choose the IAM role created by CloudFormation (e.g. StackName-MSKConnectRole-RandomGUID), and then click Next.
- Keep the default settings on Security, then click Next.
- For Logs, choose Deliver to Amazon CloudWatch Logs and Browse to search for the log group created for MSK Connect by CloudFormation (e.g. StackName-MSKConnectLogGroup-RandomGUID), and then click Next.
- Review and choose Create connector. Wait for a few minutes until the connector is successfully created.
- You may scroll down to the bottom and find the logs of the connector in CloudWatch Logs. This will be very helpful for you to troubleshoot the MSK Connect connectors.
- Login to your Salesforce Developer account and search for Leads in the search box. Click on New to create a new Lead in Salesforce to generate some test data.
- In the previous SSH session of the EC2 Client, run the following command to subscribe to the topic mskconnect-salesforce-topic, replacing BootstrapBrokerString with the value that you saved when you viewed the cluster's client information.
kafka/kafka_2.12-2.2.1/bin/kafka-console-consumer.sh --bootstrap-server <BootstrapBrokerString> --topic mskconnect-salesforce-topic --from-beginning
Download the Amazon S3 Sink Connector from Confluent: https://www.confluent.io/hub/confluentinc/kafka-connect-s3
Please note that this connector is a Community Connector and is provided under the Confluent Community License.
- Click and download the ZIP file:
- Upload the ZIP file to the S3 bucket created by CloudFormation (e.g. mskconnect-bucket-AccountId-Region-RandomGUID).
- Open the Amazon MSK console at https://console.aws.amazon.com/msk/
In the left pane expand MSK Connect, then choose Custom plugins.
-
Choose Create custom plugin, and choose the ZIP file uploaded in Step 2.
-
Enter s3-sink-connector-plugin for the custom plugin name, then choose Create custom plugin.
- Wait for a few minutes for the creation process to complete. You should see the following message in a banner at the top of the browser window when completed.
- Open the Amazon MSK console at https://console.aws.amazon.com/msk/
Choose the MSK Cluster created by CloudFormation (e.g. MSKCluster-RandomGUID), then click on View client information.
- Copy the Bootstrap servers and Apache ZooKeeper connection strings, you will need to use them in the later steps.
- Open the Amazon EC2 console at https://console.aws.amazon.com/ec2/
Find the EC2 instance named KafkaClientInstance, right-click on it and choose Connect.
- Go to the SSH client tab and follow the instructions to ssh into the EC2 instance.
- Run the following command on the EC2 instance, replacing ZookeeperConnectString with the value that you saved when you viewed the cluster's client information.
kafka/kafka_2.12-2.2.1/bin/kafka-topics.sh --create --zookeeper <ZookeeperConnectString> --replication-factor 2 --partitions 1 --topic mskconnect-s3-sink-topic
If the command succeeds, you should see the message: Created topic mskconnect-s3-sink-topic.
- Open the Amazon MSK console at https://console.aws.amazon.com/msk/
In the left pane expand MSK Connect, then choose Connectors.
- Click Create connector, and choose the s3-sink-connector-plugin, and then click Next.
- For the Connector name, enter mskconnect-s3-sink-connector. Choose the MSK cluster type and select the MSK Cluster created by CloudFormation (e.g. MSKCluster-RandomGUID).
- For the Connector configuration, paste the following values and update the corresponding fields with
<--PLACEHOLDER-->
(e.g. s3.region=us-east-1):
connector.class=io.confluent.connect.s3.S3SinkConnector
s3.region=<--YOUR-S3-BUCKET-REGION-->
flush.size=1
schema.compatibility=NONE
tasks.max=1
topics=mskconnect-s3-sink-topic
format.class=io.confluent.connect.s3.format.json.JsonFormat
partitioner.class=io.confluent.connect.storage.partitioner.DefaultPartitioner
value.converter=org.apache.kafka.connect.storage.StringConverter
storage.class=io.confluent.connect.s3.storage.S3Storage
s3.bucket.name=<--YOUR-S3-BUCKET-NAME-->
key.converter=org.apache.kafka.connect.storage.StringConverter
For details on the configuration properties, refer to the Amazon S3 Sink Connector documentation.
flush.size
Number of records written to store before invoking file commits.
We specify
flush.size=1
in the workshop to showcase each console message produced in the next section will be written out as a new file in the S3 bucket.
- For the Access permissions, choose the IAM role created by CloudFormation (e.g. StackName-MSKConnectRole-RandomGUID), and then click Next.
- Keep the default settings on Security, then click Next.
- For Logs, choose Deliver to Amazon CloudWatch Logs and Browse to search for the log group created for MSK Connect by CloudFormation (e.g. StackName-MSKConnectLogGroup-RandomGUID), and then click Next.
- Review and choose Create connector. Wait for a few minutes until the connector is successfully created.
- You may scroll down to the bottom and find the logs of the connector in CloudWatch Logs. This will be very helpful for you to troubleshoot the MSK Connect connectors.
- In the previous SSH session of the EC2 Client, run the following command to start the console Producer tool to publish messages from the console to the topic mskconnect-s3-sink-topic, replacing BootstrapBrokerString with the value that you saved when you viewed the cluster's client information.
kafka/kafka_2.12-2.2.1/bin/kafka-console-producer.sh --broker-list <BootstrapBrokerString> --producer.config kafka/kafka_2.12-2.2.1/bin/client.properties --topic mskconnect-s3-sink-topic
- Type in any message that you would like to send, and then press Enter to send the line to your MSK cluster as a separate message. Feel free to repeat this step a few times to generate some testing messages.
- Open the Amazon S3 console at https://console.aws.amazon.com/s3/
You should see that a new folder topics/ is created,
Click into the topics/ folder and you will find your topic mskconnect-s3-sink-topic/,
Click into the mskconnect-s3-sink-topic/ folder and you will find the topic partitions (e.g. only 1 partition partition=0/),
Click into one of the partitions (e.g. partition=0/) folder and you will find your console messages saved as JSON files,
Click into one of the JSON files (e.g. mskconnect-s3-sink-topic+0+0000000000.json), choose the Query with S3 Select option under Object actions:
Click the Run SQL query button with both Input and Output settings set to JSON format, you should be able to see the message that you sent from the EC2 client instance:
Follow the instructions below to clean up the workshop environment.
- First we will need to clean up the created S3 bucket. Note that this step will delete all the data stored in the S3 bucket. Open the Amazon S3 console at https://console.aws.amazon.com/s3/
Search for mskconnect in the search box, select the target bucket (e.g. mskconnect-bucket-AccountId-Region-RandomGUID) and click the Empty button on the top right.
- On the Empty Bucket screen, type permanently delete to confirm deletion and click Empty.
- You should see a confirmation that the S3 bucket is successfully emptied.
- Next we will delete the MSK Connect connectors that we created previously. Open the Amazon MSK console at https://console.aws.amazon.com/msk/
In the left pane choose Connectors under MSK Connect, select the target Connector to be deleted (e.g. mskconnect-salesforce-connector) and click the Delete button on the top right.
- Confirm deletion by typing in the connector name (e.g. mskconnect-salesforce-connector) and click Delete.
- You should see a confirmation that the connector is successfully deleted. Repeat the steps to delete all the connectors.
- Finally we will delete the CloudFormation stack to clean up all the related resources created in the workshop. Open the AWS CloudFormation console at https://console.aws.amazon.com/cloudformation/
Select the CloudFormation stack that you have created during the Lab Setup section (e.g. mskconnect-workshop), click the Delete button to the top right.
- Confirm deletion by clicking Delete stack.
- You should see that the stack deletion is initiated now. Once the stack is deleted, it will disappear from the stack list. You may refresh the page by clicking the refresh button.
The above steps should clean up all the resources (except the MSK Connect Custom Plugins, see Known Issues) that we have previously provisioned in the MSK Connect Workshop.
- You may find that the MSK Connect custom plugins cannot be cleaned up on the Amazon MSK Console. At the time of writing (15 Oct 2021), the DeleteCustomPlugin API operation is currently missing from the Amazon MSK Connect API Reference.
- Aaron Chong - aaronchong888
This workshop is built referencing the AWS Developer Guides, AWS Blog, and other AWS workshop contents listed below:
- Amazon MSK Developer Guide
- AWS CloudFormation User Guide
- Introducing Amazon MSK Connect – Stream Data to and from Your Apache Kafka Clusters Using Managed Connectors
- Amazon AppFlow Workshop
- Amazon MSK Labs