Git Product home page Git Product logo

amazon-kinesis-analytics-streaming-etl's Introduction

Streaming ETL with Apache Flink and Amazon Kinesis Data Analytics

A streaming ETL pipeline based on Apache Flink and Amazon Kinesis Data Analytics (KDA).

Apache Flink is a framework and distributed processing engine for processing data streams. AWS provides a fully managed service for Apache Flink through Amazon Kinesis Data Analytics, enabling you to quickly build and easily run sophisticated streaming applications with low operational overhead.

The architecture takes advantage of several capabilities that can be achieved when running Apache Flink with Kinesis Data Analytics. Specifically, the architecture supports:

  • Private Networks Connectivity: connect to resources in your Amazon Virtual Private Cloud (Amazon VPC), in your data center using a VPN connection, or in a remote region using a VPC peering connection
  • Multiple sources and sinks: read and write data not only from Amazon Kinesis data streams but also from Apache Kafka or Amazon Managed Streaming for Apache Kafka (Amazon MSK) clusters
  • Data partitioning: determine the partitioning of data that is ingested into Amazon Simple Storage Service (Amazon S3) based on information extracted from the event payload
  • Multiple Elasticsearch indices and custom document ids: fan-out from a single input stream to different Elasticsearch indices and explicitly control the document id
  • Exactly-once semantics: avoid duplicates when ingesting and delivering data between Apache Kafka, Amazon S3, and Amazon Elasticsearch Service (Amazon ES).

You can find a further details and a more thorough description and discussion of the architecture on the AWS Big Data Blog.

Explore the pipeline in your own AWS account

If you want to explore and play around with the architecture, launch this AWS CloudFormation template in your AWS account. The template creates a Kinesis data stream and replays a historic set of set of taxi trips into the data stream. The events are then read by a Kinesis Data Analytics application and persisted to Amazon S3 in Apache Parquet format and partitioned by event time.

Launch CloudFormation Stack

To populate the Kinesis data stream, we use a Java application that replays a public dataset of historic taxi trips made in New York City into the data stream. Each event describes a taxi trip made in New York City and includes timestamps for the start and end of a trip, information on the boroughs the trip started and ended in, and various details on the fare of the trip.

The Java application has already been downloaded to an EC2 instance that was provisioned by AWS CloudFormation. You just need to connect to the instance and execute the JAR file to start ingesting events into the stream. Follow the link next to ConnectToInstance in the output section of the CloudFormation template to connect to the EC2 instance and execute the following command:

$ java -jar /tmp/amazon-kinesis-replay-*.jar -noWatermark -objectPrefix artifacts/kinesis-analytics-taxi-consumer/taxi-trips-partitioned.json.lz4/dropoff_year=2018/ -speedup 3600 -streamName «Kinesis stream name»

You can obtain the commands with the correct parameters from the output section of the AWS CloudFormation template that you executed previously.

Configuration options

With the default configuration, the Kinesis Data Analytics application reads events from a Kinesis data stream and writes them in Apache Parquet format and partitioned by type and event time to Amazon S3.

The application supports further sources and sinks, in addition to those that are created by the CloudFormation template. You can create additional resources, eg, an Amazon MSK cluster and an Amazon Elasticsearch Service domain, and configure them as additional sources and sinks by adding them to the application's runtime properties.

Configuring additional sources and sinks

You can choose between a single Kinesis data stream and a Kafka topic as input source. In addition you can add a Kinesis data stream, a Kafka topic, an S3 bucket, and Amazon Elasticsearch Service as one or several sinks to the application.

For services that are integrated with AWS Identity and Access Management (IAM), the application uses the role that is configured for the Kinesis Data Analytics application to sign requests. However, the role itself has only permissions that are required to read from the provisioned Kinesis data stream and the S3 bucket, respectively. So when you want to connect to additional resources, you need to add further permissions to the role.

Amazon Kinesis Data Streams

Amazon Kinesis Data Streams can be used as source and sink. To configure a Kinesis data stream as a source or sink, include these parameters in the application's properties.

  • InputKinesisStream: the name of the Kinesis data stream to read data from
  • InputStreamRegion (optional): the region of the Kinesis data stream, defaults to the region of the KDA application
  • OutputKinesisStream: the name of the Kinesis data stream to write data to
  • OutputStreamRegion (optional): the region of the Kinesis data stream, defaults to the region of the KDA application

The Kinesis data stream can be in the same or in a different region. You also can use an Interface VPC Endpoints (AWS PrivateLink) to privately connect to your Kinesis data stream.

Amazon Managed Streaming for Apache Kafka (Amazon MSK)

Apache Kafka and Amazon MSK clusters can be used as sources and sinks by including the following parameters in the application's properties:

  • InputKafkaBootstrapServers: comma separated list of broker DNS names and port pairs for the initial connection
  • InputKafkaTopic: the name of the topic to read data from
  • OutputKafkaBootstrapServers: comma separated list of broker DNS names and port pairs for the initial connection
  • OutputKafkaTopic: the name of the topic to read data from

To access Kafka or MSK clusters, you need to configure the Kinesis Data Analytics application to connect your VPC. For details, see Enabling VPC support below.

Amazon Simple Storage Service (Amazon S3)

To write output to Amazon S3 include the following parameters in the application's properties:

  • OutputBucket: name of the Amazon S3 bucket to persist data to
  • ParquetConversion optional: whether output should be converted to Apache Parquet format, defaults to true

The output on Amazon S3 will be partitioned by boroughs and by the pickup time of the respective events. You can also use a Gateway VPC Endpoint to privately connect to your S3 buckets.

Amazon Elasticsearch Service

To write output to Elasticsearch include the following parameters in the application's properties:

  • OutputElasticsearchEndpoint: the URL to the Elasticsearch endpoint
  • ElasticsearchRegion (optional): the region of the Elasticsearch endpoint, defaults to the region of the KDA application

The payload of the event determines the index and ids of documents written to Amazon Elasticsearch Service. More precisely, the event type is mapped to an index (and document type) and the trip id is mapped to the document id.

You can also access Amazon ES domains that are deployed to a VPC by enabling VPC support.

Enabling VPC support

To connect the Kinesis Data Analytics application to subnets in your VPC you need to enable VPC support.

Kinesis Data Analytics then creates Elastic Network Interfaces (ENIs) in one or more of the subnets provided in your VPC configuration for the application and can access resources that have network connectivity from the configured subnets. This also includes resources that are not directly contained in these subnets but are reachable over a VPN connection or through VPC peering.

When you are leveraging VPC peering, eg, to connect to an MSK cluster or Amazon ES domain in a peered VPC, make sure to enable DNS Resolution Support for the VPC Peering Connection. Otherwise, the Kinesis Data Analytics application is not able to resolve the DNS names of the endpoints in the peered VPC.

When you enable VPC support and need to connect to endpoints that are available over the public internet at the same time, make sure that a NAT gateway has been configured for the respective subnets.

Gateway and Interface VPC Endpoints

You can also use Gateway and Interface VPC Endpoints to privately connect to supported AWS services.

In case you are using Interface VPC Endpoints, it's recommended to enable private DNS so that the application can reach the service using its default DNS hostname.

License

This library is licensed under the MIT-0 License. See the LICENSE file.

amazon-kinesis-analytics-streaming-etl's People

Contributors

sthm avatar amazon-auto avatar dependabot[bot] avatar

Watchers

James Cloos avatar

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.