Sign In Start Free Trial
Account

Add to playlist

Create a Playlist

Modal Close icon
You need to login to use this feature.
  • Scalable Data Streaming with Amazon Kinesis
  • Toc
  • feedback
Scalable Data Streaming with Amazon Kinesis

Scalable Data Streaming with Amazon Kinesis

By : Makota, Brian Maguire, Gagne, Chakrabarti
5 (4)
close
Scalable Data Streaming with Amazon Kinesis

Scalable Data Streaming with Amazon Kinesis

5 (4)
By: Makota, Brian Maguire, Gagne, Chakrabarti

Overview of this book

Amazon Kinesis is a collection of secure, serverless, durable, and highly available purpose-built data streaming services. This data streaming service provides APIs and client SDKs that enable you to produce and consume data at scale. Scalable Data Streaming with Amazon Kinesis begins with a quick overview of the core concepts of data streams, along with the essentials of the AWS Kinesis landscape. You'll then explore the requirements of the use case shown through the book to help you get started and cover the key pain points encountered in the data stream life cycle. As you advance, you'll get to grips with the architectural components of Kinesis, understand how they are configured to build data pipelines, and delve into the applications that connect to them for consumption and processing. You'll also build a Kinesis data pipeline from scratch and learn how to implement and apply practical solutions. Moving on, you'll learn how to configure Kinesis on a cloud platform. Finally, you’ll learn how other AWS services can be integrated into Kinesis. These services include Redshift, Dynamo Database, AWS S3, Elastic Search, and third-party applications such as Splunk. By the end of this AWS book, you’ll be able to build and deploy your own Kinesis data pipelines with Kinesis Data Streams (KDS), Kinesis Data Firehose (KFH), Kinesis Video Streams (KVS), and Kinesis Data Analytics (KDA).
Table of Contents (13 chapters)
close
1
Section 1: Introduction to Data Streaming and Amazon Kinesis
5
Section 2: Deep Dive into Kinesis
10
Section 3: Integrations

Discovering Amazon Kinesis Firehose

The core construct in KDF is a delivery stream. Data is ingested into a delivery stream from a source, and data is then delivered by the delivery stream to a configured destination.

The following diagram illustrates producers sending data to Amazon KDF and the delivery destinations it supports:

Figure 5.1 – Producers sending data to Amazon KDF and the delivery destinations it supports

Let's talk about the delivery stream itself first, and then we will talk about each of the supported destinations in detail, followed by the various mechanisms to send data to KDF.

Understanding KDF delivery streams

The purpose of using a KDF delivery stream is to deliver streaming data to one of the supported destinations. Each delivery stream supports a single destination, so if the requirement is to send the same streaming data to multiple destinations, you need to create multiple delivery streams, each pointing to a...

bookmark search playlist font-size

Change the font size

margin-width

Change margin width

day-mode

Change background colour

Close icon Search
Country selected

Close icon Your notes and bookmarks

Delete Bookmark

Modal Close icon
Are you sure you want to delete it?
Cancel
Yes, Delete