libZSservicesZSamazonka-kinesisZSamazonka-kinesis
Copyright(c) 2013-2021 Brendan Hay
LicenseMozilla Public License, v. 2.0.
MaintainerBrendan Hay <brendan.g.hay+amazonka@gmail.com>
Stabilityauto-generated
Portabilitynon-portable (GHC extensions)
Safe HaskellNone

Amazonka.Kinesis.PutRecord

Description

Writes a single data record into an Amazon Kinesis data stream. Call PutRecord to send data into the stream for real-time ingestion and subsequent processing, one record at a time. Each shard can support writes up to 1,000 records per second, up to a maximum data write total of 1 MiB per second.

You must specify the name of the stream that captures, stores, and transports the data; a partition key; and the data blob itself.

The data blob can be any type of data; for example, a segment from a log file, geographic/location data, website clickstream data, and so on.

The partition key is used by Kinesis Data Streams to distribute data across shards. Kinesis Data Streams segregates the data records that belong to a stream into multiple shards, using the partition key associated with each data record to determine the shard to which a given data record belongs.

Partition keys are Unicode strings, with a maximum length limit of 256 characters for each key. An MD5 hash function is used to map partition keys to 128-bit integer values and to map associated data records to shards using the hash key ranges of the shards. You can override hashing the partition key to determine the shard by explicitly specifying a hash value using the ExplicitHashKey parameter. For more information, see Adding Data to a Stream in the Amazon Kinesis Data Streams Developer Guide.

PutRecord returns the shard ID of where the data record was placed and the sequence number that was assigned to the data record.

Sequence numbers increase over time and are specific to a shard within a stream, not across all shards within a stream. To guarantee strictly increasing ordering, write serially to a shard and use the SequenceNumberForOrdering parameter. For more information, see Adding Data to a Stream in the Amazon Kinesis Data Streams Developer Guide.

After you write a record to a stream, you cannot modify that record or its order within the stream.

If a PutRecord request cannot be processed because of insufficient provisioned throughput on the shard involved in the request, PutRecord throws ProvisionedThroughputExceededException.

By default, data records are accessible for 24 hours from the time that they are added to a stream. You can use IncreaseStreamRetentionPeriod or DecreaseStreamRetentionPeriod to modify this retention period.

Synopsis

Creating a Request

data PutRecord Source #

Represents the input for PutRecord.

See: newPutRecord smart constructor.

Constructors

PutRecord' 

Fields

  • explicitHashKey :: Maybe Text

    The hash value used to explicitly determine the shard the data record is assigned to by overriding the partition key hash.

  • sequenceNumberForOrdering :: Maybe Text

    Guarantees strictly increasing sequence numbers, for puts from the same client and to the same partition key. Usage: set the SequenceNumberForOrdering of record n to the sequence number of record n-1 (as returned in the result when putting record n-1). If this parameter is not set, records are coarsely ordered based on arrival time.

  • streamName :: Text

    The name of the stream to put the data record into.

  • data' :: Base64

    The data blob to put into the record, which is base64-encoded when the blob is serialized. When the data blob (the payload before base64-encoding) is added to the partition key size, the total size must not exceed the maximum record size (1 MiB).

  • partitionKey :: Text

    Determines which shard in the stream the data record is assigned to. Partition keys are Unicode strings with a maximum length limit of 256 characters for each key. Amazon Kinesis Data Streams uses the partition key as input to a hash function that maps the partition key and associated data to a specific shard. Specifically, an MD5 hash function is used to map partition keys to 128-bit integer values and to map associated data records to shards. As a result of this hashing mechanism, all data records with the same partition key map to the same shard within the stream.

Instances

Instances details
Eq PutRecord Source # 
Instance details

Defined in Amazonka.Kinesis.PutRecord

Read PutRecord Source # 
Instance details

Defined in Amazonka.Kinesis.PutRecord

Show PutRecord Source # 
Instance details

Defined in Amazonka.Kinesis.PutRecord

Generic PutRecord Source # 
Instance details

Defined in Amazonka.Kinesis.PutRecord

Associated Types

type Rep PutRecord :: Type -> Type #

NFData PutRecord Source # 
Instance details

Defined in Amazonka.Kinesis.PutRecord

Methods

rnf :: PutRecord -> () #

Hashable PutRecord Source # 
Instance details

Defined in Amazonka.Kinesis.PutRecord

ToJSON PutRecord Source # 
Instance details

Defined in Amazonka.Kinesis.PutRecord

AWSRequest PutRecord Source # 
Instance details

Defined in Amazonka.Kinesis.PutRecord

Associated Types

type AWSResponse PutRecord #

ToHeaders PutRecord Source # 
Instance details

Defined in Amazonka.Kinesis.PutRecord

Methods

toHeaders :: PutRecord -> [Header] #

ToPath PutRecord Source # 
Instance details

Defined in Amazonka.Kinesis.PutRecord

ToQuery PutRecord Source # 
Instance details

Defined in Amazonka.Kinesis.PutRecord

type Rep PutRecord Source # 
Instance details

Defined in Amazonka.Kinesis.PutRecord

type Rep PutRecord = D1 ('MetaData "PutRecord" "Amazonka.Kinesis.PutRecord" "libZSservicesZSamazonka-kinesisZSamazonka-kinesis" 'False) (C1 ('MetaCons "PutRecord'" 'PrefixI 'True) ((S1 ('MetaSel ('Just "explicitHashKey") 'NoSourceUnpackedness 'NoSourceStrictness 'DecidedStrict) (Rec0 (Maybe Text)) :*: S1 ('MetaSel ('Just "sequenceNumberForOrdering") 'NoSourceUnpackedness 'NoSourceStrictness 'DecidedStrict) (Rec0 (Maybe Text))) :*: (S1 ('MetaSel ('Just "streamName") 'NoSourceUnpackedness 'NoSourceStrictness 'DecidedStrict) (Rec0 Text) :*: (S1 ('MetaSel ('Just "data'") 'NoSourceUnpackedness 'NoSourceStrictness 'DecidedStrict) (Rec0 Base64) :*: S1 ('MetaSel ('Just "partitionKey") 'NoSourceUnpackedness 'NoSourceStrictness 'DecidedStrict) (Rec0 Text)))))
type AWSResponse PutRecord Source # 
Instance details

Defined in Amazonka.Kinesis.PutRecord

newPutRecord Source #

Create a value of PutRecord with all optional fields omitted.

Use generic-lens or optics to modify other optional fields.

The following record fields are available, with the corresponding lenses provided for backwards compatibility:

$sel:explicitHashKey:PutRecord', putRecord_explicitHashKey - The hash value used to explicitly determine the shard the data record is assigned to by overriding the partition key hash.

$sel:sequenceNumberForOrdering:PutRecord', putRecord_sequenceNumberForOrdering - Guarantees strictly increasing sequence numbers, for puts from the same client and to the same partition key. Usage: set the SequenceNumberForOrdering of record n to the sequence number of record n-1 (as returned in the result when putting record n-1). If this parameter is not set, records are coarsely ordered based on arrival time.

$sel:streamName:PutRecord', putRecord_streamName - The name of the stream to put the data record into.

$sel:data':PutRecord', putRecord_data - The data blob to put into the record, which is base64-encoded when the blob is serialized. When the data blob (the payload before base64-encoding) is added to the partition key size, the total size must not exceed the maximum record size (1 MiB).-- -- Note: This Lens automatically encodes and decodes Base64 data. -- The underlying isomorphism will encode to Base64 representation during -- serialisation, and decode from Base64 representation during deserialisation. -- This Lens accepts and returns only raw unencoded data.

$sel:partitionKey:PutRecord', putRecord_partitionKey - Determines which shard in the stream the data record is assigned to. Partition keys are Unicode strings with a maximum length limit of 256 characters for each key. Amazon Kinesis Data Streams uses the partition key as input to a hash function that maps the partition key and associated data to a specific shard. Specifically, an MD5 hash function is used to map partition keys to 128-bit integer values and to map associated data records to shards. As a result of this hashing mechanism, all data records with the same partition key map to the same shard within the stream.

Request Lenses

putRecord_explicitHashKey :: Lens' PutRecord (Maybe Text) Source #

The hash value used to explicitly determine the shard the data record is assigned to by overriding the partition key hash.

putRecord_sequenceNumberForOrdering :: Lens' PutRecord (Maybe Text) Source #

Guarantees strictly increasing sequence numbers, for puts from the same client and to the same partition key. Usage: set the SequenceNumberForOrdering of record n to the sequence number of record n-1 (as returned in the result when putting record n-1). If this parameter is not set, records are coarsely ordered based on arrival time.

putRecord_streamName :: Lens' PutRecord Text Source #

The name of the stream to put the data record into.

putRecord_data :: Lens' PutRecord ByteString Source #

The data blob to put into the record, which is base64-encoded when the blob is serialized. When the data blob (the payload before base64-encoding) is added to the partition key size, the total size must not exceed the maximum record size (1 MiB).-- -- Note: This Lens automatically encodes and decodes Base64 data. -- The underlying isomorphism will encode to Base64 representation during -- serialisation, and decode from Base64 representation during deserialisation. -- This Lens accepts and returns only raw unencoded data.

putRecord_partitionKey :: Lens' PutRecord Text Source #

Determines which shard in the stream the data record is assigned to. Partition keys are Unicode strings with a maximum length limit of 256 characters for each key. Amazon Kinesis Data Streams uses the partition key as input to a hash function that maps the partition key and associated data to a specific shard. Specifically, an MD5 hash function is used to map partition keys to 128-bit integer values and to map associated data records to shards. As a result of this hashing mechanism, all data records with the same partition key map to the same shard within the stream.

Destructuring the Response

data PutRecordResponse Source #

Represents the output for PutRecord.

See: newPutRecordResponse smart constructor.

Constructors

PutRecordResponse' 

Fields

  • encryptionType :: Maybe EncryptionType

    The encryption type to use on the record. This parameter can be one of the following values:

    • NONE: Do not encrypt the records in the stream.
    • KMS: Use server-side encryption on the records in the stream using a customer-managed AWS KMS key.
  • httpStatus :: Int

    The response's http status code.

  • shardId :: Text

    The shard ID of the shard where the data record was placed.

  • sequenceNumber :: Text

    The sequence number identifier that was assigned to the put data record. The sequence number for the record is unique across all records in the stream. A sequence number is the identifier associated with every record put into the stream.

Instances

Instances details
Eq PutRecordResponse Source # 
Instance details

Defined in Amazonka.Kinesis.PutRecord

Read PutRecordResponse Source # 
Instance details

Defined in Amazonka.Kinesis.PutRecord

Show PutRecordResponse Source # 
Instance details

Defined in Amazonka.Kinesis.PutRecord

Generic PutRecordResponse Source # 
Instance details

Defined in Amazonka.Kinesis.PutRecord

Associated Types

type Rep PutRecordResponse :: Type -> Type #

NFData PutRecordResponse Source # 
Instance details

Defined in Amazonka.Kinesis.PutRecord

Methods

rnf :: PutRecordResponse -> () #

type Rep PutRecordResponse Source # 
Instance details

Defined in Amazonka.Kinesis.PutRecord

type Rep PutRecordResponse = D1 ('MetaData "PutRecordResponse" "Amazonka.Kinesis.PutRecord" "libZSservicesZSamazonka-kinesisZSamazonka-kinesis" 'False) (C1 ('MetaCons "PutRecordResponse'" 'PrefixI 'True) ((S1 ('MetaSel ('Just "encryptionType") 'NoSourceUnpackedness 'NoSourceStrictness 'DecidedStrict) (Rec0 (Maybe EncryptionType)) :*: S1 ('MetaSel ('Just "httpStatus") 'NoSourceUnpackedness 'NoSourceStrictness 'DecidedStrict) (Rec0 Int)) :*: (S1 ('MetaSel ('Just "shardId") 'NoSourceUnpackedness 'NoSourceStrictness 'DecidedStrict) (Rec0 Text) :*: S1 ('MetaSel ('Just "sequenceNumber") 'NoSourceUnpackedness 'NoSourceStrictness 'DecidedStrict) (Rec0 Text))))

newPutRecordResponse Source #

Create a value of PutRecordResponse with all optional fields omitted.

Use generic-lens or optics to modify other optional fields.

The following record fields are available, with the corresponding lenses provided for backwards compatibility:

$sel:encryptionType:PutRecordResponse', putRecordResponse_encryptionType - The encryption type to use on the record. This parameter can be one of the following values:

  • NONE: Do not encrypt the records in the stream.
  • KMS: Use server-side encryption on the records in the stream using a customer-managed AWS KMS key.

$sel:httpStatus:PutRecordResponse', putRecordResponse_httpStatus - The response's http status code.

$sel:shardId:PutRecordResponse', putRecordResponse_shardId - The shard ID of the shard where the data record was placed.

$sel:sequenceNumber:PutRecordResponse', putRecordResponse_sequenceNumber - The sequence number identifier that was assigned to the put data record. The sequence number for the record is unique across all records in the stream. A sequence number is the identifier associated with every record put into the stream.

Response Lenses

putRecordResponse_encryptionType :: Lens' PutRecordResponse (Maybe EncryptionType) Source #

The encryption type to use on the record. This parameter can be one of the following values:

  • NONE: Do not encrypt the records in the stream.
  • KMS: Use server-side encryption on the records in the stream using a customer-managed AWS KMS key.

putRecordResponse_httpStatus :: Lens' PutRecordResponse Int Source #

The response's http status code.

putRecordResponse_shardId :: Lens' PutRecordResponse Text Source #

The shard ID of the shard where the data record was placed.

putRecordResponse_sequenceNumber :: Lens' PutRecordResponse Text Source #

The sequence number identifier that was assigned to the put data record. The sequence number for the record is unique across all records in the stream. A sequence number is the identifier associated with every record put into the stream.