{-# LANGUAGE DeriveGeneric #-} {-# LANGUAGE DuplicateRecordFields #-} {-# LANGUAGE NamedFieldPuns #-} {-# LANGUAGE OverloadedStrings #-} {-# LANGUAGE RecordWildCards #-} {-# LANGUAGE StrictData #-} {-# LANGUAGE TypeFamilies #-} {-# LANGUAGE NoImplicitPrelude #-} {-# OPTIONS_GHC -fno-warn-unused-binds #-} {-# OPTIONS_GHC -fno-warn-unused-imports #-} {-# OPTIONS_GHC -fno-warn-unused-matches #-} -- Derived from AWS service descriptions, licensed under Apache 2.0. -- | -- Module : Amazonka.Kinesis.SplitShard -- Copyright : (c) 2013-2023 Brendan Hay -- License : Mozilla Public License, v. 2.0. -- Maintainer : Brendan Hay -- Stability : auto-generated -- Portability : non-portable (GHC extensions) -- -- Splits a shard into two new shards in the Kinesis data stream, to -- increase the stream\'s capacity to ingest and transport data. -- @SplitShard@ is called when there is a need to increase the overall -- capacity of a stream because of an expected increase in the volume of -- data records being ingested. This API is only supported for the data -- streams with the provisioned capacity mode. -- -- When invoking this API, it is recommended you use the @StreamARN@ input -- parameter rather than the @StreamName@ input parameter. -- -- You can also use @SplitShard@ when a shard appears to be approaching its -- maximum utilization; for example, the producers sending data into the -- specific shard are suddenly sending more than previously anticipated. -- You can also call @SplitShard@ to increase stream capacity, so that more -- Kinesis Data Streams applications can simultaneously read data from the -- stream for real-time processing. -- -- You must specify the shard to be split and the new hash key, which is -- the position in the shard where the shard gets split in two. In many -- cases, the new hash key might be the average of the beginning and ending -- hash key, but it can be any hash key value in the range being mapped -- into the shard. For more information, see -- -- in the /Amazon Kinesis Data Streams Developer Guide/. -- -- You can use DescribeStreamSummary and the ListShards APIs to determine -- the shard ID and hash key values for the @ShardToSplit@ and -- @NewStartingHashKey@ parameters that are specified in the @SplitShard@ -- request. -- -- @SplitShard@ is an asynchronous operation. Upon receiving a @SplitShard@ -- request, Kinesis Data Streams immediately returns a response and sets -- the stream status to @UPDATING@. After the operation is completed, -- Kinesis Data Streams sets the stream status to @ACTIVE@. Read and write -- operations continue to work while the stream is in the @UPDATING@ state. -- -- You can use DescribeStreamSummary to check the status of the stream, -- which is returned in @StreamStatus@. If the stream is in the @ACTIVE@ -- state, you can call @SplitShard@. -- -- If the specified stream does not exist, DescribeStreamSummary returns a -- @ResourceNotFoundException@. If you try to create more shards than are -- authorized for your account, you receive a @LimitExceededException@. -- -- For the default shard limit for an Amazon Web Services account, see -- -- in the /Amazon Kinesis Data Streams Developer Guide/. To increase this -- limit, -- . -- -- If you try to operate on too many streams simultaneously using -- CreateStream, DeleteStream, MergeShards, and\/or SplitShard, you receive -- a @LimitExceededException@. -- -- @SplitShard@ has a limit of five transactions per second per account. module Amazonka.Kinesis.SplitShard ( -- * Creating a Request SplitShard (..), newSplitShard, -- * Request Lenses splitShard_streamARN, splitShard_streamName, splitShard_shardToSplit, splitShard_newStartingHashKey, -- * Destructuring the Response SplitShardResponse (..), newSplitShardResponse, ) where import qualified Amazonka.Core as Core import qualified Amazonka.Core.Lens.Internal as Lens import qualified Amazonka.Data as Data import Amazonka.Kinesis.Types import qualified Amazonka.Prelude as Prelude import qualified Amazonka.Request as Request import qualified Amazonka.Response as Response -- | Represents the input for @SplitShard@. -- -- /See:/ 'newSplitShard' smart constructor. data SplitShard = SplitShard' { -- | The ARN of the stream. streamARN :: Prelude.Maybe Prelude.Text, -- | The name of the stream for the shard split. streamName :: Prelude.Maybe Prelude.Text, -- | The shard ID of the shard to split. shardToSplit :: Prelude.Text, -- | A hash key value for the starting hash key of one of the child shards -- created by the split. The hash key range for a given shard constitutes a -- set of ordered contiguous positive integers. The value for -- @NewStartingHashKey@ must be in the range of hash keys being mapped into -- the shard. The @NewStartingHashKey@ hash key value and all higher hash -- key values in hash key range are distributed to one of the child shards. -- All the lower hash key values in the range are distributed to the other -- child shard. newStartingHashKey' :: Prelude.Text } deriving (Prelude.Eq, Prelude.Read, Prelude.Show, Prelude.Generic) -- | -- Create a value of 'SplitShard' with all optional fields omitted. -- -- Use or to modify other optional fields. -- -- The following record fields are available, with the corresponding lenses provided -- for backwards compatibility: -- -- 'streamARN', 'splitShard_streamARN' - The ARN of the stream. -- -- 'streamName', 'splitShard_streamName' - The name of the stream for the shard split. -- -- 'shardToSplit', 'splitShard_shardToSplit' - The shard ID of the shard to split. -- -- 'newStartingHashKey'', 'splitShard_newStartingHashKey' - A hash key value for the starting hash key of one of the child shards -- created by the split. The hash key range for a given shard constitutes a -- set of ordered contiguous positive integers. The value for -- @NewStartingHashKey@ must be in the range of hash keys being mapped into -- the shard. The @NewStartingHashKey@ hash key value and all higher hash -- key values in hash key range are distributed to one of the child shards. -- All the lower hash key values in the range are distributed to the other -- child shard. newSplitShard :: -- | 'shardToSplit' Prelude.Text -> -- | 'newStartingHashKey'' Prelude.Text -> SplitShard newSplitShard pShardToSplit_ pNewStartingHashKey_ = SplitShard' { streamARN = Prelude.Nothing, streamName = Prelude.Nothing, shardToSplit = pShardToSplit_, newStartingHashKey' = pNewStartingHashKey_ } -- | The ARN of the stream. splitShard_streamARN :: Lens.Lens' SplitShard (Prelude.Maybe Prelude.Text) splitShard_streamARN = Lens.lens (\SplitShard' {streamARN} -> streamARN) (\s@SplitShard' {} a -> s {streamARN = a} :: SplitShard) -- | The name of the stream for the shard split. splitShard_streamName :: Lens.Lens' SplitShard (Prelude.Maybe Prelude.Text) splitShard_streamName = Lens.lens (\SplitShard' {streamName} -> streamName) (\s@SplitShard' {} a -> s {streamName = a} :: SplitShard) -- | The shard ID of the shard to split. splitShard_shardToSplit :: Lens.Lens' SplitShard Prelude.Text splitShard_shardToSplit = Lens.lens (\SplitShard' {shardToSplit} -> shardToSplit) (\s@SplitShard' {} a -> s {shardToSplit = a} :: SplitShard) -- | A hash key value for the starting hash key of one of the child shards -- created by the split. The hash key range for a given shard constitutes a -- set of ordered contiguous positive integers. The value for -- @NewStartingHashKey@ must be in the range of hash keys being mapped into -- the shard. The @NewStartingHashKey@ hash key value and all higher hash -- key values in hash key range are distributed to one of the child shards. -- All the lower hash key values in the range are distributed to the other -- child shard. splitShard_newStartingHashKey :: Lens.Lens' SplitShard Prelude.Text splitShard_newStartingHashKey = Lens.lens (\SplitShard' {newStartingHashKey'} -> newStartingHashKey') (\s@SplitShard' {} a -> s {newStartingHashKey' = a} :: SplitShard) instance Core.AWSRequest SplitShard where type AWSResponse SplitShard = SplitShardResponse request overrides = Request.postJSON (overrides defaultService) response = Response.receiveNull SplitShardResponse' instance Prelude.Hashable SplitShard where hashWithSalt _salt SplitShard' {..} = _salt `Prelude.hashWithSalt` streamARN `Prelude.hashWithSalt` streamName `Prelude.hashWithSalt` shardToSplit `Prelude.hashWithSalt` newStartingHashKey' instance Prelude.NFData SplitShard where rnf SplitShard' {..} = Prelude.rnf streamARN `Prelude.seq` Prelude.rnf streamName `Prelude.seq` Prelude.rnf shardToSplit `Prelude.seq` Prelude.rnf newStartingHashKey' instance Data.ToHeaders SplitShard where toHeaders = Prelude.const ( Prelude.mconcat [ "X-Amz-Target" Data.=# ( "Kinesis_20131202.SplitShard" :: Prelude.ByteString ), "Content-Type" Data.=# ( "application/x-amz-json-1.1" :: Prelude.ByteString ) ] ) instance Data.ToJSON SplitShard where toJSON SplitShard' {..} = Data.object ( Prelude.catMaybes [ ("StreamARN" Data..=) Prelude.<$> streamARN, ("StreamName" Data..=) Prelude.<$> streamName, Prelude.Just ("ShardToSplit" Data..= shardToSplit), Prelude.Just ("NewStartingHashKey" Data..= newStartingHashKey') ] ) instance Data.ToPath SplitShard where toPath = Prelude.const "/" instance Data.ToQuery SplitShard where toQuery = Prelude.const Prelude.mempty -- | /See:/ 'newSplitShardResponse' smart constructor. data SplitShardResponse = SplitShardResponse' { } deriving (Prelude.Eq, Prelude.Read, Prelude.Show, Prelude.Generic) -- | -- Create a value of 'SplitShardResponse' with all optional fields omitted. -- -- Use or to modify other optional fields. newSplitShardResponse :: SplitShardResponse newSplitShardResponse = SplitShardResponse' instance Prelude.NFData SplitShardResponse where rnf _ = ()