1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768697071727374757677787980818283848586878889909192939495969798991001011021031041051061071081091101111121131141151161171181191201211221231241251261271281291301311321331341351361371381391401411421431441451461471481491501511521531541551561571581591601611621631641651661671681691701711721731741751761771781791801811821831841851861871881891901911921931941951961971981992002012022032042052062072082092102112122132142152162172182192202212222232242252262272282292302312322332342352362372382392402412422432442452462472482492502512522532542552562572582592602612622632642652662672682692702712722732742752762772782792802812822832842852862872882892902912922932942952962972982993003013023033043053063073083093103113123133143153163173183193203213223233243253263273283293303313323333343353363373383393403413423433443453463473483493503513523533543553563573583593603613623633643653663673683693703713723733743753763773783793803813823833843853863873883893903913923933943953963973983994004014024034044054064074084094104114124134144154164174184194204214224234244254264274284294304314324334344354364374384394404414424434444454464474484494504514524534544554564574584594604614624634644654664674684694704714724734744754764774784794804814824834844854864874884894904914924934944954964974984995005015025035045055065075085095105115125135145155165175185195205215225235245255265275285295305315325335345355365375385395405415425435445455465475485495505515525535545555565575585595605615625635645655665675685695705715725735745755765775785795805815825835845855865875885895905915925935945955965975985996006016026036046056066076086096106116126136146156166176186196206216226236246256266276286296306316326336346356366376386396406416426436446456466476486496506516526536546556566576586596606616626636646656666676686696706716726736746756766776786796806816826836846856866876886896906916926936946956966976986997007017027037047057067077087097107117127137147157167177187197207217227237247257267277287297307317327337347357367377387397407417427437447457467477487497507517527537547557567577587597607617627637647657667677687697707717727737747757767777787797807817827837847857867877887897907917927937947957967977987998008018028038048058068078088098108118128138148158168178188198208218228238248258268278288298308318328338348358368378388398408418428438448458468478488498508518528538548558568578588598608618628638648658668678688698708718728738748758768778788798808818828838848858868878888898908918928938948958968978988999009019029039049059069079089099109119129139149159169179189199209219229239249259269279289299309319329339349359369379389399409419429439449459469479489499509519529539549559569579589599609619629639649659669679689699709719729739749759769779789799809819829839849859869879889899909919929939949959969979989991000100110021003100410051006100710081009101010111012101310141015101610171018101910201021102210231024102510261027102810291030103110321033103410351036103710381039104010411042104310441045104610471048104910501051105210531054105510561057105810591060106110621063106410651066106710681069107010711072107310741075107610771078107910801081108210831084108510861087108810891090109110921093109410951096109710981099110011011102110311041105110611071108110911101111111211131114111511161117111811191120112111221123112411251126112711281129113011311132113311341135113611371138113911401141114211431144114511461147114811491150115111521153115411551156115711581159116011611162116311641165116611671168116911701171117211731174117511761177117811791180118111821183118411851186118711881189119011911192119311941195119611971198119912001201120212031204120512061207120812091210121112121213121412151216121712181219122012211222122312241225122612271228122912301231123212331234123512361237123812391240124112421243124412451246124712481249125012511252125312541255125612571258125912601261126212631264126512661267126812691270127112721273127412751276127712781279128012811282128312841285128612871288128912901291129212931294129512961297129812991300130113021303130413051306130713081309131013111312131313141315131613171318131913201321132213231324132513261327132813291330133113321333133413351336133713381339134013411342134313441345134613471348134913501351135213531354135513561357135813591360136113621363136413651366136713681369137013711372137313741375137613771378137913801381138213831384138513861387138813891390139113921393139413951396139713981399140014011402140314041405140614071408140914101411141214131414141514161417141814191420142114221423142414251426142714281429143014311432143314341435143614371438143914401441144214431444144514461447144814491450145114521453145414551456145714581459146014611462146314641465146614671468146914701471147214731474147514761477147814791480148114821483148414851486148714881489149014911492149314941495149614971498149915001501150215031504150515061507150815091510151115121513151415151516151715181519152015211522152315241525152615271528152915301531153215331534153515361537153815391540154115421543154415451546154715481549155015511552155315541555155615571558155915601561156215631564156515661567156815691570157115721573157415751576157715781579158015811582158315841585158615871588158915901591159215931594159515961597159815991600160116021603160416051606160716081609161016111612161316141615161616171618161916201621162216231624162516261627162816291630163116321633163416351636163716381639164016411642164316441645164616471648164916501651165216531654165516561657165816591660166116621663166416651666166716681669167016711672167316741675167616771678167916801681168216831684168516861687168816891690169116921693169416951696169716981699170017011702170317041705170617071708170917101711171217131714 |
- // THIS FILE IS AUTOMATICALLY GENERATED. DO NOT EDIT.
- // Package kinesis provides a client for Amazon Kinesis.
- package kinesis
- import (
- "github.com/aws/aws-sdk-go/aws/awsutil"
- "github.com/aws/aws-sdk-go/aws/request"
- )
- const opAddTagsToStream = "AddTagsToStream"
- // AddTagsToStreamRequest generates a request for the AddTagsToStream operation.
- func (c *Kinesis) AddTagsToStreamRequest(input *AddTagsToStreamInput) (req *request.Request, output *AddTagsToStreamOutput) {
- op := &request.Operation{
- Name: opAddTagsToStream,
- HTTPMethod: "POST",
- HTTPPath: "/",
- }
- if input == nil {
- input = &AddTagsToStreamInput{}
- }
- req = c.newRequest(op, input, output)
- output = &AddTagsToStreamOutput{}
- req.Data = output
- return
- }
- // Adds or updates tags for the specified Amazon Kinesis stream. Each stream
- // can have up to 10 tags.
- //
- // If tags have already been assigned to the stream, AddTagsToStream overwrites
- // any existing tags that correspond to the specified tag keys.
- func (c *Kinesis) AddTagsToStream(input *AddTagsToStreamInput) (*AddTagsToStreamOutput, error) {
- req, out := c.AddTagsToStreamRequest(input)
- err := req.Send()
- return out, err
- }
- const opCreateStream = "CreateStream"
- // CreateStreamRequest generates a request for the CreateStream operation.
- func (c *Kinesis) CreateStreamRequest(input *CreateStreamInput) (req *request.Request, output *CreateStreamOutput) {
- op := &request.Operation{
- Name: opCreateStream,
- HTTPMethod: "POST",
- HTTPPath: "/",
- }
- if input == nil {
- input = &CreateStreamInput{}
- }
- req = c.newRequest(op, input, output)
- output = &CreateStreamOutput{}
- req.Data = output
- return
- }
- // Creates a Amazon Kinesis stream. A stream captures and transports data records
- // that are continuously emitted from different data sources or producers. Scale-out
- // within an Amazon Kinesis stream is explicitly supported by means of shards,
- // which are uniquely identified groups of data records in an Amazon Kinesis
- // stream.
- //
- // You specify and control the number of shards that a stream is composed of.
- // Each open shard can support up to 5 read transactions per second, up to a
- // maximum total of 2 MB of data read per second. Each shard can support up
- // to 1000 records written per second, up to a maximum total of 1 MB data written
- // per second. You can add shards to a stream if the amount of data input increases
- // and you can remove shards if the amount of data input decreases.
- //
- // The stream name identifies the stream. The name is scoped to the AWS account
- // used by the application. It is also scoped by region. That is, two streams
- // in two different accounts can have the same name, and two streams in the
- // same account, but in two different regions, can have the same name.
- //
- // CreateStream is an asynchronous operation. Upon receiving a CreateStream
- // request, Amazon Kinesis immediately returns and sets the stream status to
- // CREATING. After the stream is created, Amazon Kinesis sets the stream status
- // to ACTIVE. You should perform read and write operations only on an ACTIVE
- // stream.
- //
- // You receive a LimitExceededException when making a CreateStream request
- // if you try to do one of the following:
- //
- // Have more than five streams in the CREATING state at any point in time.
- // Create more shards than are authorized for your account. For the default
- // shard limit for an AWS account, see Amazon Kinesis Limits (http://docs.aws.amazon.com/kinesis/latest/dev/service-sizes-and-limits.html).
- // If you need to increase this limit, contact AWS Support (http://docs.aws.amazon.com/general/latest/gr/aws_service_limits.html)
- //
- // You can use DescribeStream to check the stream status, which is returned
- // in StreamStatus.
- //
- // CreateStream has a limit of 5 transactions per second per account.
- func (c *Kinesis) CreateStream(input *CreateStreamInput) (*CreateStreamOutput, error) {
- req, out := c.CreateStreamRequest(input)
- err := req.Send()
- return out, err
- }
- const opDeleteStream = "DeleteStream"
- // DeleteStreamRequest generates a request for the DeleteStream operation.
- func (c *Kinesis) DeleteStreamRequest(input *DeleteStreamInput) (req *request.Request, output *DeleteStreamOutput) {
- op := &request.Operation{
- Name: opDeleteStream,
- HTTPMethod: "POST",
- HTTPPath: "/",
- }
- if input == nil {
- input = &DeleteStreamInput{}
- }
- req = c.newRequest(op, input, output)
- output = &DeleteStreamOutput{}
- req.Data = output
- return
- }
- // Deletes a stream and all its shards and data. You must shut down any applications
- // that are operating on the stream before you delete the stream. If an application
- // attempts to operate on a deleted stream, it will receive the exception ResourceNotFoundException.
- //
- // If the stream is in the ACTIVE state, you can delete it. After a DeleteStream
- // request, the specified stream is in the DELETING state until Amazon Kinesis
- // completes the deletion.
- //
- // Note: Amazon Kinesis might continue to accept data read and write operations,
- // such as PutRecord, PutRecords, and GetRecords, on a stream in the DELETING
- // state until the stream deletion is complete.
- //
- // When you delete a stream, any shards in that stream are also deleted, and
- // any tags are dissociated from the stream.
- //
- // You can use the DescribeStream operation to check the state of the stream,
- // which is returned in StreamStatus.
- //
- // DeleteStream has a limit of 5 transactions per second per account.
- func (c *Kinesis) DeleteStream(input *DeleteStreamInput) (*DeleteStreamOutput, error) {
- req, out := c.DeleteStreamRequest(input)
- err := req.Send()
- return out, err
- }
- const opDescribeStream = "DescribeStream"
- // DescribeStreamRequest generates a request for the DescribeStream operation.
- func (c *Kinesis) DescribeStreamRequest(input *DescribeStreamInput) (req *request.Request, output *DescribeStreamOutput) {
- op := &request.Operation{
- Name: opDescribeStream,
- HTTPMethod: "POST",
- HTTPPath: "/",
- Paginator: &request.Paginator{
- InputTokens: []string{"ExclusiveStartShardId"},
- OutputTokens: []string{"StreamDescription.Shards[-1].ShardId"},
- LimitToken: "Limit",
- TruncationToken: "StreamDescription.HasMoreShards",
- },
- }
- if input == nil {
- input = &DescribeStreamInput{}
- }
- req = c.newRequest(op, input, output)
- output = &DescribeStreamOutput{}
- req.Data = output
- return
- }
- // Describes the specified stream.
- //
- // The information about the stream includes its current status, its Amazon
- // Resource Name (ARN), and an array of shard objects. For each shard object,
- // there is information about the hash key and sequence number ranges that the
- // shard spans, and the IDs of any earlier shards that played in a role in creating
- // the shard. A sequence number is the identifier associated with every record
- // ingested in the Amazon Kinesis stream. The sequence number is assigned when
- // a record is put into the stream.
- //
- // You can limit the number of returned shards using the Limit parameter. The
- // number of shards in a stream may be too large to return from a single call
- // to DescribeStream. You can detect this by using the HasMoreShards flag in
- // the returned output. HasMoreShards is set to true when there is more data
- // available.
- //
- // DescribeStream is a paginated operation. If there are more shards available,
- // you can request them using the shard ID of the last shard returned. Specify
- // this ID in the ExclusiveStartShardId parameter in a subsequent request to
- // DescribeStream.
- //
- // DescribeStream has a limit of 10 transactions per second per account.
- func (c *Kinesis) DescribeStream(input *DescribeStreamInput) (*DescribeStreamOutput, error) {
- req, out := c.DescribeStreamRequest(input)
- err := req.Send()
- return out, err
- }
- func (c *Kinesis) DescribeStreamPages(input *DescribeStreamInput, fn func(p *DescribeStreamOutput, lastPage bool) (shouldContinue bool)) error {
- page, _ := c.DescribeStreamRequest(input)
- return page.EachPage(func(p interface{}, lastPage bool) bool {
- return fn(p.(*DescribeStreamOutput), lastPage)
- })
- }
- const opGetRecords = "GetRecords"
- // GetRecordsRequest generates a request for the GetRecords operation.
- func (c *Kinesis) GetRecordsRequest(input *GetRecordsInput) (req *request.Request, output *GetRecordsOutput) {
- op := &request.Operation{
- Name: opGetRecords,
- HTTPMethod: "POST",
- HTTPPath: "/",
- }
- if input == nil {
- input = &GetRecordsInput{}
- }
- req = c.newRequest(op, input, output)
- output = &GetRecordsOutput{}
- req.Data = output
- return
- }
- // Gets data records from a shard.
- //
- // Specify a shard iterator using the ShardIterator parameter. The shard iterator
- // specifies the position in the shard from which you want to start reading
- // data records sequentially. If there are no records available in the portion
- // of the shard that the iterator points to, GetRecords returns an empty list.
- // Note that it might take multiple calls to get to a portion of the shard that
- // contains records.
- //
- // You can scale by provisioning multiple shards. Your application should have
- // one thread per shard, each reading continuously from its stream. To read
- // from a stream continually, call GetRecords in a loop. Use GetShardIterator
- // to get the shard iterator to specify in the first GetRecords call. GetRecords
- // returns a new shard iterator in NextShardIterator. Specify the shard iterator
- // returned in NextShardIterator in subsequent calls to GetRecords. Note that
- // if the shard has been closed, the shard iterator can't return more data and
- // GetRecords returns null in NextShardIterator. You can terminate the loop
- // when the shard is closed, or when the shard iterator reaches the record with
- // the sequence number or other attribute that marks it as the last record to
- // process.
- //
- // Each data record can be up to 50 KB in size, and each shard can read up
- // to 2 MB per second. You can ensure that your calls don't exceed the maximum
- // supported size or throughput by using the Limit parameter to specify the
- // maximum number of records that GetRecords can return. Consider your average
- // record size when determining this limit. For example, if your average record
- // size is 40 KB, you can limit the data returned to about 1 MB per call by
- // specifying 25 as the limit.
- //
- // The size of the data returned by GetRecords will vary depending on the utilization
- // of the shard. The maximum size of data that GetRecords can return is 10 MB.
- // If a call returns this amount of data, subsequent calls made within the next
- // 5 seconds throw ProvisionedThroughputExceededException. If there is insufficient
- // provisioned throughput on the shard, subsequent calls made within the next
- // 1 second throw ProvisionedThroughputExceededException. Note that GetRecords
- // won't return any data when it throws an exception. For this reason, we recommend
- // that you wait one second between calls to GetRecords; however, it's possible
- // that the application will get exceptions for longer than 1 second.
- //
- // To detect whether the application is falling behind in processing, you can
- // use the MillisBehindLatest response attribute. You can also monitor the amount
- // of data in a stream using the CloudWatch metrics. For more information, see
- // Monitoring Amazon Kinesis with Amazon CloudWatch (http://docs.aws.amazon.com/kinesis/latest/dev/monitoring_with_cloudwatch.html)
- // in the Amazon Kinesis Developer Guide.
- func (c *Kinesis) GetRecords(input *GetRecordsInput) (*GetRecordsOutput, error) {
- req, out := c.GetRecordsRequest(input)
- err := req.Send()
- return out, err
- }
- const opGetShardIterator = "GetShardIterator"
- // GetShardIteratorRequest generates a request for the GetShardIterator operation.
- func (c *Kinesis) GetShardIteratorRequest(input *GetShardIteratorInput) (req *request.Request, output *GetShardIteratorOutput) {
- op := &request.Operation{
- Name: opGetShardIterator,
- HTTPMethod: "POST",
- HTTPPath: "/",
- }
- if input == nil {
- input = &GetShardIteratorInput{}
- }
- req = c.newRequest(op, input, output)
- output = &GetShardIteratorOutput{}
- req.Data = output
- return
- }
- // Gets a shard iterator. A shard iterator expires five minutes after it is
- // returned to the requester.
- //
- // A shard iterator specifies the position in the shard from which to start
- // reading data records sequentially. A shard iterator specifies this position
- // using the sequence number of a data record in a shard. A sequence number
- // is the identifier associated with every record ingested in the Amazon Kinesis
- // stream. The sequence number is assigned when a record is put into the stream.
- //
- // You must specify the shard iterator type. For example, you can set the ShardIteratorType
- // parameter to read exactly from the position denoted by a specific sequence
- // number by using the AT_SEQUENCE_NUMBER shard iterator type, or right after
- // the sequence number by using the AFTER_SEQUENCE_NUMBER shard iterator type,
- // using sequence numbers returned by earlier calls to PutRecord, PutRecords,
- // GetRecords, or DescribeStream. You can specify the shard iterator type TRIM_HORIZON
- // in the request to cause ShardIterator to point to the last untrimmed record
- // in the shard in the system, which is the oldest data record in the shard.
- // Or you can point to just after the most recent record in the shard, by using
- // the shard iterator type LATEST, so that you always read the most recent data
- // in the shard.
- //
- // When you repeatedly read from an Amazon Kinesis stream use a GetShardIterator
- // request to get the first shard iterator for use in your first GetRecords
- // request and then use the shard iterator returned by the GetRecords request
- // in NextShardIterator for subsequent reads. A new shard iterator is returned
- // by every GetRecords request in NextShardIterator, which you use in the ShardIterator
- // parameter of the next GetRecords request.
- //
- // If a GetShardIterator request is made too often, you receive a ProvisionedThroughputExceededException.
- // For more information about throughput limits, see GetRecords.
- //
- // If the shard is closed, the iterator can't return more data, and GetShardIterator
- // returns null for its ShardIterator. A shard can be closed using SplitShard
- // or MergeShards.
- //
- // GetShardIterator has a limit of 5 transactions per second per account per
- // open shard.
- func (c *Kinesis) GetShardIterator(input *GetShardIteratorInput) (*GetShardIteratorOutput, error) {
- req, out := c.GetShardIteratorRequest(input)
- err := req.Send()
- return out, err
- }
- const opListStreams = "ListStreams"
- // ListStreamsRequest generates a request for the ListStreams operation.
- func (c *Kinesis) ListStreamsRequest(input *ListStreamsInput) (req *request.Request, output *ListStreamsOutput) {
- op := &request.Operation{
- Name: opListStreams,
- HTTPMethod: "POST",
- HTTPPath: "/",
- Paginator: &request.Paginator{
- InputTokens: []string{"ExclusiveStartStreamName"},
- OutputTokens: []string{"StreamNames[-1]"},
- LimitToken: "Limit",
- TruncationToken: "HasMoreStreams",
- },
- }
- if input == nil {
- input = &ListStreamsInput{}
- }
- req = c.newRequest(op, input, output)
- output = &ListStreamsOutput{}
- req.Data = output
- return
- }
- // Lists your streams.
- //
- // The number of streams may be too large to return from a single call to
- // ListStreams. You can limit the number of returned streams using the Limit
- // parameter. If you do not specify a value for the Limit parameter, Amazon
- // Kinesis uses the default limit, which is currently 10.
- //
- // You can detect if there are more streams available to list by using the
- // HasMoreStreams flag from the returned output. If there are more streams available,
- // you can request more streams by using the name of the last stream returned
- // by the ListStreams request in the ExclusiveStartStreamName parameter in a
- // subsequent request to ListStreams. The group of stream names returned by
- // the subsequent request is then added to the list. You can continue this process
- // until all the stream names have been collected in the list.
- //
- // ListStreams has a limit of 5 transactions per second per account.
- func (c *Kinesis) ListStreams(input *ListStreamsInput) (*ListStreamsOutput, error) {
- req, out := c.ListStreamsRequest(input)
- err := req.Send()
- return out, err
- }
- func (c *Kinesis) ListStreamsPages(input *ListStreamsInput, fn func(p *ListStreamsOutput, lastPage bool) (shouldContinue bool)) error {
- page, _ := c.ListStreamsRequest(input)
- return page.EachPage(func(p interface{}, lastPage bool) bool {
- return fn(p.(*ListStreamsOutput), lastPage)
- })
- }
- const opListTagsForStream = "ListTagsForStream"
- // ListTagsForStreamRequest generates a request for the ListTagsForStream operation.
- func (c *Kinesis) ListTagsForStreamRequest(input *ListTagsForStreamInput) (req *request.Request, output *ListTagsForStreamOutput) {
- op := &request.Operation{
- Name: opListTagsForStream,
- HTTPMethod: "POST",
- HTTPPath: "/",
- }
- if input == nil {
- input = &ListTagsForStreamInput{}
- }
- req = c.newRequest(op, input, output)
- output = &ListTagsForStreamOutput{}
- req.Data = output
- return
- }
- // Lists the tags for the specified Amazon Kinesis stream.
- func (c *Kinesis) ListTagsForStream(input *ListTagsForStreamInput) (*ListTagsForStreamOutput, error) {
- req, out := c.ListTagsForStreamRequest(input)
- err := req.Send()
- return out, err
- }
- const opMergeShards = "MergeShards"
- // MergeShardsRequest generates a request for the MergeShards operation.
- func (c *Kinesis) MergeShardsRequest(input *MergeShardsInput) (req *request.Request, output *MergeShardsOutput) {
- op := &request.Operation{
- Name: opMergeShards,
- HTTPMethod: "POST",
- HTTPPath: "/",
- }
- if input == nil {
- input = &MergeShardsInput{}
- }
- req = c.newRequest(op, input, output)
- output = &MergeShardsOutput{}
- req.Data = output
- return
- }
- // Merges two adjacent shards in a stream and combines them into a single shard
- // to reduce the stream's capacity to ingest and transport data. Two shards
- // are considered adjacent if the union of the hash key ranges for the two shards
- // form a contiguous set with no gaps. For example, if you have two shards,
- // one with a hash key range of 276...381 and the other with a hash key range
- // of 382...454, then you could merge these two shards into a single shard that
- // would have a hash key range of 276...454. After the merge, the single child
- // shard receives data for all hash key values covered by the two parent shards.
- //
- // MergeShards is called when there is a need to reduce the overall capacity
- // of a stream because of excess capacity that is not being used. You must specify
- // the shard to be merged and the adjacent shard for a stream. For more information
- // about merging shards, see Merge Two Shards (http://docs.aws.amazon.com/kinesis/latest/dev/kinesis-using-sdk-java-resharding-merge.html)
- // in the Amazon Kinesis Developer Guide.
- //
- // If the stream is in the ACTIVE state, you can call MergeShards. If a stream
- // is in the CREATING, UPDATING, or DELETING state, MergeShards returns a ResourceInUseException.
- // If the specified stream does not exist, MergeShards returns a ResourceNotFoundException.
- //
- // You can use DescribeStream to check the state of the stream, which is returned
- // in StreamStatus.
- //
- // MergeShards is an asynchronous operation. Upon receiving a MergeShards request,
- // Amazon Kinesis immediately returns a response and sets the StreamStatus to
- // UPDATING. After the operation is completed, Amazon Kinesis sets the StreamStatus
- // to ACTIVE. Read and write operations continue to work while the stream is
- // in the UPDATING state.
- //
- // You use DescribeStream to determine the shard IDs that are specified in
- // the MergeShards request.
- //
- // If you try to operate on too many streams in parallel using CreateStream,
- // DeleteStream, MergeShards or SplitShard, you will receive a LimitExceededException.
- //
- // MergeShards has limit of 5 transactions per second per account.
- func (c *Kinesis) MergeShards(input *MergeShardsInput) (*MergeShardsOutput, error) {
- req, out := c.MergeShardsRequest(input)
- err := req.Send()
- return out, err
- }
- const opPutRecord = "PutRecord"
- // PutRecordRequest generates a request for the PutRecord operation.
- func (c *Kinesis) PutRecordRequest(input *PutRecordInput) (req *request.Request, output *PutRecordOutput) {
- op := &request.Operation{
- Name: opPutRecord,
- HTTPMethod: "POST",
- HTTPPath: "/",
- }
- if input == nil {
- input = &PutRecordInput{}
- }
- req = c.newRequest(op, input, output)
- output = &PutRecordOutput{}
- req.Data = output
- return
- }
- // Puts (writes) a single data record from a producer into an Amazon Kinesis
- // stream. Call PutRecord to send data from the producer into the Amazon Kinesis
- // stream for real-time ingestion and subsequent processing, one record at a
- // time. Each shard can support up to 1000 records written per second, up to
- // a maximum total of 1 MB data written per second.
- //
- // You must specify the name of the stream that captures, stores, and transports
- // the data; a partition key; and the data blob itself.
- //
- // The data blob can be any type of data; for example, a segment from a log
- // file, geographic/location data, website clickstream data, and so on.
- //
- // The partition key is used by Amazon Kinesis to distribute data across shards.
- // Amazon Kinesis segregates the data records that belong to a data stream into
- // multiple shards, using the partition key associated with each data record
- // to determine which shard a given data record belongs to.
- //
- // Partition keys are Unicode strings, with a maximum length limit of 256 characters
- // for each key. An MD5 hash function is used to map partition keys to 128-bit
- // integer values and to map associated data records to shards using the hash
- // key ranges of the shards. You can override hashing the partition key to determine
- // the shard by explicitly specifying a hash value using the ExplicitHashKey
- // parameter. For more information, see Adding Data to a Stream (http://docs.aws.amazon.com/kinesis/latest/dev/kinesis-using-sdk-java-add-data-to-stream.html)
- // in the Amazon Kinesis Developer Guide.
- //
- // PutRecord returns the shard ID of where the data record was placed and the
- // sequence number that was assigned to the data record.
- //
- // Sequence numbers generally increase over time. To guarantee strictly increasing
- // ordering, use the SequenceNumberForOrdering parameter. For more information,
- // see Adding Data to a Stream (http://docs.aws.amazon.com/kinesis/latest/dev/kinesis-using-sdk-java-add-data-to-stream.html)
- // in the Amazon Kinesis Developer Guide.
- //
- // If a PutRecord request cannot be processed because of insufficient provisioned
- // throughput on the shard involved in the request, PutRecord throws ProvisionedThroughputExceededException.
- //
- // Data records are accessible for only 24 hours from the time that they are
- // added to an Amazon Kinesis stream.
- func (c *Kinesis) PutRecord(input *PutRecordInput) (*PutRecordOutput, error) {
- req, out := c.PutRecordRequest(input)
- err := req.Send()
- return out, err
- }
- const opPutRecords = "PutRecords"
- // PutRecordsRequest generates a request for the PutRecords operation.
- func (c *Kinesis) PutRecordsRequest(input *PutRecordsInput) (req *request.Request, output *PutRecordsOutput) {
- op := &request.Operation{
- Name: opPutRecords,
- HTTPMethod: "POST",
- HTTPPath: "/",
- }
- if input == nil {
- input = &PutRecordsInput{}
- }
- req = c.newRequest(op, input, output)
- output = &PutRecordsOutput{}
- req.Data = output
- return
- }
- // Puts (writes) multiple data records from a producer into an Amazon Kinesis
- // stream in a single call (also referred to as a PutRecords request). Use this
- // operation to send data from a data producer into the Amazon Kinesis stream
- // for real-time ingestion and processing. Each shard can support up to 1000
- // records written per second, up to a maximum total of 1 MB data written per
- // second.
- //
- // You must specify the name of the stream that captures, stores, and transports
- // the data; and an array of request Records, with each record in the array
- // requiring a partition key and data blob.
- //
- // The data blob can be any type of data; for example, a segment from a log
- // file, geographic/location data, website clickstream data, and so on.
- //
- // The partition key is used by Amazon Kinesis as input to a hash function
- // that maps the partition key and associated data to a specific shard. An MD5
- // hash function is used to map partition keys to 128-bit integer values and
- // to map associated data records to shards. As a result of this hashing mechanism,
- // all data records with the same partition key map to the same shard within
- // the stream. For more information, see Adding Data to a Stream (http://docs.aws.amazon.com/kinesis/latest/dev/kinesis-using-sdk-java-add-data-to-stream.html)
- // in the Amazon Kinesis Developer Guide.
- //
- // Each record in the Records array may include an optional parameter, ExplicitHashKey,
- // which overrides the partition key to shard mapping. This parameter allows
- // a data producer to determine explicitly the shard where the record is stored.
- // For more information, see Adding Multiple Records with PutRecords (http://docs.aws.amazon.com/kinesis/latest/dev/kinesis-using-sdk-java-add-data-to-stream.html#kinesis-using-sdk-java-putrecords)
- // in the Amazon Kinesis Developer Guide.
- //
- // The PutRecords response includes an array of response Records. Each record
- // in the response array directly correlates with a record in the request array
- // using natural ordering, from the top to the bottom of the request and response.
- // The response Records array always includes the same number of records as
- // the request array.
- //
- // The response Records array includes both successfully and unsuccessfully
- // processed records. Amazon Kinesis attempts to process all records in each
- // PutRecords request. A single record failure does not stop the processing
- // of subsequent records.
- //
- // A successfully-processed record includes ShardId and SequenceNumber values.
- // The ShardId parameter identifies the shard in the stream where the record
- // is stored. The SequenceNumber parameter is an identifier assigned to the
- // put record, unique to all records in the stream.
- //
- // An unsuccessfully-processed record includes ErrorCode and ErrorMessage values.
- // ErrorCode reflects the type of error and can be one of the following values:
- // ProvisionedThroughputExceededException or InternalFailure. ErrorMessage provides
- // more detailed information about the ProvisionedThroughputExceededException
- // exception including the account ID, stream name, and shard ID of the record
- // that was throttled. For more information about partially successful responses,
- // see Adding Multiple Records with PutRecords (http://docs.aws.amazon.com/kinesis/latest/dev/kinesis-using-sdk-java-add-data-to-stream.html#kinesis-using-sdk-java-putrecords)
- // in the Amazon Kinesis Developer Guide.
- //
- // Data records are accessible for only 24 hours from the time that they are
- // added to an Amazon Kinesis stream.
- func (c *Kinesis) PutRecords(input *PutRecordsInput) (*PutRecordsOutput, error) {
- req, out := c.PutRecordsRequest(input)
- err := req.Send()
- return out, err
- }
- const opRemoveTagsFromStream = "RemoveTagsFromStream"
- // RemoveTagsFromStreamRequest generates a request for the RemoveTagsFromStream operation.
- func (c *Kinesis) RemoveTagsFromStreamRequest(input *RemoveTagsFromStreamInput) (req *request.Request, output *RemoveTagsFromStreamOutput) {
- op := &request.Operation{
- Name: opRemoveTagsFromStream,
- HTTPMethod: "POST",
- HTTPPath: "/",
- }
- if input == nil {
- input = &RemoveTagsFromStreamInput{}
- }
- req = c.newRequest(op, input, output)
- output = &RemoveTagsFromStreamOutput{}
- req.Data = output
- return
- }
- // Deletes tags from the specified Amazon Kinesis stream.
- //
- // If you specify a tag that does not exist, it is ignored.
- func (c *Kinesis) RemoveTagsFromStream(input *RemoveTagsFromStreamInput) (*RemoveTagsFromStreamOutput, error) {
- req, out := c.RemoveTagsFromStreamRequest(input)
- err := req.Send()
- return out, err
- }
- const opSplitShard = "SplitShard"
- // SplitShardRequest generates a request for the SplitShard operation.
- func (c *Kinesis) SplitShardRequest(input *SplitShardInput) (req *request.Request, output *SplitShardOutput) {
- op := &request.Operation{
- Name: opSplitShard,
- HTTPMethod: "POST",
- HTTPPath: "/",
- }
- if input == nil {
- input = &SplitShardInput{}
- }
- req = c.newRequest(op, input, output)
- output = &SplitShardOutput{}
- req.Data = output
- return
- }
- // Splits a shard into two new shards in the stream, to increase the stream's
- // capacity to ingest and transport data. SplitShard is called when there is
- // a need to increase the overall capacity of stream because of an expected
- // increase in the volume of data records being ingested.
- //
- // You can also use SplitShard when a shard appears to be approaching its maximum
- // utilization, for example, when the set of producers sending data into the
- // specific shard are suddenly sending more than previously anticipated. You
- // can also call SplitShard to increase stream capacity, so that more Amazon
- // Kinesis applications can simultaneously read data from the stream for real-time
- // processing.
- //
- // You must specify the shard to be split and the new hash key, which is the
- // position in the shard where the shard gets split in two. In many cases, the
- // new hash key might simply be the average of the beginning and ending hash
- // key, but it can be any hash key value in the range being mapped into the
- // shard. For more information about splitting shards, see Split a Shard (http://docs.aws.amazon.com/kinesis/latest/dev/kinesis-using-sdk-java-resharding-split.html)
- // in the Amazon Kinesis Developer Guide.
- //
- // You can use DescribeStream to determine the shard ID and hash key values
- // for the ShardToSplit and NewStartingHashKey parameters that are specified
- // in the SplitShard request.
- //
- // SplitShard is an asynchronous operation. Upon receiving a SplitShard request,
- // Amazon Kinesis immediately returns a response and sets the stream status
- // to UPDATING. After the operation is completed, Amazon Kinesis sets the stream
- // status to ACTIVE. Read and write operations continue to work while the stream
- // is in the UPDATING state.
- //
- // You can use DescribeStream to check the status of the stream, which is returned
- // in StreamStatus. If the stream is in the ACTIVE state, you can call SplitShard.
- // If a stream is in CREATING or UPDATING or DELETING states, DescribeStream
- // returns a ResourceInUseException.
- //
- // If the specified stream does not exist, DescribeStream returns a ResourceNotFoundException.
- // If you try to create more shards than are authorized for your account, you
- // receive a LimitExceededException.
- //
- // For the default shard limit for an AWS account, see Amazon Kinesis Limits
- // (http://docs.aws.amazon.com/kinesis/latest/dev/service-sizes-and-limits.html).
- // If you need to increase this limit, contact AWS Support (http://docs.aws.amazon.com/general/latest/gr/aws_service_limits.html)
- //
- // If you try to operate on too many streams in parallel using CreateStream,
- // DeleteStream, MergeShards or SplitShard, you receive a LimitExceededException.
- //
- // SplitShard has limit of 5 transactions per second per account.
- func (c *Kinesis) SplitShard(input *SplitShardInput) (*SplitShardOutput, error) {
- req, out := c.SplitShardRequest(input)
- err := req.Send()
- return out, err
- }
- // Represents the input for AddTagsToStream.
- type AddTagsToStreamInput struct {
- // The name of the stream.
- StreamName *string `type:"string" required:"true"`
- // The set of key-value pairs to use to create the tags.
- Tags map[string]*string `type:"map" required:"true"`
- metadataAddTagsToStreamInput `json:"-" xml:"-"`
- }
- type metadataAddTagsToStreamInput struct {
- SDKShapeTraits bool `type:"structure"`
- }
- // String returns the string representation
- func (s AddTagsToStreamInput) String() string {
- return awsutil.Prettify(s)
- }
- // GoString returns the string representation
- func (s AddTagsToStreamInput) GoString() string {
- return s.String()
- }
- type AddTagsToStreamOutput struct {
- metadataAddTagsToStreamOutput `json:"-" xml:"-"`
- }
- type metadataAddTagsToStreamOutput struct {
- SDKShapeTraits bool `type:"structure"`
- }
- // String returns the string representation
- func (s AddTagsToStreamOutput) String() string {
- return awsutil.Prettify(s)
- }
- // GoString returns the string representation
- func (s AddTagsToStreamOutput) GoString() string {
- return s.String()
- }
- // Represents the input for CreateStream.
- type CreateStreamInput struct {
- // The number of shards that the stream will use. The throughput of the stream
- // is a function of the number of shards; more shards are required for greater
- // provisioned throughput.
- //
- // DefaultShardLimit;
- ShardCount *int64 `type:"integer" required:"true"`
- // A name to identify the stream. The stream name is scoped to the AWS account
- // used by the application that creates the stream. It is also scoped by region.
- // That is, two streams in two different AWS accounts can have the same name,
- // and two streams in the same AWS account, but in two different regions, can
- // have the same name.
- StreamName *string `type:"string" required:"true"`
- metadataCreateStreamInput `json:"-" xml:"-"`
- }
- type metadataCreateStreamInput struct {
- SDKShapeTraits bool `type:"structure"`
- }
- // String returns the string representation
- func (s CreateStreamInput) String() string {
- return awsutil.Prettify(s)
- }
- // GoString returns the string representation
- func (s CreateStreamInput) GoString() string {
- return s.String()
- }
- type CreateStreamOutput struct {
- metadataCreateStreamOutput `json:"-" xml:"-"`
- }
- type metadataCreateStreamOutput struct {
- SDKShapeTraits bool `type:"structure"`
- }
- // String returns the string representation
- func (s CreateStreamOutput) String() string {
- return awsutil.Prettify(s)
- }
- // GoString returns the string representation
- func (s CreateStreamOutput) GoString() string {
- return s.String()
- }
- // Represents the input for DeleteStream.
- type DeleteStreamInput struct {
- // The name of the stream to delete.
- StreamName *string `type:"string" required:"true"`
- metadataDeleteStreamInput `json:"-" xml:"-"`
- }
- type metadataDeleteStreamInput struct {
- SDKShapeTraits bool `type:"structure"`
- }
- // String returns the string representation
- func (s DeleteStreamInput) String() string {
- return awsutil.Prettify(s)
- }
- // GoString returns the string representation
- func (s DeleteStreamInput) GoString() string {
- return s.String()
- }
- type DeleteStreamOutput struct {
- metadataDeleteStreamOutput `json:"-" xml:"-"`
- }
- type metadataDeleteStreamOutput struct {
- SDKShapeTraits bool `type:"structure"`
- }
- // String returns the string representation
- func (s DeleteStreamOutput) String() string {
- return awsutil.Prettify(s)
- }
- // GoString returns the string representation
- func (s DeleteStreamOutput) GoString() string {
- return s.String()
- }
- // Represents the input for DescribeStream.
- type DescribeStreamInput struct {
- // The shard ID of the shard to start with.
- ExclusiveStartShardId *string `type:"string"`
- // The maximum number of shards to return.
- Limit *int64 `type:"integer"`
- // The name of the stream to describe.
- StreamName *string `type:"string" required:"true"`
- metadataDescribeStreamInput `json:"-" xml:"-"`
- }
- type metadataDescribeStreamInput struct {
- SDKShapeTraits bool `type:"structure"`
- }
- // String returns the string representation
- func (s DescribeStreamInput) String() string {
- return awsutil.Prettify(s)
- }
- // GoString returns the string representation
- func (s DescribeStreamInput) GoString() string {
- return s.String()
- }
- // Represents the output for DescribeStream.
- type DescribeStreamOutput struct {
- // The current status of the stream, the stream ARN, an array of shard objects
- // that comprise the stream, and states whether there are more shards available.
- StreamDescription *StreamDescription `type:"structure" required:"true"`
- metadataDescribeStreamOutput `json:"-" xml:"-"`
- }
- type metadataDescribeStreamOutput struct {
- SDKShapeTraits bool `type:"structure"`
- }
- // String returns the string representation
- func (s DescribeStreamOutput) String() string {
- return awsutil.Prettify(s)
- }
- // GoString returns the string representation
- func (s DescribeStreamOutput) GoString() string {
- return s.String()
- }
- // Represents the input for GetRecords.
- type GetRecordsInput struct {
- // The maximum number of records to return. Specify a value of up to 10,000.
- // If you specify a value that is greater than 10,000, GetRecords throws InvalidArgumentException.
- Limit *int64 `type:"integer"`
- // The position in the shard from which you want to start sequentially reading
- // data records. A shard iterator specifies this position using the sequence
- // number of a data record in the shard.
- ShardIterator *string `type:"string" required:"true"`
- metadataGetRecordsInput `json:"-" xml:"-"`
- }
- type metadataGetRecordsInput struct {
- SDKShapeTraits bool `type:"structure"`
- }
- // String returns the string representation
- func (s GetRecordsInput) String() string {
- return awsutil.Prettify(s)
- }
- // GoString returns the string representation
- func (s GetRecordsInput) GoString() string {
- return s.String()
- }
- // Represents the output for GetRecords.
- type GetRecordsOutput struct {
- // The number of milliseconds the GetRecords response is from the tip of the
- // stream, indicating how far behind current time the consumer is. A value of
- // zero indicates record processing is caught up, and there are no new records
- // to process at this moment.
- MillisBehindLatest *int64 `type:"long"`
- // The next position in the shard from which to start sequentially reading data
- // records. If set to null, the shard has been closed and the requested iterator
- // will not return any more data.
- NextShardIterator *string `type:"string"`
- // The data records retrieved from the shard.
- Records []*Record `type:"list" required:"true"`
- metadataGetRecordsOutput `json:"-" xml:"-"`
- }
- type metadataGetRecordsOutput struct {
- SDKShapeTraits bool `type:"structure"`
- }
- // String returns the string representation
- func (s GetRecordsOutput) String() string {
- return awsutil.Prettify(s)
- }
- // GoString returns the string representation
- func (s GetRecordsOutput) GoString() string {
- return s.String()
- }
- // Represents the input for GetShardIterator.
- type GetShardIteratorInput struct {
- // The shard ID of the shard to get the iterator for.
- ShardId *string `type:"string" required:"true"`
- // Determines how the shard iterator is used to start reading data records from
- // the shard.
- //
- // The following are the valid shard iterator types:
- //
- // AT_SEQUENCE_NUMBER - Start reading exactly from the position denoted by
- // a specific sequence number. AFTER_SEQUENCE_NUMBER - Start reading right after
- // the position denoted by a specific sequence number. TRIM_HORIZON - Start
- // reading at the last untrimmed record in the shard in the system, which is
- // the oldest data record in the shard. LATEST - Start reading just after the
- // most recent record in the shard, so that you always read the most recent
- // data in the shard.
- ShardIteratorType *string `type:"string" required:"true" enum:"ShardIteratorType"`
- // The sequence number of the data record in the shard from which to start reading
- // from.
- StartingSequenceNumber *string `type:"string"`
- // The name of the stream.
- StreamName *string `type:"string" required:"true"`
- metadataGetShardIteratorInput `json:"-" xml:"-"`
- }
- type metadataGetShardIteratorInput struct {
- SDKShapeTraits bool `type:"structure"`
- }
- // String returns the string representation
- func (s GetShardIteratorInput) String() string {
- return awsutil.Prettify(s)
- }
- // GoString returns the string representation
- func (s GetShardIteratorInput) GoString() string {
- return s.String()
- }
- // Represents the output for GetShardIterator.
- type GetShardIteratorOutput struct {
- // The position in the shard from which to start reading data records sequentially.
- // A shard iterator specifies this position using the sequence number of a data
- // record in a shard.
- ShardIterator *string `type:"string"`
- metadataGetShardIteratorOutput `json:"-" xml:"-"`
- }
- type metadataGetShardIteratorOutput struct {
- SDKShapeTraits bool `type:"structure"`
- }
- // String returns the string representation
- func (s GetShardIteratorOutput) String() string {
- return awsutil.Prettify(s)
- }
- // GoString returns the string representation
- func (s GetShardIteratorOutput) GoString() string {
- return s.String()
- }
- // The range of possible hash key values for the shard, which is a set of ordered
- // contiguous positive integers.
- type HashKeyRange struct {
- // The ending hash key of the hash key range.
- EndingHashKey *string `type:"string" required:"true"`
- // The starting hash key of the hash key range.
- StartingHashKey *string `type:"string" required:"true"`
- metadataHashKeyRange `json:"-" xml:"-"`
- }
- type metadataHashKeyRange struct {
- SDKShapeTraits bool `type:"structure"`
- }
- // String returns the string representation
- func (s HashKeyRange) String() string {
- return awsutil.Prettify(s)
- }
- // GoString returns the string representation
- func (s HashKeyRange) GoString() string {
- return s.String()
- }
- // Represents the input for ListStreams.
- type ListStreamsInput struct {
- // The name of the stream to start the list with.
- ExclusiveStartStreamName *string `type:"string"`
- // The maximum number of streams to list.
- Limit *int64 `type:"integer"`
- metadataListStreamsInput `json:"-" xml:"-"`
- }
- type metadataListStreamsInput struct {
- SDKShapeTraits bool `type:"structure"`
- }
- // String returns the string representation
- func (s ListStreamsInput) String() string {
- return awsutil.Prettify(s)
- }
- // GoString returns the string representation
- func (s ListStreamsInput) GoString() string {
- return s.String()
- }
- // Represents the output for ListStreams.
- type ListStreamsOutput struct {
- // If set to true, there are more streams available to list.
- HasMoreStreams *bool `type:"boolean" required:"true"`
- // The names of the streams that are associated with the AWS account making
- // the ListStreams request.
- StreamNames []*string `type:"list" required:"true"`
- metadataListStreamsOutput `json:"-" xml:"-"`
- }
- type metadataListStreamsOutput struct {
- SDKShapeTraits bool `type:"structure"`
- }
- // String returns the string representation
- func (s ListStreamsOutput) String() string {
- return awsutil.Prettify(s)
- }
- // GoString returns the string representation
- func (s ListStreamsOutput) GoString() string {
- return s.String()
- }
- // Represents the input for ListTagsForStream.
- type ListTagsForStreamInput struct {
- // The key to use as the starting point for the list of tags. If this parameter
- // is set, ListTagsForStream gets all tags that occur after ExclusiveStartTagKey.
- ExclusiveStartTagKey *string `type:"string"`
- // The number of tags to return. If this number is less than the total number
- // of tags associated with the stream, HasMoreTags is set to true. To list additional
- // tags, set ExclusiveStartTagKey to the last key in the response.
- Limit *int64 `type:"integer"`
- // The name of the stream.
- StreamName *string `type:"string" required:"true"`
- metadataListTagsForStreamInput `json:"-" xml:"-"`
- }
- type metadataListTagsForStreamInput struct {
- SDKShapeTraits bool `type:"structure"`
- }
- // String returns the string representation
- func (s ListTagsForStreamInput) String() string {
- return awsutil.Prettify(s)
- }
- // GoString returns the string representation
- func (s ListTagsForStreamInput) GoString() string {
- return s.String()
- }
- // Represents the output for ListTagsForStream.
- type ListTagsForStreamOutput struct {
- // If set to true, more tags are available. To request additional tags, set
- // ExclusiveStartTagKey to the key of the last tag returned.
- HasMoreTags *bool `type:"boolean" required:"true"`
- // A list of tags associated with StreamName, starting with the first tag after
- // ExclusiveStartTagKey and up to the specified Limit.
- Tags []*Tag `type:"list" required:"true"`
- metadataListTagsForStreamOutput `json:"-" xml:"-"`
- }
- type metadataListTagsForStreamOutput struct {
- SDKShapeTraits bool `type:"structure"`
- }
- // String returns the string representation
- func (s ListTagsForStreamOutput) String() string {
- return awsutil.Prettify(s)
- }
- // GoString returns the string representation
- func (s ListTagsForStreamOutput) GoString() string {
- return s.String()
- }
- // Represents the input for MergeShards.
- type MergeShardsInput struct {
- // The shard ID of the adjacent shard for the merge.
- AdjacentShardToMerge *string `type:"string" required:"true"`
- // The shard ID of the shard to combine with the adjacent shard for the merge.
- ShardToMerge *string `type:"string" required:"true"`
- // The name of the stream for the merge.
- StreamName *string `type:"string" required:"true"`
- metadataMergeShardsInput `json:"-" xml:"-"`
- }
- type metadataMergeShardsInput struct {
- SDKShapeTraits bool `type:"structure"`
- }
- // String returns the string representation
- func (s MergeShardsInput) String() string {
- return awsutil.Prettify(s)
- }
- // GoString returns the string representation
- func (s MergeShardsInput) GoString() string {
- return s.String()
- }
- type MergeShardsOutput struct {
- metadataMergeShardsOutput `json:"-" xml:"-"`
- }
- type metadataMergeShardsOutput struct {
- SDKShapeTraits bool `type:"structure"`
- }
- // String returns the string representation
- func (s MergeShardsOutput) String() string {
- return awsutil.Prettify(s)
- }
- // GoString returns the string representation
- func (s MergeShardsOutput) GoString() string {
- return s.String()
- }
- // Represents the input for PutRecord.
- type PutRecordInput struct {
- // The data blob to put into the record, which is base64-encoded when the blob
- // is serialized. The maximum size of the data blob (the payload before base64-encoding)
- // is 50 kilobytes (KB)
- Data []byte `type:"blob" required:"true"`
- // The hash value used to explicitly determine the shard the data record is
- // assigned to by overriding the partition key hash.
- ExplicitHashKey *string `type:"string"`
- // Determines which shard in the stream the data record is assigned to. Partition
- // keys are Unicode strings with a maximum length limit of 256 characters for
- // each key. Amazon Kinesis uses the partition key as input to a hash function
- // that maps the partition key and associated data to a specific shard. Specifically,
- // an MD5 hash function is used to map partition keys to 128-bit integer values
- // and to map associated data records to shards. As a result of this hashing
- // mechanism, all data records with the same partition key will map to the same
- // shard within the stream.
- PartitionKey *string `type:"string" required:"true"`
- // Guarantees strictly increasing sequence numbers, for puts from the same client
- // and to the same partition key. Usage: set the SequenceNumberForOrdering of
- // record n to the sequence number of record n-1 (as returned in the result
- // when putting record n-1). If this parameter is not set, records will be coarsely
- // ordered based on arrival time.
- SequenceNumberForOrdering *string `type:"string"`
- // The name of the stream to put the data record into.
- StreamName *string `type:"string" required:"true"`
- metadataPutRecordInput `json:"-" xml:"-"`
- }
- type metadataPutRecordInput struct {
- SDKShapeTraits bool `type:"structure"`
- }
- // String returns the string representation
- func (s PutRecordInput) String() string {
- return awsutil.Prettify(s)
- }
- // GoString returns the string representation
- func (s PutRecordInput) GoString() string {
- return s.String()
- }
- // Represents the output for PutRecord.
- type PutRecordOutput struct {
- // The sequence number identifier that was assigned to the put data record.
- // The sequence number for the record is unique across all records in the stream.
- // A sequence number is the identifier associated with every record put into
- // the stream.
- SequenceNumber *string `type:"string" required:"true"`
- // The shard ID of the shard where the data record was placed.
- ShardId *string `type:"string" required:"true"`
- metadataPutRecordOutput `json:"-" xml:"-"`
- }
- type metadataPutRecordOutput struct {
- SDKShapeTraits bool `type:"structure"`
- }
- // String returns the string representation
- func (s PutRecordOutput) String() string {
- return awsutil.Prettify(s)
- }
- // GoString returns the string representation
- func (s PutRecordOutput) GoString() string {
- return s.String()
- }
- // A PutRecords request.
- type PutRecordsInput struct {
- // The records associated with the request.
- Records []*PutRecordsRequestEntry `type:"list" required:"true"`
- // The stream name associated with the request.
- StreamName *string `type:"string" required:"true"`
- metadataPutRecordsInput `json:"-" xml:"-"`
- }
- type metadataPutRecordsInput struct {
- SDKShapeTraits bool `type:"structure"`
- }
- // String returns the string representation
- func (s PutRecordsInput) String() string {
- return awsutil.Prettify(s)
- }
- // GoString returns the string representation
- func (s PutRecordsInput) GoString() string {
- return s.String()
- }
- // PutRecords results.
- type PutRecordsOutput struct {
- // The number of unsuccessfully processed records in a PutRecords request.
- FailedRecordCount *int64 `type:"integer"`
- // An array of successfully and unsuccessfully processed record results, correlated
- // with the request by natural ordering. A record that is successfully added
- // to your Amazon Kinesis stream includes SequenceNumber and ShardId in the
- // result. A record that fails to be added to your Amazon Kinesis stream includes
- // ErrorCode and ErrorMessage in the result.
- Records []*PutRecordsResultEntry `type:"list" required:"true"`
- metadataPutRecordsOutput `json:"-" xml:"-"`
- }
- type metadataPutRecordsOutput struct {
- SDKShapeTraits bool `type:"structure"`
- }
- // String returns the string representation
- func (s PutRecordsOutput) String() string {
- return awsutil.Prettify(s)
- }
- // GoString returns the string representation
- func (s PutRecordsOutput) GoString() string {
- return s.String()
- }
- // Represents the output for PutRecords.
- type PutRecordsRequestEntry struct {
- // The data blob to put into the record, which is base64-encoded when the blob
- // is serialized. The maximum size of the data blob (the payload before base64-encoding)
- // is 50 kilobytes (KB)
- Data []byte `type:"blob" required:"true"`
- // The hash value used to determine explicitly the shard that the data record
- // is assigned to by overriding the partition key hash.
- ExplicitHashKey *string `type:"string"`
- // Determines which shard in the stream the data record is assigned to. Partition
- // keys are Unicode strings with a maximum length limit of 256 characters for
- // each key. Amazon Kinesis uses the partition key as input to a hash function
- // that maps the partition key and associated data to a specific shard. Specifically,
- // an MD5 hash function is used to map partition keys to 128-bit integer values
- // and to map associated data records to shards. As a result of this hashing
- // mechanism, all data records with the same partition key map to the same shard
- // within the stream.
- PartitionKey *string `type:"string" required:"true"`
- metadataPutRecordsRequestEntry `json:"-" xml:"-"`
- }
- type metadataPutRecordsRequestEntry struct {
- SDKShapeTraits bool `type:"structure"`
- }
- // String returns the string representation
- func (s PutRecordsRequestEntry) String() string {
- return awsutil.Prettify(s)
- }
- // GoString returns the string representation
- func (s PutRecordsRequestEntry) GoString() string {
- return s.String()
- }
- // Represents the result of an individual record from a PutRecords request.
- // A record that is successfully added to your Amazon Kinesis stream includes
- // SequenceNumber and ShardId in the result. A record that fails to be added
- // to your Amazon Kinesis stream includes ErrorCode and ErrorMessage in the
- // result.
- type PutRecordsResultEntry struct {
- // The error code for an individual record result. ErrorCodes can be either
- // ProvisionedThroughputExceededException or InternalFailure.
- ErrorCode *string `type:"string"`
- // The error message for an individual record result. An ErrorCode value of
- // ProvisionedThroughputExceededException has an error message that includes
- // the account ID, stream name, and shard ID. An ErrorCode value of InternalFailure
- // has the error message "Internal Service Failure".
- ErrorMessage *string `type:"string"`
- // The sequence number for an individual record result.
- SequenceNumber *string `type:"string"`
- // The shard ID for an individual record result.
- ShardId *string `type:"string"`
- metadataPutRecordsResultEntry `json:"-" xml:"-"`
- }
- type metadataPutRecordsResultEntry struct {
- SDKShapeTraits bool `type:"structure"`
- }
- // String returns the string representation
- func (s PutRecordsResultEntry) String() string {
- return awsutil.Prettify(s)
- }
- // GoString returns the string representation
- func (s PutRecordsResultEntry) GoString() string {
- return s.String()
- }
- // The unit of data of the Amazon Kinesis stream, which is composed of a sequence
- // number, a partition key, and a data blob.
- type Record struct {
- // The data blob. The data in the blob is both opaque and immutable to the Amazon
- // Kinesis service, which does not inspect, interpret, or change the data in
- // the blob in any way. The maximum size of the data blob (the payload before
- // base64-encoding) is 50 kilobytes (KB)
- Data []byte `type:"blob" required:"true"`
- // Identifies which shard in the stream the data record is assigned to.
- PartitionKey *string `type:"string" required:"true"`
- // The unique identifier for the record in the Amazon Kinesis stream.
- SequenceNumber *string `type:"string" required:"true"`
- metadataRecord `json:"-" xml:"-"`
- }
- type metadataRecord struct {
- SDKShapeTraits bool `type:"structure"`
- }
- // String returns the string representation
- func (s Record) String() string {
- return awsutil.Prettify(s)
- }
- // GoString returns the string representation
- func (s Record) GoString() string {
- return s.String()
- }
- // Represents the input for RemoveTagsFromStream.
- type RemoveTagsFromStreamInput struct {
- // The name of the stream.
- StreamName *string `type:"string" required:"true"`
- // A list of tag keys. Each corresponding tag is removed from the stream.
- TagKeys []*string `type:"list" required:"true"`
- metadataRemoveTagsFromStreamInput `json:"-" xml:"-"`
- }
- type metadataRemoveTagsFromStreamInput struct {
- SDKShapeTraits bool `type:"structure"`
- }
- // String returns the string representation
- func (s RemoveTagsFromStreamInput) String() string {
- return awsutil.Prettify(s)
- }
- // GoString returns the string representation
- func (s RemoveTagsFromStreamInput) GoString() string {
- return s.String()
- }
- type RemoveTagsFromStreamOutput struct {
- metadataRemoveTagsFromStreamOutput `json:"-" xml:"-"`
- }
- type metadataRemoveTagsFromStreamOutput struct {
- SDKShapeTraits bool `type:"structure"`
- }
- // String returns the string representation
- func (s RemoveTagsFromStreamOutput) String() string {
- return awsutil.Prettify(s)
- }
- // GoString returns the string representation
- func (s RemoveTagsFromStreamOutput) GoString() string {
- return s.String()
- }
- // The range of possible sequence numbers for the shard.
- type SequenceNumberRange struct {
- // The ending sequence number for the range. Shards that are in the OPEN state
- // have an ending sequence number of null.
- EndingSequenceNumber *string `type:"string"`
- // The starting sequence number for the range.
- StartingSequenceNumber *string `type:"string" required:"true"`
- metadataSequenceNumberRange `json:"-" xml:"-"`
- }
- type metadataSequenceNumberRange struct {
- SDKShapeTraits bool `type:"structure"`
- }
- // String returns the string representation
- func (s SequenceNumberRange) String() string {
- return awsutil.Prettify(s)
- }
- // GoString returns the string representation
- func (s SequenceNumberRange) GoString() string {
- return s.String()
- }
- // A uniquely identified group of data records in an Amazon Kinesis stream.
- type Shard struct {
- // The shard Id of the shard adjacent to the shard's parent.
- AdjacentParentShardId *string `type:"string"`
- // The range of possible hash key values for the shard, which is a set of ordered
- // contiguous positive integers.
- HashKeyRange *HashKeyRange `type:"structure" required:"true"`
- // The shard Id of the shard's parent.
- ParentShardId *string `type:"string"`
- // The range of possible sequence numbers for the shard.
- SequenceNumberRange *SequenceNumberRange `type:"structure" required:"true"`
- // The unique identifier of the shard within the Amazon Kinesis stream.
- ShardId *string `type:"string" required:"true"`
- metadataShard `json:"-" xml:"-"`
- }
- type metadataShard struct {
- SDKShapeTraits bool `type:"structure"`
- }
- // String returns the string representation
- func (s Shard) String() string {
- return awsutil.Prettify(s)
- }
- // GoString returns the string representation
- func (s Shard) GoString() string {
- return s.String()
- }
- // Represents the input for SplitShard.
- type SplitShardInput struct {
- // A hash key value for the starting hash key of one of the child shards created
- // by the split. The hash key range for a given shard constitutes a set of ordered
- // contiguous positive integers. The value for NewStartingHashKey must be in
- // the range of hash keys being mapped into the shard. The NewStartingHashKey
- // hash key value and all higher hash key values in hash key range are distributed
- // to one of the child shards. All the lower hash key values in the range are
- // distributed to the other child shard.
- NewStartingHashKey *string `type:"string" required:"true"`
- // The shard ID of the shard to split.
- ShardToSplit *string `type:"string" required:"true"`
- // The name of the stream for the shard split.
- StreamName *string `type:"string" required:"true"`
- metadataSplitShardInput `json:"-" xml:"-"`
- }
- type metadataSplitShardInput struct {
- SDKShapeTraits bool `type:"structure"`
- }
- // String returns the string representation
- func (s SplitShardInput) String() string {
- return awsutil.Prettify(s)
- }
- // GoString returns the string representation
- func (s SplitShardInput) GoString() string {
- return s.String()
- }
- type SplitShardOutput struct {
- metadataSplitShardOutput `json:"-" xml:"-"`
- }
- type metadataSplitShardOutput struct {
- SDKShapeTraits bool `type:"structure"`
- }
- // String returns the string representation
- func (s SplitShardOutput) String() string {
- return awsutil.Prettify(s)
- }
- // GoString returns the string representation
- func (s SplitShardOutput) GoString() string {
- return s.String()
- }
- // Represents the output for DescribeStream.
- type StreamDescription struct {
- // If set to true, more shards in the stream are available to describe.
- HasMoreShards *bool `type:"boolean" required:"true"`
- // The shards that comprise the stream.
- Shards []*Shard `type:"list" required:"true"`
- // The Amazon Resource Name (ARN) for the stream being described.
- StreamARN *string `type:"string" required:"true"`
- // The name of the stream being described.
- StreamName *string `type:"string" required:"true"`
- // The current status of the stream being described.
- //
- // The stream status is one of the following states:
- //
- // CREATING - The stream is being created. Amazon Kinesis immediately returns
- // and sets StreamStatus to CREATING. DELETING - The stream is being deleted.
- // The specified stream is in the DELETING state until Amazon Kinesis completes
- // the deletion. ACTIVE - The stream exists and is ready for read and write
- // operations or deletion. You should perform read and write operations only
- // on an ACTIVE stream. UPDATING - Shards in the stream are being merged or
- // split. Read and write operations continue to work while the stream is in
- // the UPDATING state.
- StreamStatus *string `type:"string" required:"true" enum:"StreamStatus"`
- metadataStreamDescription `json:"-" xml:"-"`
- }
- type metadataStreamDescription struct {
- SDKShapeTraits bool `type:"structure"`
- }
- // String returns the string representation
- func (s StreamDescription) String() string {
- return awsutil.Prettify(s)
- }
- // GoString returns the string representation
- func (s StreamDescription) GoString() string {
- return s.String()
- }
- // Metadata assigned to the stream, consisting of a key-value pair.
- type Tag struct {
- // A unique identifier for the tag. Maximum length: 128 characters. Valid characters:
- // Unicode letters, digits, white space, _ . / = + - % @
- Key *string `type:"string" required:"true"`
- // An optional string, typically used to describe or define the tag. Maximum
- // length: 256 characters. Valid characters: Unicode letters, digits, white
- // space, _ . / = + - % @
- Value *string `type:"string"`
- metadataTag `json:"-" xml:"-"`
- }
- type metadataTag struct {
- SDKShapeTraits bool `type:"structure"`
- }
- // String returns the string representation
- func (s Tag) String() string {
- return awsutil.Prettify(s)
- }
- // GoString returns the string representation
- func (s Tag) GoString() string {
- return s.String()
- }
- const (
- // @enum ShardIteratorType
- ShardIteratorTypeAtSequenceNumber = "AT_SEQUENCE_NUMBER"
- // @enum ShardIteratorType
- ShardIteratorTypeAfterSequenceNumber = "AFTER_SEQUENCE_NUMBER"
- // @enum ShardIteratorType
- ShardIteratorTypeTrimHorizon = "TRIM_HORIZON"
- // @enum ShardIteratorType
- ShardIteratorTypeLatest = "LATEST"
- )
- const (
- // @enum StreamStatus
- StreamStatusCreating = "CREATING"
- // @enum StreamStatus
- StreamStatusDeleting = "DELETING"
- // @enum StreamStatus
- StreamStatusActive = "ACTIVE"
- // @enum StreamStatus
- StreamStatusUpdating = "UPDATING"
- )
|