Skip to content

Commit

Permalink
feat(AWSFirehose): update models to latest (#5121)
Browse files Browse the repository at this point in the history
  • Loading branch information
awsmobilesdk authored Dec 28, 2023
1 parent 1ead57c commit 53e7dbf
Show file tree
Hide file tree
Showing 6 changed files with 127 additions and 9 deletions.
35 changes: 35 additions & 0 deletions AWSKinesis/AWSFirehoseModel.h
Original file line number Diff line number Diff line change
Expand Up @@ -26,6 +26,7 @@ typedef NS_ENUM(NSInteger, AWSFirehoseErrorType) {
AWSFirehoseErrorConcurrentModification,
AWSFirehoseErrorInvalidArgument,
AWSFirehoseErrorInvalidKMSResource,
AWSFirehoseErrorInvalidSource,
AWSFirehoseErrorLimitExceeded,
AWSFirehoseErrorResourceInUse,
AWSFirehoseErrorResourceNotFound,
Expand Down Expand Up @@ -312,6 +313,7 @@ typedef NS_ENUM(NSInteger, AWSFirehoseSplunkS3BackupMode) {
@class AWSFirehoseSchemaConfiguration;
@class AWSFirehoseSerializer;
@class AWSFirehoseSourceDescription;
@class AWSFirehoseSplunkBufferingHints;
@class AWSFirehoseSplunkDestinationConfiguration;
@class AWSFirehoseSplunkDestinationDescription;
@class AWSFirehoseSplunkDestinationUpdate;
Expand Down Expand Up @@ -3001,13 +3003,36 @@ typedef NS_ENUM(NSInteger, AWSFirehoseSplunkS3BackupMode) {

@end

/**
<p>The buffering options. If no value is specified, the default values for Splunk are used.</p>
*/
@interface AWSFirehoseSplunkBufferingHints : AWSModel


/**
<p>Buffer incoming data for the specified period of time, in seconds, before delivering it to the destination. The default value is 60 (1 minute).</p>
*/
@property (nonatomic, strong) NSNumber * _Nullable intervalInSeconds;

/**
<p>Buffer incoming data to the specified size, in MBs, before delivering it to the destination. The default value is 5. </p>
*/
@property (nonatomic, strong) NSNumber * _Nullable sizeInMBs;

@end

/**
<p>Describes the configuration of a destination in Splunk.</p>
Required parameters: [HECEndpoint, HECEndpointType, HECToken, S3Configuration]
*/
@interface AWSFirehoseSplunkDestinationConfiguration : AWSModel


/**
<p>The buffering options. If no value is specified, the default values for Splunk are used.</p>
*/
@property (nonatomic, strong) AWSFirehoseSplunkBufferingHints * _Nullable bufferingHints;

/**
<p>The Amazon CloudWatch logging options for your delivery stream.</p>
*/
Expand Down Expand Up @@ -3061,6 +3086,11 @@ typedef NS_ENUM(NSInteger, AWSFirehoseSplunkS3BackupMode) {
@interface AWSFirehoseSplunkDestinationDescription : AWSModel


/**
<p>The buffering options. If no value is specified, the default values for Splunk are used.</p>
*/
@property (nonatomic, strong) AWSFirehoseSplunkBufferingHints * _Nullable bufferingHints;

/**
<p>The Amazon CloudWatch logging options for your delivery stream.</p>
*/
Expand Down Expand Up @@ -3114,6 +3144,11 @@ typedef NS_ENUM(NSInteger, AWSFirehoseSplunkS3BackupMode) {
@interface AWSFirehoseSplunkDestinationUpdate : AWSModel


/**
<p>The buffering options. If no value is specified, the default values for Splunk are used.</p>
*/
@property (nonatomic, strong) AWSFirehoseSplunkBufferingHints * _Nullable bufferingHints;

/**
<p>The Amazon CloudWatch logging options for your delivery stream.</p>
*/
Expand Down
30 changes: 30 additions & 0 deletions AWSKinesis/AWSFirehoseModel.m
Original file line number Diff line number Diff line change
Expand Up @@ -3505,6 +3505,21 @@ + (NSValueTransformer *)MSKSourceDescriptionJSONTransformer {

@end

@implementation AWSFirehoseSplunkBufferingHints

+ (BOOL)supportsSecureCoding {
return YES;
}

+ (NSDictionary *)JSONKeyPathsByPropertyKey {
return @{
@"intervalInSeconds" : @"IntervalInSeconds",
@"sizeInMBs" : @"SizeInMBs",
};
}

@end

@implementation AWSFirehoseSplunkDestinationConfiguration

+ (BOOL)supportsSecureCoding {
Expand All @@ -3513,6 +3528,7 @@ + (BOOL)supportsSecureCoding {

+ (NSDictionary *)JSONKeyPathsByPropertyKey {
return @{
@"bufferingHints" : @"BufferingHints",
@"cloudWatchLoggingOptions" : @"CloudWatchLoggingOptions",
@"HECAcknowledgmentTimeoutInSeconds" : @"HECAcknowledgmentTimeoutInSeconds",
@"HECEndpoint" : @"HECEndpoint",
Expand All @@ -3525,6 +3541,10 @@ + (NSDictionary *)JSONKeyPathsByPropertyKey {
};
}

+ (NSValueTransformer *)bufferingHintsJSONTransformer {
return [NSValueTransformer awsmtl_JSONDictionaryTransformerWithModelClass:[AWSFirehoseSplunkBufferingHints class]];
}

+ (NSValueTransformer *)cloudWatchLoggingOptionsJSONTransformer {
return [NSValueTransformer awsmtl_JSONDictionaryTransformerWithModelClass:[AWSFirehoseCloudWatchLoggingOptions class]];
}
Expand Down Expand Up @@ -3593,6 +3613,7 @@ + (BOOL)supportsSecureCoding {

+ (NSDictionary *)JSONKeyPathsByPropertyKey {
return @{
@"bufferingHints" : @"BufferingHints",
@"cloudWatchLoggingOptions" : @"CloudWatchLoggingOptions",
@"HECAcknowledgmentTimeoutInSeconds" : @"HECAcknowledgmentTimeoutInSeconds",
@"HECEndpoint" : @"HECEndpoint",
Expand All @@ -3605,6 +3626,10 @@ + (NSDictionary *)JSONKeyPathsByPropertyKey {
};
}

+ (NSValueTransformer *)bufferingHintsJSONTransformer {
return [NSValueTransformer awsmtl_JSONDictionaryTransformerWithModelClass:[AWSFirehoseSplunkBufferingHints class]];
}

+ (NSValueTransformer *)cloudWatchLoggingOptionsJSONTransformer {
return [NSValueTransformer awsmtl_JSONDictionaryTransformerWithModelClass:[AWSFirehoseCloudWatchLoggingOptions class]];
}
Expand Down Expand Up @@ -3673,6 +3698,7 @@ + (BOOL)supportsSecureCoding {

+ (NSDictionary *)JSONKeyPathsByPropertyKey {
return @{
@"bufferingHints" : @"BufferingHints",
@"cloudWatchLoggingOptions" : @"CloudWatchLoggingOptions",
@"HECAcknowledgmentTimeoutInSeconds" : @"HECAcknowledgmentTimeoutInSeconds",
@"HECEndpoint" : @"HECEndpoint",
Expand All @@ -3685,6 +3711,10 @@ + (NSDictionary *)JSONKeyPathsByPropertyKey {
};
}

+ (NSValueTransformer *)bufferingHintsJSONTransformer {
return [NSValueTransformer awsmtl_JSONDictionaryTransformerWithModelClass:[AWSFirehoseSplunkBufferingHints class]];
}

+ (NSValueTransformer *)cloudWatchLoggingOptionsJSONTransformer {
return [NSValueTransformer awsmtl_JSONDictionaryTransformerWithModelClass:[AWSFirehoseCloudWatchLoggingOptions class]];
}
Expand Down
57 changes: 52 additions & 5 deletions AWSKinesis/AWSFirehoseResources.m
Original file line number Diff line number Diff line change
Expand Up @@ -151,6 +151,7 @@ - (NSString *)definitionString {
{\"shape\":\"ResourceNotFoundException\"},\
{\"shape\":\"InvalidArgumentException\"},\
{\"shape\":\"InvalidKMSResourceException\"},\
{\"shape\":\"InvalidSourceException\"},\
{\"shape\":\"ServiceUnavailableException\"}\
],\
\"documentation\":\"<p>Writes a single data record into an Amazon Kinesis Data Firehose delivery stream. To write multiple data records into a delivery stream, use <a>PutRecordBatch</a>. Applications using these operations are referred to as producers.</p> <p>By default, each delivery stream can take in up to 2,000 transactions per second, 5,000 records per second, or 5 MB per second. If you use <a>PutRecord</a> and <a>PutRecordBatch</a>, the limits are an aggregate across these two operations for each delivery stream. For more information about limits and how to request an increase, see <a href=\\\"https://docs.aws.amazon.com/firehose/latest/dev/limits.html\\\">Amazon Kinesis Data Firehose Limits</a>. </p> <p>Kinesis Data Firehose accumulates and publishes a particular metric for a customer account in one minute intervals. It is possible that the bursts of incoming bytes/records ingested to a delivery stream last only for a few seconds. Due to this, the actual spikes in the traffic might not be fully visible in the customer's 1 minute CloudWatch metrics.</p> <p>You must specify the name of the delivery stream and the data record when using <a>PutRecord</a>. The data record consists of a data blob that can be up to 1,000 KiB in size, and any kind of data. For example, it can be a segment from a log file, geographic location data, website clickstream data, and so on.</p> <p>Kinesis Data Firehose buffers records before delivering them to the destination. To disambiguate the data blobs at the destination, a common solution is to use delimiters in the data, such as a newline (<code>\\\\n</code>) or some other character unique within the data. This allows the consumer application to parse individual data items when reading the data from the destination.</p> <p>The <code>PutRecord</code> operation returns a <code>RecordId</code>, which is a unique string assigned to each record. Producer applications can use this ID for purposes such as auditability and investigation.</p> <p>If the <code>PutRecord</code> operation throws a <code>ServiceUnavailableException</code>, the API is automatically reinvoked (retried) 3 times. If the exception persists, it is possible that the throughput limits have been exceeded for the delivery stream. </p> <p>Re-invoking the Put API operations (for example, PutRecord and PutRecordBatch) can result in data duplicates. For larger data assets, allow for a longer time out before retrying Put API operations.</p> <p>Data records sent to Kinesis Data Firehose are stored for 24 hours from the time they are added to a delivery stream as it tries to send the records to the destination. If the destination is unreachable for more than 24 hours, the data is no longer available.</p> <important> <p>Don't concatenate two or more base64 strings to form the data fields of your records. Instead, concatenate the raw data, then perform base64 encoding.</p> </important>\"\
Expand All @@ -167,6 +168,7 @@ - (NSString *)definitionString {
{\"shape\":\"ResourceNotFoundException\"},\
{\"shape\":\"InvalidArgumentException\"},\
{\"shape\":\"InvalidKMSResourceException\"},\
{\"shape\":\"InvalidSourceException\"},\
{\"shape\":\"ServiceUnavailableException\"}\
],\
\"documentation\":\"<p>Writes multiple data records into a delivery stream in a single call, which can achieve higher throughput per producer than when writing single records. To write single data records into a delivery stream, use <a>PutRecord</a>. Applications using these operations are referred to as producers.</p> <p>Kinesis Data Firehose accumulates and publishes a particular metric for a customer account in one minute intervals. It is possible that the bursts of incoming bytes/records ingested to a delivery stream last only for a few seconds. Due to this, the actual spikes in the traffic might not be fully visible in the customer's 1 minute CloudWatch metrics.</p> <p>For information about service quota, see <a href=\\\"https://docs.aws.amazon.com/firehose/latest/dev/limits.html\\\">Amazon Kinesis Data Firehose Quota</a>.</p> <p>Each <a>PutRecordBatch</a> request supports up to 500 records. Each record in the request can be as large as 1,000 KB (before base64 encoding), up to a limit of 4 MB for the entire request. These limits cannot be changed.</p> <p>You must specify the name of the delivery stream and the data record when using <a>PutRecord</a>. The data record consists of a data blob that can be up to 1,000 KB in size, and any kind of data. For example, it could be a segment from a log file, geographic location data, website clickstream data, and so on.</p> <p>Kinesis Data Firehose buffers records before delivering them to the destination. To disambiguate the data blobs at the destination, a common solution is to use delimiters in the data, such as a newline (<code>\\\\n</code>) or some other character unique within the data. This allows the consumer application to parse individual data items when reading the data from the destination.</p> <p>The <a>PutRecordBatch</a> response includes a count of failed records, <code>FailedPutCount</code>, and an array of responses, <code>RequestResponses</code>. Even if the <a>PutRecordBatch</a> call succeeds, the value of <code>FailedPutCount</code> may be greater than 0, indicating that there are records for which the operation didn't succeed. Each entry in the <code>RequestResponses</code> array provides additional information about the processed record. It directly correlates with a record in the request array using the same ordering, from the top to the bottom. The response array always includes the same number of records as the request array. <code>RequestResponses</code> includes both successfully and unsuccessfully processed records. Kinesis Data Firehose tries to process all records in each <a>PutRecordBatch</a> request. A single record failure does not stop the processing of subsequent records. </p> <p>A successfully processed record includes a <code>RecordId</code> value, which is unique for the record. An unsuccessfully processed record includes <code>ErrorCode</code> and <code>ErrorMessage</code> values. <code>ErrorCode</code> reflects the type of error, and is one of the following values: <code>ServiceUnavailableException</code> or <code>InternalFailure</code>. <code>ErrorMessage</code> provides more detailed information about the error.</p> <p>If there is an internal server error or a timeout, the write might have completed or it might have failed. If <code>FailedPutCount</code> is greater than 0, retry the request, resending only those records that might have failed processing. This minimizes the possible duplicate records and also reduces the total bytes sent (and corresponding charges). We recommend that you handle any duplicates at the destination.</p> <p>If <a>PutRecordBatch</a> throws <code>ServiceUnavailableException</code>, the API is automatically reinvoked (retried) 3 times. If the exception persists, it is possible that the throughput limits have been exceeded for the delivery stream.</p> <p>Re-invoking the Put API operations (for example, PutRecord and PutRecordBatch) can result in data duplicates. For larger data assets, allow for a longer time out before retrying Put API operations.</p> <p>Data records sent to Kinesis Data Firehose are stored for 24 hours from the time they are added to a delivery stream as it attempts to send the records to the destination. If the destination is unreachable for more than 24 hours, the data is no longer available.</p> <important> <p>Don't concatenate two or more base64 strings to form the data fields of your records. Instead, concatenate the raw data, then perform base64 encoding.</p> </important>\"\
Expand Down Expand Up @@ -277,7 +279,7 @@ - (NSString *)definitionString {
\"AmazonOpenSearchServerlessBufferingIntervalInSeconds\":{\
\"type\":\"integer\",\
\"max\":900,\
\"min\":60\
\"min\":0\
},\
\"AmazonOpenSearchServerlessBufferingSizeInMBs\":{\
\"type\":\"integer\",\
Expand Down Expand Up @@ -437,7 +439,7 @@ - (NSString *)definitionString {
\"AmazonopensearchserviceBufferingIntervalInSeconds\":{\
\"type\":\"integer\",\
\"max\":900,\
\"min\":60\
\"min\":0\
},\
\"AmazonopensearchserviceBufferingSizeInMBs\":{\
\"type\":\"integer\",\
Expand Down Expand Up @@ -1230,7 +1232,7 @@ - (NSString *)definitionString {
\"ElasticsearchBufferingIntervalInSeconds\":{\
\"type\":\"integer\",\
\"max\":900,\
\"min\":60\
\"min\":0\
},\
\"ElasticsearchBufferingSizeInMBs\":{\
\"type\":\"integer\",\
Expand Down Expand Up @@ -1773,7 +1775,7 @@ - (NSString *)definitionString {
\"HttpEndpointBufferingIntervalInSeconds\":{\
\"type\":\"integer\",\
\"max\":900,\
\"min\":60\
\"min\":0\
},\
\"HttpEndpointBufferingSizeInMBs\":{\
\"type\":\"integer\",\
Expand Down Expand Up @@ -2002,7 +2004,7 @@ - (NSString *)definitionString {
\"IntervalInSeconds\":{\
\"type\":\"integer\",\
\"max\":900,\
\"min\":60\
\"min\":0\
},\
\"InvalidArgumentException\":{\
\"type\":\"structure\",\
Expand All @@ -2024,6 +2026,15 @@ - (NSString *)definitionString {
\"documentation\":\"<p>Kinesis Data Firehose throws this exception when an attempt to put records or to start or stop delivery stream encryption fails. This happens when the KMS service throws one of the following exception types: <code>AccessDeniedException</code>, <code>InvalidStateException</code>, <code>DisabledException</code>, or <code>NotFoundException</code>.</p>\",\
\"exception\":true\
},\
\"InvalidSourceException\":{\
\"type\":\"structure\",\
\"members\":{\
\"code\":{\"shape\":\"ErrorCode\"},\
\"message\":{\"shape\":\"ErrorMessage\"}\
},\
\"documentation\":\"<p>Only requests from CloudWatch Logs are supported when CloudWatch Logs decompression is enabled.</p>\",\
\"exception\":true\
},\
\"KMSEncryptionConfig\":{\
\"type\":\"structure\",\
\"required\":[\"AWSKMSKeyARN\"],\
Expand Down Expand Up @@ -3074,6 +3085,30 @@ - (NSString *)definitionString {
},\
\"documentation\":\"<p>Details about a Kinesis data stream used as the source for a Kinesis Data Firehose delivery stream.</p>\"\
},\
\"SplunkBufferingHints\":{\
\"type\":\"structure\",\
\"members\":{\
\"IntervalInSeconds\":{\
\"shape\":\"SplunkBufferingIntervalInSeconds\",\
\"documentation\":\"<p>Buffer incoming data for the specified period of time, in seconds, before delivering it to the destination. The default value is 60 (1 minute).</p>\"\
},\
\"SizeInMBs\":{\
\"shape\":\"SplunkBufferingSizeInMBs\",\
\"documentation\":\"<p>Buffer incoming data to the specified size, in MBs, before delivering it to the destination. The default value is 5. </p>\"\
}\
},\
\"documentation\":\"<p>The buffering options. If no value is specified, the default values for Splunk are used.</p>\"\
},\
\"SplunkBufferingIntervalInSeconds\":{\
\"type\":\"integer\",\
\"max\":60,\
\"min\":0\
},\
\"SplunkBufferingSizeInMBs\":{\
\"type\":\"integer\",\
\"max\":5,\
\"min\":1\
},\
\"SplunkDestinationConfiguration\":{\
\"type\":\"structure\",\
\"required\":[\
Expand Down Expand Up @@ -3118,6 +3153,10 @@ - (NSString *)definitionString {
\"CloudWatchLoggingOptions\":{\
\"shape\":\"CloudWatchLoggingOptions\",\
\"documentation\":\"<p>The Amazon CloudWatch logging options for your delivery stream.</p>\"\
},\
\"BufferingHints\":{\
\"shape\":\"SplunkBufferingHints\",\
\"documentation\":\"<p>The buffering options. If no value is specified, the default values for Splunk are used.</p>\"\
}\
},\
\"documentation\":\"<p>Describes the configuration of a destination in Splunk.</p>\"\
Expand Down Expand Up @@ -3160,6 +3199,10 @@ - (NSString *)definitionString {
\"CloudWatchLoggingOptions\":{\
\"shape\":\"CloudWatchLoggingOptions\",\
\"documentation\":\"<p>The Amazon CloudWatch logging options for your delivery stream.</p>\"\
},\
\"BufferingHints\":{\
\"shape\":\"SplunkBufferingHints\",\
\"documentation\":\"<p>The buffering options. If no value is specified, the default values for Splunk are used.</p>\"\
}\
},\
\"documentation\":\"<p>Describes a destination in Splunk.</p>\"\
Expand Down Expand Up @@ -3202,6 +3245,10 @@ - (NSString *)definitionString {
\"CloudWatchLoggingOptions\":{\
\"shape\":\"CloudWatchLoggingOptions\",\
\"documentation\":\"<p>The Amazon CloudWatch logging options for your delivery stream.</p>\"\
},\
\"BufferingHints\":{\
\"shape\":\"SplunkBufferingHints\",\
\"documentation\":\"<p>The buffering options. If no value is specified, the default values for Splunk are used.</p>\"\
}\
},\
\"documentation\":\"<p>Describes an update for a destination in Splunk.</p>\"\
Expand Down
Loading

0 comments on commit 53e7dbf

Please sign in to comment.