# frozen_string_literal: true
# WARNING ABOUT GENERATED CODE
#
# This file is generated. See the contributing guide for more information:
# https://github.com/aws/aws-sdk-ruby/blob/version-3/CONTRIBUTING.md
#
# WARNING ABOUT GENERATED CODE
require 'seahorse/client/plugins/content_length'
require 'aws-sdk-core/plugins/credentials_configuration'
require 'aws-sdk-core/plugins/logging'
require 'aws-sdk-core/plugins/param_converter'
require 'aws-sdk-core/plugins/param_validator'
require 'aws-sdk-core/plugins/user_agent'
require 'aws-sdk-core/plugins/helpful_socket_errors'
require 'aws-sdk-core/plugins/retry_errors'
require 'aws-sdk-core/plugins/global_configuration'
require 'aws-sdk-core/plugins/regional_endpoint'
require 'aws-sdk-core/plugins/endpoint_discovery'
require 'aws-sdk-core/plugins/endpoint_pattern'
require 'aws-sdk-core/plugins/response_paging'
require 'aws-sdk-core/plugins/stub_responses'
require 'aws-sdk-core/plugins/idempotency_token'
require 'aws-sdk-core/plugins/invocation_id'
require 'aws-sdk-core/plugins/jsonvalue_converter'
require 'aws-sdk-core/plugins/client_metrics_plugin'
require 'aws-sdk-core/plugins/client_metrics_send_plugin'
require 'aws-sdk-core/plugins/transfer_encoding'
require 'aws-sdk-core/plugins/http_checksum'
require 'aws-sdk-core/plugins/checksum_algorithm'
require 'aws-sdk-core/plugins/request_compression'
require 'aws-sdk-core/plugins/defaults_mode'
require 'aws-sdk-core/plugins/recursion_detection'
require 'aws-sdk-core/plugins/telemetry'
require 'aws-sdk-core/plugins/sign'
require 'aws-sdk-core/plugins/protocols/rest_json'
module Aws::IoTAnalytics
# An API client for IoTAnalytics. To construct a client, you need to configure a `:region` and `:credentials`.
#
# client = Aws::IoTAnalytics::Client.new(
# region: region_name,
# credentials: credentials,
# # ...
# )
#
# For details on configuring region and credentials see
# the [developer guide](/sdk-for-ruby/v3/developer-guide/setup-config.html).
#
# See {#initialize} for a full list of supported configuration options.
class Client < Seahorse::Client::Base
include Aws::ClientStubs
@identifier = :iotanalytics
set_api(ClientApi::API)
add_plugin(Seahorse::Client::Plugins::ContentLength)
add_plugin(Aws::Plugins::CredentialsConfiguration)
add_plugin(Aws::Plugins::Logging)
add_plugin(Aws::Plugins::ParamConverter)
add_plugin(Aws::Plugins::ParamValidator)
add_plugin(Aws::Plugins::UserAgent)
add_plugin(Aws::Plugins::HelpfulSocketErrors)
add_plugin(Aws::Plugins::RetryErrors)
add_plugin(Aws::Plugins::GlobalConfiguration)
add_plugin(Aws::Plugins::RegionalEndpoint)
add_plugin(Aws::Plugins::EndpointDiscovery)
add_plugin(Aws::Plugins::EndpointPattern)
add_plugin(Aws::Plugins::ResponsePaging)
add_plugin(Aws::Plugins::StubResponses)
add_plugin(Aws::Plugins::IdempotencyToken)
add_plugin(Aws::Plugins::InvocationId)
add_plugin(Aws::Plugins::JsonvalueConverter)
add_plugin(Aws::Plugins::ClientMetricsPlugin)
add_plugin(Aws::Plugins::ClientMetricsSendPlugin)
add_plugin(Aws::Plugins::TransferEncoding)
add_plugin(Aws::Plugins::HttpChecksum)
add_plugin(Aws::Plugins::ChecksumAlgorithm)
add_plugin(Aws::Plugins::RequestCompression)
add_plugin(Aws::Plugins::DefaultsMode)
add_plugin(Aws::Plugins::RecursionDetection)
add_plugin(Aws::Plugins::Telemetry)
add_plugin(Aws::Plugins::Sign)
add_plugin(Aws::Plugins::Protocols::RestJson)
add_plugin(Aws::IoTAnalytics::Plugins::Endpoints)
# @overload initialize(options)
# @param [Hash] options
#
# @option options [Array<Seahorse::Client::Plugin>] :plugins ([]])
# A list of plugins to apply to the client. Each plugin is either a
# class name or an instance of a plugin class.
#
# @option options [required, Aws::CredentialProvider] :credentials
# Your AWS credentials. This can be an instance of any one of the
# following classes:
#
# * `Aws::Credentials` - Used for configuring static, non-refreshing
# credentials.
#
# * `Aws::SharedCredentials` - Used for loading static credentials from a
# shared file, such as `~/.aws/config`.
#
# * `Aws::AssumeRoleCredentials` - Used when you need to assume a role.
#
# * `Aws::AssumeRoleWebIdentityCredentials` - Used when you need to
# assume a role after providing credentials via the web.
#
# * `Aws::SSOCredentials` - Used for loading credentials from AWS SSO using an
# access token generated from `aws login`.
#
# * `Aws::ProcessCredentials` - Used for loading credentials from a
# process that outputs to stdout.
#
# * `Aws::InstanceProfileCredentials` - Used for loading credentials
# from an EC2 IMDS on an EC2 instance.
#
# * `Aws::ECSCredentials` - Used for loading credentials from
# instances running in ECS.
#
# * `Aws::CognitoIdentityCredentials` - Used for loading credentials
# from the Cognito Identity service.
#
# When `:credentials` are not configured directly, the following
# locations will be searched for credentials:
#
# * `Aws.config[:credentials]`
# * The `:access_key_id`, `:secret_access_key`, `:session_token`, and
# `:account_id` options.
# * ENV['AWS_ACCESS_KEY_ID'], ENV['AWS_SECRET_ACCESS_KEY'],
# ENV['AWS_SESSION_TOKEN'], and ENV['AWS_ACCOUNT_ID']
# * `~/.aws/credentials`
# * `~/.aws/config`
# * EC2/ECS IMDS instance profile - When used by default, the timeouts
# are very aggressive. Construct and pass an instance of
# `Aws::InstanceProfileCredentials` or `Aws::ECSCredentials` to
# enable retries and extended timeouts. Instance profile credential
# fetching can be disabled by setting ENV['AWS_EC2_METADATA_DISABLED']
# to true.
#
# @option options [required, String] :region
# The AWS region to connect to. The configured `:region` is
# used to determine the service `:endpoint`. When not passed,
# a default `:region` is searched for in the following locations:
#
# * `Aws.config[:region]`
# * `ENV['AWS_REGION']`
# * `ENV['AMAZON_REGION']`
# * `ENV['AWS_DEFAULT_REGION']`
# * `~/.aws/credentials`
# * `~/.aws/config`
#
# @option options [String] :access_key_id
#
# @option options [String] :account_id
#
# @option options [Boolean] :active_endpoint_cache (false)
# When set to `true`, a thread polling for endpoints will be running in
# the background every 60 secs (default). Defaults to `false`.
#
# @option options [Boolean] :adaptive_retry_wait_to_fill (true)
# Used only in `adaptive` retry mode. When true, the request will sleep
# until there is sufficent client side capacity to retry the request.
# When false, the request will raise a `RetryCapacityNotAvailableError` and will
# not retry instead of sleeping.
#
# @option options [Boolean] :client_side_monitoring (false)
# When `true`, client-side metrics will be collected for all API requests from
# this client.
#
# @option options [String] :client_side_monitoring_client_id ("")
# Allows you to provide an identifier for this client which will be attached to
# all generated client side metrics. Defaults to an empty string.
#
# @option options [String] :client_side_monitoring_host ("127.0.0.1")
# Allows you to specify the DNS hostname or IPv4 or IPv6 address that the client
# side monitoring agent is running on, where client metrics will be published via UDP.
#
# @option options [Integer] :client_side_monitoring_port (31000)
# Required for publishing client metrics. The port that the client side monitoring
# agent is running on, where client metrics will be published via UDP.
#
# @option options [Aws::ClientSideMonitoring::Publisher] :client_side_monitoring_publisher (Aws::ClientSideMonitoring::Publisher)
# Allows you to provide a custom client-side monitoring publisher class. By default,
# will use the Client Side Monitoring Agent Publisher.
#
# @option options [Boolean] :convert_params (true)
# When `true`, an attempt is made to coerce request parameters into
# the required types.
#
# @option options [Boolean] :correct_clock_skew (true)
# Used only in `standard` and adaptive retry modes. Specifies whether to apply
# a clock skew correction and retry requests with skewed client clocks.
#
# @option options [String] :defaults_mode ("legacy")
# See {Aws::DefaultsModeConfiguration} for a list of the
# accepted modes and the configuration defaults that are included.
#
# @option options [Boolean] :disable_host_prefix_injection (false)
# Set to true to disable SDK automatically adding host prefix
# to default service endpoint when available.
#
# @option options [Boolean] :disable_request_compression (false)
# When set to 'true' the request body will not be compressed
# for supported operations.
#
# @option options [String, URI::HTTPS, URI::HTTP] :endpoint
# Normally you should not configure the `:endpoint` option
# directly. This is normally constructed from the `:region`
# option. Configuring `:endpoint` is normally reserved for
# connecting to test or custom endpoints. The endpoint should
# be a URI formatted like:
#
# 'http://example.com'
# 'https://example.com'
# 'http://example.com:123'
#
# @option options [Integer] :endpoint_cache_max_entries (1000)
# Used for the maximum size limit of the LRU cache storing endpoints data
# for endpoint discovery enabled operations. Defaults to 1000.
#
# @option options [Integer] :endpoint_cache_max_threads (10)
# Used for the maximum threads in use for polling endpoints to be cached, defaults to 10.
#
# @option options [Integer] :endpoint_cache_poll_interval (60)
# When :endpoint_discovery and :active_endpoint_cache is enabled,
# Use this option to config the time interval in seconds for making
# requests fetching endpoints information. Defaults to 60 sec.
#
# @option options [Boolean] :endpoint_discovery (false)
# When set to `true`, endpoint discovery will be enabled for operations when available.
#
# @option options [Boolean] :ignore_configured_endpoint_urls
# Setting to true disables use of endpoint URLs provided via environment
# variables and the shared configuration file.
#
# @option options [Aws::Log::Formatter] :log_formatter (Aws::Log::Formatter.default)
# The log formatter.
#
# @option options [Symbol] :log_level (:info)
# The log level to send messages to the `:logger` at.
#
# @option options [Logger] :logger
# The Logger instance to send log messages to. If this option
# is not set, logging will be disabled.
#
# @option options [Integer] :max_attempts (3)
# An integer representing the maximum number attempts that will be made for
# a single request, including the initial attempt. For example,
# setting this value to 5 will result in a request being retried up to
# 4 times. Used in `standard` and `adaptive` retry modes.
#
# @option options [String] :profile ("default")
# Used when loading credentials from the shared credentials file
# at HOME/.aws/credentials. When not specified, 'default' is used.
#
# @option options [String] :request_checksum_calculation ("when_supported")
# Determines when a checksum will be calculated for request payloads. Values are:
#
# * `when_supported` - (default) When set, a checksum will be
# calculated for all request payloads of operations modeled with the
# `httpChecksum` trait where `requestChecksumRequired` is `true` and/or a
# `requestAlgorithmMember` is modeled.
# * `when_required` - When set, a checksum will only be calculated for
# request payloads of operations modeled with the `httpChecksum` trait where
# `requestChecksumRequired` is `true` or where a `requestAlgorithmMember`
# is modeled and supplied.
#
# @option options [Integer] :request_min_compression_size_bytes (10240)
# The minimum size in bytes that triggers compression for request
# bodies. The value must be non-negative integer value between 0
# and 10485780 bytes inclusive.
#
# @option options [String] :response_checksum_validation ("when_supported")
# Determines when checksum validation will be performed on response payloads. Values are:
#
# * `when_supported` - (default) When set, checksum validation is performed on all
# response payloads of operations modeled with the `httpChecksum` trait where
# `responseAlgorithms` is modeled, except when no modeled checksum algorithms
# are supported.
# * `when_required` - When set, checksum validation is not performed on
# response payloads of operations unless the checksum algorithm is supported and
# the `requestValidationModeMember` member is set to `ENABLED`.
#
# @option options [Proc] :retry_backoff
# A proc or lambda used for backoff. Defaults to 2**retries * retry_base_delay.
# This option is only used in the `legacy` retry mode.
#
# @option options [Float] :retry_base_delay (0.3)
# The base delay in seconds used by the default backoff function. This option
# is only used in the `legacy` retry mode.
#
# @option options [Symbol] :retry_jitter (:none)
# A delay randomiser function used by the default backoff function.
# Some predefined functions can be referenced by name - :none, :equal, :full,
# otherwise a Proc that takes and returns a number. This option is only used
# in the `legacy` retry mode.
#
# @see https://www.awsarchitectureblog.com/2015/03/backoff.html
#
# @option options [Integer] :retry_limit (3)
# The maximum number of times to retry failed requests. Only
# ~ 500 level server errors and certain ~ 400 level client errors
# are retried. Generally, these are throttling errors, data
# checksum errors, networking errors, timeout errors, auth errors,
# endpoint discovery, and errors from expired credentials.
# This option is only used in the `legacy` retry mode.
#
# @option options [Integer] :retry_max_delay (0)
# The maximum number of seconds to delay between retries (0 for no limit)
# used by the default backoff function. This option is only used in the
# `legacy` retry mode.
#
# @option options [String] :retry_mode ("legacy")
# Specifies which retry algorithm to use. Values are:
#
# * `legacy` - The pre-existing retry behavior. This is default value if
# no retry mode is provided.
#
# * `standard` - A standardized set of retry rules across the AWS SDKs.
# This includes support for retry quotas, which limit the number of
# unsuccessful retries a client can make.
#
# * `adaptive` - An experimental retry mode that includes all the
# functionality of `standard` mode along with automatic client side
# throttling. This is a provisional mode that may change behavior
# in the future.
#
# @option options [String] :sdk_ua_app_id
# A unique and opaque application ID that is appended to the
# User-Agent header as app/sdk_ua_app_id. It should have a
# maximum length of 50. This variable is sourced from environment
# variable AWS_SDK_UA_APP_ID or the shared config profile attribute sdk_ua_app_id.
#
# @option options [String] :secret_access_key
#
# @option options [String] :session_token
#
# @option options [Array] :sigv4a_signing_region_set
# A list of regions that should be signed with SigV4a signing. When
# not passed, a default `:sigv4a_signing_region_set` is searched for
# in the following locations:
#
# * `Aws.config[:sigv4a_signing_region_set]`
# * `ENV['AWS_SIGV4A_SIGNING_REGION_SET']`
# * `~/.aws/config`
#
# @option options [Boolean] :stub_responses (false)
# Causes the client to return stubbed responses. By default
# fake responses are generated and returned. You can specify
# the response data to return or errors to raise by calling
# {ClientStubs#stub_responses}. See {ClientStubs} for more information.
#
# ** Please note ** When response stubbing is enabled, no HTTP
# requests are made, and retries are disabled.
#
# @option options [Aws::Telemetry::TelemetryProviderBase] :telemetry_provider (Aws::Telemetry::NoOpTelemetryProvider)
# Allows you to provide a telemetry provider, which is used to
# emit telemetry data. By default, uses `NoOpTelemetryProvider` which
# will not record or emit any telemetry data. The SDK supports the
# following telemetry providers:
#
# * OpenTelemetry (OTel) - To use the OTel provider, install and require the
# `opentelemetry-sdk` gem and then, pass in an instance of a
# `Aws::Telemetry::OTelProvider` for telemetry provider.
#
# @option options [Aws::TokenProvider] :token_provider
# A Bearer Token Provider. This can be an instance of any one of the
# following classes:
#
# * `Aws::StaticTokenProvider` - Used for configuring static, non-refreshing
# tokens.
#
# * `Aws::SSOTokenProvider` - Used for loading tokens from AWS SSO using an
# access token generated from `aws login`.
#
# When `:token_provider` is not configured directly, the `Aws::TokenProviderChain`
# will be used to search for tokens configured for your profile in shared configuration files.
#
# @option options [Boolean] :use_dualstack_endpoint
# When set to `true`, dualstack enabled endpoints (with `.aws` TLD)
# will be used if available.
#
# @option options [Boolean] :use_fips_endpoint
# When set to `true`, fips compatible endpoints will be used if available.
# When a `fips` region is used, the region is normalized and this config
# is set to `true`.
#
# @option options [Boolean] :validate_params (true)
# When `true`, request parameters are validated before
# sending the request.
#
# @option options [Aws::IoTAnalytics::EndpointProvider] :endpoint_provider
# The endpoint provider used to resolve endpoints. Any object that responds to
# `#resolve_endpoint(parameters)` where `parameters` is a Struct similar to
# `Aws::IoTAnalytics::EndpointParameters`.
#
# @option options [Float] :http_continue_timeout (1)
# The number of seconds to wait for a 100-continue response before sending the
# request body. This option has no effect unless the request has "Expect"
# header set to "100-continue". Defaults to `nil` which disables this
# behaviour. This value can safely be set per request on the session.
#
# @option options [Float] :http_idle_timeout (5)
# The number of seconds a connection is allowed to sit idle before it
# is considered stale. Stale connections are closed and removed from the
# pool before making a request.
#
# @option options [Float] :http_open_timeout (15)
# The default number of seconds to wait for response data.
# This value can safely be set per-request on the session.
#
# @option options [URI::HTTP,String] :http_proxy
# A proxy to send requests through. Formatted like 'http://proxy.com:123'.
#
# @option options [Float] :http_read_timeout (60)
# The default number of seconds to wait for response data.
# This value can safely be set per-request on the session.
#
# @option options [Boolean] :http_wire_trace (false)
# When `true`, HTTP debug output will be sent to the `:logger`.
#
# @option options [Proc] :on_chunk_received
# When a Proc object is provided, it will be used as callback when each chunk
# of the response body is received. It provides three arguments: the chunk,
# the number of bytes received, and the total number of
# bytes in the response (or nil if the server did not send a `content-length`).
#
# @option options [Proc] :on_chunk_sent
# When a Proc object is provided, it will be used as callback when each chunk
# of the request body is sent. It provides three arguments: the chunk,
# the number of bytes read from the body, and the total number of
# bytes in the body.
#
# @option options [Boolean] :raise_response_errors (true)
# When `true`, response errors are raised.
#
# @option options [String] :ssl_ca_bundle
# Full path to the SSL certificate authority bundle file that should be used when
# verifying peer certificates. If you do not pass `:ssl_ca_bundle` or
# `:ssl_ca_directory` the the system default will be used if available.
#
# @option options [String] :ssl_ca_directory
# Full path of the directory that contains the unbundled SSL certificate
# authority files for verifying peer certificates. If you do
# not pass `:ssl_ca_bundle` or `:ssl_ca_directory` the the system
# default will be used if available.
#
# @option options [String] :ssl_ca_store
# Sets the X509::Store to verify peer certificate.
#
# @option options [OpenSSL::X509::Certificate] :ssl_cert
# Sets a client certificate when creating http connections.
#
# @option options [OpenSSL::PKey] :ssl_key
# Sets a client key when creating http connections.
#
# @option options [Float] :ssl_timeout
# Sets the SSL timeout in seconds
#
# @option options [Boolean] :ssl_verify_peer (true)
# When `true`, SSL peer certificates are verified when establishing a connection.
#
def initialize(*args)
super
end
# @!group API Operations
# Sends messages to a channel.
#
# @option params [required, String] :channel_name
# The name of the channel where the messages are sent.
#
# @option params [required, Array<Types::Message>] :messages
# The list of messages to be sent. Each message has the format: \{
# "messageId": "string", "payload": "string"}.
#
# The field names of message payloads (data) that you send to IoT
# Analytics:
#
# * Must contain only alphanumeric characters and undescores (\_). No
# other special characters are allowed.
#
# * Must begin with an alphabetic character or single underscore (\_).
#
# * Cannot contain hyphens (-).
#
# * In regular expression terms:
# "^\[A-Za-z\_\](\[A-Za-z0-9\]*\|\[A-Za-z0-9\]\[A-Za-z0-9\_\]*)$".
#
# * Cannot be more than 255 characters.
#
# * Are case insensitive. (Fields named foo and FOO in the same payload
# are considered duplicates.)
#
# For example, \{"temp\_01": 29} or \{"\_temp\_01": 29} are valid,
# but \{"temp-01": 29}, \{"01\_temp": 29} or \{"\_\_temp\_01": 29}
# are invalid in message payloads.
#
# @return [Types::BatchPutMessageResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
#
# * {Types::BatchPutMessageResponse#batch_put_message_error_entries #batch_put_message_error_entries} => Array<Types::BatchPutMessageErrorEntry>
#
# @example Request syntax with placeholder values
#
# resp = client.batch_put_message({
# channel_name: "ChannelName", # required
# messages: [ # required
# {
# message_id: "MessageId", # required
# payload: "data", # required
# },
# ],
# })
#
# @example Response structure
#
# resp.batch_put_message_error_entries #=> Array
# resp.batch_put_message_error_entries[0].message_id #=> String
# resp.batch_put_message_error_entries[0].error_code #=> String
# resp.batch_put_message_error_entries[0].error_message #=> String
#
# @overload batch_put_message(params = {})
# @param [Hash] params ({})
def batch_put_message(params = {}, options = {})
req = build_request(:batch_put_message, params)
req.send_request(options)
end
# Cancels the reprocessing of data through the pipeline.
#
# @option params [required, String] :pipeline_name
# The name of pipeline for which data reprocessing is canceled.
#
# @option params [required, String] :reprocessing_id
# The ID of the reprocessing task (returned by
# `StartPipelineReprocessing`).
#
# @return [Struct] Returns an empty {Seahorse::Client::Response response}.
#
# @example Request syntax with placeholder values
#
# resp = client.cancel_pipeline_reprocessing({
# pipeline_name: "PipelineName", # required
# reprocessing_id: "ReprocessingId", # required
# })
#
# @overload cancel_pipeline_reprocessing(params = {})
# @param [Hash] params ({})
def cancel_pipeline_reprocessing(params = {}, options = {})
req = build_request(:cancel_pipeline_reprocessing, params)
req.send_request(options)
end
# Used to create a channel. A channel collects data from an MQTT topic
# and archives the raw, unprocessed messages before publishing the data
# to a pipeline.
#
# @option params [required, String] :channel_name
# The name of the channel.
#
# @option params [Types::ChannelStorage] :channel_storage
# Where channel data is stored. You can choose one of `serviceManagedS3`
# or `customerManagedS3` storage. If not specified, the default is
# `serviceManagedS3`. You can't change this storage option after the
# channel is created.
#
# @option params [Types::RetentionPeriod] :retention_period
# How long, in days, message data is kept for the channel. When
# `customerManagedS3` storage is selected, this parameter is ignored.
#
# @option params [Array<Types::Tag>] :tags
# Metadata which can be used to manage the channel.
#
# @return [Types::CreateChannelResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
#
# * {Types::CreateChannelResponse#channel_name #channel_name} => String
# * {Types::CreateChannelResponse#channel_arn #channel_arn} => String
# * {Types::CreateChannelResponse#retention_period #retention_period} => Types::RetentionPeriod
#
# @example Request syntax with placeholder values
#
# resp = client.create_channel({
# channel_name: "ChannelName", # required
# channel_storage: {
# service_managed_s3: {
# },
# customer_managed_s3: {
# bucket: "BucketName", # required
# key_prefix: "S3KeyPrefix",
# role_arn: "RoleArn", # required
# },
# },
# retention_period: {
# unlimited: false,
# number_of_days: 1,
# },
# tags: [
# {
# key: "TagKey", # required
# value: "TagValue", # required
# },
# ],
# })
#
# @example Response structure
#
# resp.channel_name #=> String
# resp.channel_arn #=> String
# resp.retention_period.unlimited #=> Boolean
# resp.retention_period.number_of_days #=> Integer
#
# @overload create_channel(params = {})
# @param [Hash] params ({})
def create_channel(params = {}, options = {})
req = build_request(:create_channel, params)
req.send_request(options)
end
# Used to create a dataset. A dataset stores data retrieved from a data
# store by applying a `queryAction` (a SQL query) or a `containerAction`
# (executing a containerized application). This operation creates the
# skeleton of a dataset. The dataset can be populated manually by
# calling `CreateDatasetContent` or automatically according to a trigger
# you specify.
#
# @option params [required, String] :dataset_name
# The name of the dataset.
#
# @option params [required, Array<Types::DatasetAction>] :actions
# A list of actions that create the dataset contents.
#
# @option params [Array<Types::DatasetTrigger>] :triggers
# A list of triggers. A trigger causes dataset contents to be populated
# at a specified time interval or when another dataset's contents are
# created. The list of triggers can be empty or contain up to five
# `DataSetTrigger` objects.
#
# @option params [Array<Types::DatasetContentDeliveryRule>] :content_delivery_rules
# When dataset contents are created, they are delivered to destinations
# specified here.
#
# @option params [Types::RetentionPeriod] :retention_period
# Optional. How long, in days, versions of dataset contents are kept for
# the dataset. If not specified or set to `null`, versions of dataset
# contents are retained for at most 90 days. The number of versions of
# dataset contents retained is determined by the
# `versioningConfiguration` parameter. For more information, see [
# Keeping Multiple Versions of IoT Analytics datasets][1] in the *IoT
# Analytics User Guide*.
#
#
#
# [1]: https://docs.aws.amazon.com/iotanalytics/latest/userguide/getting-started.html#aws-iot-analytics-dataset-versions
#
# @option params [Types::VersioningConfiguration] :versioning_configuration
# Optional. How many versions of dataset contents are kept. If not
# specified or set to null, only the latest version plus the latest
# succeeded version (if they are different) are kept for the time period
# specified by the `retentionPeriod` parameter. For more information,
# see [Keeping Multiple Versions of IoT Analytics datasets][1] in the
# *IoT Analytics User Guide*.
#
#
#
# [1]: https://docs.aws.amazon.com/iotanalytics/latest/userguide/getting-started.html#aws-iot-analytics-dataset-versions
#
# @option params [Array<Types::Tag>] :tags
# Metadata which can be used to manage the dataset.
#
# @option params [Array<Types::LateDataRule>] :late_data_rules
# A list of data rules that send notifications to CloudWatch, when data
# arrives late. To specify `lateDataRules`, the dataset must use a
# [DeltaTimer][1] filter.
#
#
#
# [1]: https://docs.aws.amazon.com/iotanalytics/latest/APIReference/API_DeltaTime.html
#
# @return [Types::CreateDatasetResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
#
# * {Types::CreateDatasetResponse#dataset_name #dataset_name} => String
# * {Types::CreateDatasetResponse#dataset_arn #dataset_arn} => String
# * {Types::CreateDatasetResponse#retention_period #retention_period} => Types::RetentionPeriod
#
# @example Request syntax with placeholder values
#
# resp = client.create_dataset({
# dataset_name: "DatasetName", # required
# actions: [ # required
# {
# action_name: "DatasetActionName",
# query_action: {
# sql_query: "SqlQuery", # required
# filters: [
# {
# delta_time: {
# offset_seconds: 1, # required
# time_expression: "TimeExpression", # required
# },
# },
# ],
# },
# container_action: {
# image: "Image", # required
# execution_role_arn: "RoleArn", # required
# resource_configuration: { # required
# compute_type: "ACU_1", # required, accepts ACU_1, ACU_2
# volume_size_in_gb: 1, # required
# },
# variables: [
# {
# name: "VariableName", # required
# string_value: "StringValue",
# double_value: 1.0,
# dataset_content_version_value: {
# dataset_name: "DatasetName", # required
# },
# output_file_uri_value: {
# file_name: "OutputFileName", # required
# },
# },
# ],
# },
# },
# ],
# triggers: [
# {
# schedule: {
# expression: "ScheduleExpression",
# },
# dataset: {
# name: "DatasetName", # required
# },
# },
# ],
# content_delivery_rules: [
# {
# entry_name: "EntryName",
# destination: { # required
# iot_events_destination_configuration: {
# input_name: "IotEventsInputName", # required
# role_arn: "RoleArn", # required
# },
# s3_destination_configuration: {
# bucket: "BucketName", # required
# key: "BucketKeyExpression", # required
# glue_configuration: {
# table_name: "GlueTableName", # required
# database_name: "GlueDatabaseName", # required
# },
# role_arn: "RoleArn", # required
# },
# },
# },
# ],
# retention_period: {
# unlimited: false,
# number_of_days: 1,
# },
# versioning_configuration: {
# unlimited: false,
# max_versions: 1,
# },
# tags: [
# {
# key: "TagKey", # required
# value: "TagValue", # required
# },
# ],
# late_data_rules: [
# {
# rule_name: "LateDataRuleName",
# rule_configuration: { # required
# delta_time_session_window_configuration: {
# timeout_in_minutes: 1, # required
# },
# },
# },
# ],
# })
#
# @example Response structure
#
# resp.dataset_name #=> String
# resp.dataset_arn #=> String
# resp.retention_period.unlimited #=> Boolean
# resp.retention_period.number_of_days #=> Integer
#
# @overload create_dataset(params = {})
# @param [Hash] params ({})
def create_dataset(params = {}, options = {})
req = build_request(:create_dataset, params)
req.send_request(options)
end
# Creates the content of a dataset by applying a `queryAction` (a SQL
# query) or a `containerAction` (executing a containerized application).
#
# @option params [required, String] :dataset_name
# The name of the dataset.
#
# @option params [String] :version_id
# The version ID of the dataset content. To specify `versionId` for a
# dataset content, the dataset must use a [DeltaTimer][1] filter.
#
#
#
# [1]: https://docs.aws.amazon.com/iotanalytics/latest/APIReference/API_DeltaTime.html
#
# @return [Types::CreateDatasetContentResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
#
# * {Types::CreateDatasetContentResponse#version_id #version_id} => String
#
# @example Request syntax with placeholder values
#
# resp = client.create_dataset_content({
# dataset_name: "DatasetName", # required
# version_id: "DatasetContentVersion",
# })
#
# @example Response structure
#
# resp.version_id #=> String
#
# @overload create_dataset_content(params = {})
# @param [Hash] params ({})
def create_dataset_content(params = {}, options = {})
req = build_request(:create_dataset_content, params)
req.send_request(options)
end
# Creates a data store, which is a repository for messages.
#
# @option params [required, String] :datastore_name
# The name of the data store.
#
# @option params [Types::DatastoreStorage] :datastore_storage
# Where data in a data store is stored.. You can choose
# `serviceManagedS3` storage, `customerManagedS3` storage, or
# `iotSiteWiseMultiLayerStorage` storage. The default is
# `serviceManagedS3`. You can't change the choice of Amazon S3 storage
# after your data store is created.
#
# @option params [Types::RetentionPeriod] :retention_period
# How long, in days, message data is kept for the data store. When
# `customerManagedS3` storage is selected, this parameter is ignored.
#
# @option params [Array<Types::Tag>] :tags
# Metadata which can be used to manage the data store.
#
# @option params [Types::FileFormatConfiguration] :file_format_configuration
# Contains the configuration information of file formats. IoT Analytics
# data stores support JSON and [Parquet][1].
#
# The default file format is JSON. You can specify only one format.
#
# You can't change the file format after you create the data store.
#
#
#
# [1]: https://parquet.apache.org/
#
# @option params [Types::DatastorePartitions] :datastore_partitions
# Contains information about the partition dimensions in a data store.
#
# @return [Types::CreateDatastoreResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
#
# * {Types::CreateDatastoreResponse#datastore_name #datastore_name} => String
# * {Types::CreateDatastoreResponse#datastore_arn #datastore_arn} => String
# * {Types::CreateDatastoreResponse#retention_period #retention_period} => Types::RetentionPeriod
#
# @example Request syntax with placeholder values
#
# resp = client.create_datastore({
# datastore_name: "DatastoreName", # required
# datastore_storage: {
# service_managed_s3: {
# },
# customer_managed_s3: {
# bucket: "BucketName", # required
# key_prefix: "S3KeyPrefix",
# role_arn: "RoleArn", # required
# },
# iot_site_wise_multi_layer_storage: {
# customer_managed_s3_storage: { # required
# bucket: "BucketName", # required
# key_prefix: "S3KeyPrefix",
# },
# },
# },
# retention_period: {
# unlimited: false,
# number_of_days: 1,
# },
# tags: [
# {
# key: "TagKey", # required
# value: "TagValue", # required
# },
# ],
# file_format_configuration: {
# json_configuration: {
# },
# parquet_configuration: {
# schema_definition: {
# columns: [
# {
# name: "ColumnName", # required
# type: "ColumnDataType", # required
# },
# ],
# },
# },
# },
# datastore_partitions: {
# partitions: [
# {
# attribute_partition: {
# attribute_name: "PartitionAttributeName", # required
# },
# timestamp_partition: {
# attribute_name: "PartitionAttributeName", # required
# timestamp_format: "TimestampFormat",
# },
# },
# ],
# },
# })
#
# @example Response structure
#
# resp.datastore_name #=> String
# resp.datastore_arn #=> String
# resp.retention_period.unlimited #=> Boolean
# resp.retention_period.number_of_days #=> Integer
#
# @overload create_datastore(params = {})
# @param [Hash] params ({})
def create_datastore(params = {}, options = {})
req = build_request(:create_datastore, params)
req.send_request(options)
end
# Creates a pipeline. A pipeline consumes messages from a channel and
# allows you to process the messages before storing them in a data
# store. You must specify both a `channel` and a `datastore` activity
# and, optionally, as many as 23 additional activities in the
# `pipelineActivities` array.
#
# @option params [required, String] :pipeline_name
# The name of the pipeline.
#
# @option params [required, Array<Types::PipelineActivity>] :pipeline_activities
# A list of `PipelineActivity` objects. Activities perform
# transformations on your messages, such as removing, renaming or adding
# message attributes; filtering messages based on attribute values;
# invoking your Lambda unctions on messages for advanced processing; or
# performing mathematical transformations to normalize device data.
#
# The list can be 2-25 `PipelineActivity` objects and must contain both
# a `channel` and a `datastore` activity. Each entry in the list must
# contain only one activity. For example:
#
# `pipelineActivities = [ { "channel": { ... } }, { "lambda": { ... } },
# ... ]`
#
# @option params [Array<Types::Tag>] :tags
# Metadata which can be used to manage the pipeline.
#
# @return [Types::CreatePipelineResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
#
# * {Types::CreatePipelineResponse#pipeline_name #pipeline_name} => String
# * {Types::CreatePipelineResponse#pipeline_arn #pipeline_arn} => String
#
# @example Request syntax with placeholder values
#
# resp = client.create_pipeline({
# pipeline_name: "PipelineName", # required
# pipeline_activities: [ # required
# {
# channel: {
# name: "ActivityName", # required
# channel_name: "ChannelName", # required
# next: "ActivityName",
# },
# lambda: {
# name: "ActivityName", # required
# lambda_name: "LambdaName", # required
# batch_size: 1, # required
# next: "ActivityName",
# },
# datastore: {
# name: "ActivityName", # required
# datastore_name: "DatastoreName", # required
# },
# add_attributes: {
# name: "ActivityName", # required
# attributes: { # required
# "AttributeName" => "AttributeName",
# },
# next: "ActivityName",
# },
# remove_attributes: {
# name: "ActivityName", # required
# attributes: ["AttributeName"], # required
# next: "ActivityName",
# },
# select_attributes: {
# name: "ActivityName", # required
# attributes: ["AttributeName"], # required
# next: "ActivityName",
# },
# filter: {
# name: "ActivityName", # required
# filter: "FilterExpression", # required
# next: "ActivityName",
# },
# math: {
# name: "ActivityName", # required
# attribute: "AttributeName", # required
# math: "MathExpression", # required
# next: "ActivityName",
# },
# device_registry_enrich: {
# name: "ActivityName", # required
# attribute: "AttributeName", # required
# thing_name: "AttributeName", # required
# role_arn: "RoleArn", # required
# next: "ActivityName",
# },
# device_shadow_enrich: {
# name: "ActivityName", # required
# attribute: "AttributeName", # required
# thing_name: "AttributeName", # required
# role_arn: "RoleArn", # required
# next: "ActivityName",
# },
# },
# ],
# tags: [
# {
# key: "TagKey", # required
# value: "TagValue", # required
# },
# ],
# })
#
# @example Response structure
#
# resp.pipeline_name #=> String
# resp.pipeline_arn #=> String
#
# @overload create_pipeline(params = {})
# @param [Hash] params ({})
def create_pipeline(params = {}, options = {})
req = build_request(:create_pipeline, params)
req.send_request(options)
end
# Deletes the specified channel.
#
# @option params [required, String] :channel_name
# The name of the channel to delete.
#
# @return [Struct] Returns an empty {Seahorse::Client::Response response}.
#
# @example Request syntax with placeholder values
#
# resp = client.delete_channel({
# channel_name: "ChannelName", # required
# })
#
# @overload delete_channel(params = {})
# @param [Hash] params ({})
def delete_channel(params = {}, options = {})
req = build_request(:delete_channel, params)
req.send_request(options)
end
# Deletes the specified dataset.
#
# You do not have to delete the content of the dataset before you
# perform this operation.
#
# @option params [required, String] :dataset_name
# The name of the dataset to delete.
#
# @return [Struct] Returns an empty {Seahorse::Client::Response response}.
#
# @example Request syntax with placeholder values
#
# resp = client.delete_dataset({
# dataset_name: "DatasetName", # required
# })
#
# @overload delete_dataset(params = {})
# @param [Hash] params ({})
def delete_dataset(params = {}, options = {})
req = build_request(:delete_dataset, params)
req.send_request(options)
end
# Deletes the content of the specified dataset.
#
# @option params [required, String] :dataset_name
# The name of the dataset whose content is deleted.
#
# @option params [String] :version_id
# The version of the dataset whose content is deleted. You can also use
# the strings "$LATEST" or "$LATEST\_SUCCEEDED" to delete the latest
# or latest successfully completed data set. If not specified,
# "$LATEST\_SUCCEEDED" is the default.
#
# @return [Struct] Returns an empty {Seahorse::Client::Response response}.
#
# @example Request syntax with placeholder values
#
# resp = client.delete_dataset_content({
# dataset_name: "DatasetName", # required
# version_id: "DatasetContentVersion",
# })
#
# @overload delete_dataset_content(params = {})
# @param [Hash] params ({})
def delete_dataset_content(params = {}, options = {})
req = build_request(:delete_dataset_content, params)
req.send_request(options)
end
# Deletes the specified data store.
#
# @option params [required, String] :datastore_name
# The name of the data store to delete.
#
# @return [Struct] Returns an empty {Seahorse::Client::Response response}.
#
# @example Request syntax with placeholder values
#
# resp = client.delete_datastore({
# datastore_name: "DatastoreName", # required
# })
#
# @overload delete_datastore(params = {})
# @param [Hash] params ({})
def delete_datastore(params = {}, options = {})
req = build_request(:delete_datastore, params)
req.send_request(options)
end
# Deletes the specified pipeline.
#
# @option params [required, String] :pipeline_name
# The name of the pipeline to delete.
#
# @return [Struct] Returns an empty {Seahorse::Client::Response response}.
#
# @example Request syntax with placeholder values
#
# resp = client.delete_pipeline({
# pipeline_name: "PipelineName", # required
# })
#
# @overload delete_pipeline(params = {})
# @param [Hash] params ({})
def delete_pipeline(params = {}, options = {})
req = build_request(:delete_pipeline, params)
req.send_request(options)
end
# Retrieves information about a channel.
#
# @option params [required, String] :channel_name
# The name of the channel whose information is retrieved.
#
# @option params [Boolean] :include_statistics
# If true, additional statistical information about the channel is
# included in the response. This feature can't be used with a channel
# whose S3 storage is customer-managed.
#
# @return [Types::DescribeChannelResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
#
# * {Types::DescribeChannelResponse#channel #channel} => Types::Channel
# * {Types::DescribeChannelResponse#statistics #statistics} => Types::ChannelStatistics
#
# @example Request syntax with placeholder values
#
# resp = client.describe_channel({
# channel_name: "ChannelName", # required
# include_statistics: false,
# })
#
# @example Response structure
#
# resp.channel.name #=> String
# resp.channel.storage.customer_managed_s3.bucket #=> String
# resp.channel.storage.customer_managed_s3.key_prefix #=> String
# resp.channel.storage.customer_managed_s3.role_arn #=> String
# resp.channel.arn #=> String
# resp.channel.status #=> String, one of "CREATING", "ACTIVE", "DELETING"
# resp.channel.retention_period.unlimited #=> Boolean
# resp.channel.retention_period.number_of_days #=> Integer
# resp.channel.creation_time #=> Time
# resp.channel.last_update_time #=> Time
# resp.channel.last_message_arrival_time #=> Time
# resp.statistics.size.estimated_size_in_bytes #=> Float
# resp.statistics.size.estimated_on #=> Time
#
# @overload describe_channel(params = {})
# @param [Hash] params ({})
def describe_channel(params = {}, options = {})
req = build_request(:describe_channel, params)
req.send_request(options)
end
# Retrieves information about a dataset.
#
# @option params [required, String] :dataset_name
# The name of the dataset whose information is retrieved.
#
# @return [Types::DescribeDatasetResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
#
# * {Types::DescribeDatasetResponse#dataset #dataset} => Types::Dataset
#
# @example Request syntax with placeholder values
#
# resp = client.describe_dataset({
# dataset_name: "DatasetName", # required
# })
#
# @example Response structure
#
# resp.dataset.name #=> String
# resp.dataset.arn #=> String
# resp.dataset.actions #=> Array
# resp.dataset.actions[0].action_name #=> String
# resp.dataset.actions[0].query_action.sql_query #=> String
# resp.dataset.actions[0].query_action.filters #=> Array
# resp.dataset.actions[0].query_action.filters[0].delta_time.offset_seconds #=> Integer
# resp.dataset.actions[0].query_action.filters[0].delta_time.time_expression #=> String
# resp.dataset.actions[0].container_action.image #=> String
# resp.dataset.actions[0].container_action.execution_role_arn #=> String
# resp.dataset.actions[0].container_action.resource_configuration.compute_type #=> String, one of "ACU_1", "ACU_2"
# resp.dataset.actions[0].container_action.resource_configuration.volume_size_in_gb #=> Integer
# resp.dataset.actions[0].container_action.variables #=> Array
# resp.dataset.actions[0].container_action.variables[0].name #=> String
# resp.dataset.actions[0].container_action.variables[0].string_value #=> String
# resp.dataset.actions[0].container_action.variables[0].double_value #=> Float
# resp.dataset.actions[0].container_action.variables[0].dataset_content_version_value.dataset_name #=> String
# resp.dataset.actions[0].container_action.variables[0].output_file_uri_value.file_name #=> String
# resp.dataset.triggers #=> Array
# resp.dataset.triggers[0].schedule.expression #=> String
# resp.dataset.triggers[0].dataset.name #=> String
# resp.dataset.content_delivery_rules #=> Array
# resp.dataset.content_delivery_rules[0].entry_name #=> String
# resp.dataset.content_delivery_rules[0].destination.iot_events_destination_configuration.input_name #=> String
# resp.dataset.content_delivery_rules[0].destination.iot_events_destination_configuration.role_arn #=> String
# resp.dataset.content_delivery_rules[0].destination.s3_destination_configuration.bucket #=> String
# resp.dataset.content_delivery_rules[0].destination.s3_destination_configuration.key #=> String
# resp.dataset.content_delivery_rules[0].destination.s3_destination_configuration.glue_configuration.table_name #=> String
# resp.dataset.content_delivery_rules[0].destination.s3_destination_configuration.glue_configuration.database_name #=> String
# resp.dataset.content_delivery_rules[0].destination.s3_destination_configuration.role_arn #=> String
# resp.dataset.status #=> String, one of "CREATING", "ACTIVE", "DELETING"
# resp.dataset.creation_time #=> Time
# resp.dataset.last_update_time #=> Time
# resp.dataset.retention_period.unlimited #=> Boolean
# resp.dataset.retention_period.number_of_days #=> Integer
# resp.dataset.versioning_configuration.unlimited #=> Boolean
# resp.dataset.versioning_configuration.max_versions #=> Integer
# resp.dataset.late_data_rules #=> Array
# resp.dataset.late_data_rules[0].rule_name #=> String
# resp.dataset.late_data_rules[0].rule_configuration.delta_time_session_window_configuration.timeout_in_minutes #=> Integer
#
# @overload describe_dataset(params = {})
# @param [Hash] params ({})
def describe_dataset(params = {}, options = {})
req = build_request(:describe_dataset, params)
req.send_request(options)
end
# Retrieves information about a data store.
#
# @option params [required, String] :datastore_name
# The name of the data store
#
# @option params [Boolean] :include_statistics
# If true, additional statistical information about the data store is
# included in the response. This feature can't be used with a data
# store whose S3 storage is customer-managed.
#
# @return [Types::DescribeDatastoreResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
#
# * {Types::DescribeDatastoreResponse#datastore #datastore} => Types::Datastore
# * {Types::DescribeDatastoreResponse#statistics #statistics} => Types::DatastoreStatistics
#
# @example Request syntax with placeholder values
#
# resp = client.describe_datastore({
# datastore_name: "DatastoreName", # required
# include_statistics: false,
# })
#
# @example Response structure
#
# resp.datastore.name #=> String
# resp.datastore.storage.customer_managed_s3.bucket #=> String
# resp.datastore.storage.customer_managed_s3.key_prefix #=> String
# resp.datastore.storage.customer_managed_s3.role_arn #=> String
# resp.datastore.storage.iot_site_wise_multi_layer_storage.customer_managed_s3_storage.bucket #=> String
# resp.datastore.storage.iot_site_wise_multi_layer_storage.customer_managed_s3_storage.key_prefix #=> String
# resp.datastore.arn #=> String
# resp.datastore.status #=> String, one of "CREATING", "ACTIVE", "DELETING"
# resp.datastore.retention_period.unlimited #=> Boolean
# resp.datastore.retention_period.number_of_days #=> Integer
# resp.datastore.creation_time #=> Time
# resp.datastore.last_update_time #=> Time
# resp.datastore.last_message_arrival_time #=> Time
# resp.datastore.file_format_configuration.parquet_configuration.schema_definition.columns #=> Array
# resp.datastore.file_format_configuration.parquet_configuration.schema_definition.columns[0].name #=> String
# resp.datastore.file_format_configuration.parquet_configuration.schema_definition.columns[0].type #=> String
# resp.datastore.datastore_partitions.partitions #=> Array
# resp.datastore.datastore_partitions.partitions[0].attribute_partition.attribute_name #=> String
# resp.datastore.datastore_partitions.partitions[0].timestamp_partition.attribute_name #=> String
# resp.datastore.datastore_partitions.partitions[0].timestamp_partition.timestamp_format #=> String
# resp.statistics.size.estimated_size_in_bytes #=> Float
# resp.statistics.size.estimated_on #=> Time
#
# @overload describe_datastore(params = {})
# @param [Hash] params ({})
def describe_datastore(params = {}, options = {})
req = build_request(:describe_datastore, params)
req.send_request(options)
end
# Retrieves the current settings of the IoT Analytics logging options.
#
# @return [Types::DescribeLoggingOptionsResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
#
# * {Types::DescribeLoggingOptionsResponse#logging_options #logging_options} => Types::LoggingOptions
#
# @example Response structure
#
# resp.logging_options.role_arn #=> String
# resp.logging_options.level #=> String, one of "ERROR"
# resp.logging_options.enabled #=> Boolean
#
# @overload describe_logging_options(params = {})
# @param [Hash] params ({})
def describe_logging_options(params = {}, options = {})
req = build_request(:describe_logging_options, params)
req.send_request(options)
end
# Retrieves information about a pipeline.
#
# @option params [required, String] :pipeline_name
# The name of the pipeline whose information is retrieved.
#
# @return [Types::DescribePipelineResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
#
# * {Types::DescribePipelineResponse#pipeline #pipeline} => Types::Pipeline
#
# @example Request syntax with placeholder values
#
# resp = client.describe_pipeline({
# pipeline_name: "PipelineName", # required
# })
#
# @example Response structure
#
# resp.pipeline.name #=> String
# resp.pipeline.arn #=> String
# resp.pipeline.activities #=> Array
# resp.pipeline.activities[0].channel.name #=> String
# resp.pipeline.activities[0].channel.channel_name #=> String
# resp.pipeline.activities[0].channel.next #=> String
# resp.pipeline.activities[0].lambda.name #=> String
# resp.pipeline.activities[0].lambda.lambda_name #=> String
# resp.pipeline.activities[0].lambda.batch_size #=> Integer
# resp.pipeline.activities[0].lambda.next #=> String
# resp.pipeline.activities[0].datastore.name #=> String
# resp.pipeline.activities[0].datastore.datastore_name #=> String
# resp.pipeline.activities[0].add_attributes.name #=> String
# resp.pipeline.activities[0].add_attributes.attributes #=> Hash
# resp.pipeline.activities[0].add_attributes.attributes["AttributeName"] #=> String
# resp.pipeline.activities[0].add_attributes.next #=> String
# resp.pipeline.activities[0].remove_attributes.name #=> String
# resp.pipeline.activities[0].remove_attributes.attributes #=> Array
# resp.pipeline.activities[0].remove_attributes.attributes[0] #=> String
# resp.pipeline.activities[0].remove_attributes.next #=> String
# resp.pipeline.activities[0].select_attributes.name #=> String
# resp.pipeline.activities[0].select_attributes.attributes #=> Array
# resp.pipeline.activities[0].select_attributes.attributes[0] #=> String
# resp.pipeline.activities[0].select_attributes.next #=> String
# resp.pipeline.activities[0].filter.name #=> String
# resp.pipeline.activities[0].filter.filter #=> String
# resp.pipeline.activities[0].filter.next #=> String
# resp.pipeline.activities[0].math.name #=> String
# resp.pipeline.activities[0].math.attribute #=> String
# resp.pipeline.activities[0].math.math #=> String
# resp.pipeline.activities[0].math.next #=> String
# resp.pipeline.activities[0].device_registry_enrich.name #=> String
# resp.pipeline.activities[0].device_registry_enrich.attribute #=> String
# resp.pipeline.activities[0].device_registry_enrich.thing_name #=> String
# resp.pipeline.activities[0].device_registry_enrich.role_arn #=> String
# resp.pipeline.activities[0].device_registry_enrich.next #=> String
# resp.pipeline.activities[0].device_shadow_enrich.name #=> String
# resp.pipeline.activities[0].device_shadow_enrich.attribute #=> String
# resp.pipeline.activities[0].device_shadow_enrich.thing_name #=> String
# resp.pipeline.activities[0].device_shadow_enrich.role_arn #=> String
# resp.pipeline.activities[0].device_shadow_enrich.next #=> String
# resp.pipeline.reprocessing_summaries #=> Array
# resp.pipeline.reprocessing_summaries[0].id #=> String
# resp.pipeline.reprocessing_summaries[0].status #=> String, one of "RUNNING", "SUCCEEDED", "CANCELLED", "FAILED"
# resp.pipeline.reprocessing_summaries[0].creation_time #=> Time
# resp.pipeline.creation_time #=> Time
# resp.pipeline.last_update_time #=> Time
#
# @overload describe_pipeline(params = {})
# @param [Hash] params ({})
def describe_pipeline(params = {}, options = {})
req = build_request(:describe_pipeline, params)
req.send_request(options)
end
# Retrieves the contents of a dataset as presigned URIs.
#
# @option params [required, String] :dataset_name
# The name of the dataset whose contents are retrieved.
#
# @option params [String] :version_id
# The version of the dataset whose contents are retrieved. You can also
# use the strings "$LATEST" or "$LATEST\_SUCCEEDED" to retrieve the
# contents of the latest or latest successfully completed dataset. If
# not specified, "$LATEST\_SUCCEEDED" is the default.
#
# @return [Types::GetDatasetContentResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
#
# * {Types::GetDatasetContentResponse#entries #data.entries} => Array<Types::DatasetEntry> (This method conflicts with a method on Response, call it through the data member)
# * {Types::GetDatasetContentResponse#timestamp #timestamp} => Time
# * {Types::GetDatasetContentResponse#status #status} => Types::DatasetContentStatus
#
# @example Request syntax with placeholder values
#
# resp = client.get_dataset_content({
# dataset_name: "DatasetName", # required
# version_id: "DatasetContentVersion",
# })
#
# @example Response structure
#
# resp.data.entries #=> Array
# resp.data.entries[0].entry_name #=> String
# resp.data.entries[0].data_uri #=> String
# resp.timestamp #=> Time
# resp.status.state #=> String, one of "CREATING", "SUCCEEDED", "FAILED"
# resp.status.reason #=> String
#
# @overload get_dataset_content(params = {})
# @param [Hash] params ({})
def get_dataset_content(params = {}, options = {})
req = build_request(:get_dataset_content, params)
req.send_request(options)
end
# Retrieves a list of channels.
#
# @option params [String] :next_token
# The token for the next set of results.
#
# @option params [Integer] :max_results
# The maximum number of results to return in this request.
#
# The default value is 100.
#
# @return [Types::ListChannelsResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
#
# * {Types::ListChannelsResponse#channel_summaries #channel_summaries} => Array<Types::ChannelSummary>
# * {Types::ListChannelsResponse#next_token #next_token} => String
#
# The returned {Seahorse::Client::Response response} is a pageable response and is Enumerable. For details on usage see {Aws::PageableResponse PageableResponse}.
#
# @example Request syntax with placeholder values
#
# resp = client.list_channels({
# next_token: "NextToken",
# max_results: 1,
# })
#
# @example Response structure
#
# resp.channel_summaries #=> Array
# resp.channel_summaries[0].channel_name #=> String
# resp.channel_summaries[0].channel_storage.customer_managed_s3.bucket #=> String
# resp.channel_summaries[0].channel_storage.customer_managed_s3.key_prefix #=> String
# resp.channel_summaries[0].channel_storage.customer_managed_s3.role_arn #=> String
# resp.channel_summaries[0].status #=> String, one of "CREATING", "ACTIVE", "DELETING"
# resp.channel_summaries[0].creation_time #=> Time
# resp.channel_summaries[0].last_update_time #=> Time
# resp.channel_summaries[0].last_message_arrival_time #=> Time
# resp.next_token #=> String
#
# @overload list_channels(params = {})
# @param [Hash] params ({})
def list_channels(params = {}, options = {})
req = build_request(:list_channels, params)
req.send_request(options)
end
# Lists information about dataset contents that have been created.
#
# @option params [required, String] :dataset_name
# The name of the dataset whose contents information you want to list.
#
# @option params [String] :next_token
# The token for the next set of results.
#
# @option params [Integer] :max_results
# The maximum number of results to return in this request.
#
# @option params [Time,DateTime,Date,Integer,String] :scheduled_on_or_after
# A filter to limit results to those dataset contents whose creation is
# scheduled on or after the given time. See the field
# `triggers.schedule` in the `CreateDataset` request. (timestamp)
#
# @option params [Time,DateTime,Date,Integer,String] :scheduled_before
# A filter to limit results to those dataset contents whose creation is
# scheduled before the given time. See the field `triggers.schedule` in
# the `CreateDataset` request. (timestamp)
#
# @return [Types::ListDatasetContentsResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
#
# * {Types::ListDatasetContentsResponse#dataset_content_summaries #dataset_content_summaries} => Array<Types::DatasetContentSummary>
# * {Types::ListDatasetContentsResponse#next_token #next_token} => String
#
# The returned {Seahorse::Client::Response response} is a pageable response and is Enumerable. For details on usage see {Aws::PageableResponse PageableResponse}.
#
# @example Request syntax with placeholder values
#
# resp = client.list_dataset_contents({
# dataset_name: "DatasetName", # required
# next_token: "NextToken",
# max_results: 1,
# scheduled_on_or_after: Time.now,
# scheduled_before: Time.now,
# })
#
# @example Response structure
#
# resp.dataset_content_summaries #=> Array
# resp.dataset_content_summaries[0].version #=> String
# resp.dataset_content_summaries[0].status.state #=> String, one of "CREATING", "SUCCEEDED", "FAILED"
# resp.dataset_content_summaries[0].status.reason #=> String
# resp.dataset_content_summaries[0].creation_time #=> Time
# resp.dataset_content_summaries[0].schedule_time #=> Time
# resp.dataset_content_summaries[0].completion_time #=> Time
# resp.next_token #=> String
#
# @overload list_dataset_contents(params = {})
# @param [Hash] params ({})
def list_dataset_contents(params = {}, options = {})
req = build_request(:list_dataset_contents, params)
req.send_request(options)
end
# Retrieves information about datasets.
#
# @option params [String] :next_token
# The token for the next set of results.
#
# @option params [Integer] :max_results
# The maximum number of results to return in this request.
#
# The default value is 100.
#
# @return [Types::ListDatasetsResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
#
# * {Types::ListDatasetsResponse#dataset_summaries #dataset_summaries} => Array<Types::DatasetSummary>
# * {Types::ListDatasetsResponse#next_token #next_token} => String
#
# The returned {Seahorse::Client::Response response} is a pageable response and is Enumerable. For details on usage see {Aws::PageableResponse PageableResponse}.
#
# @example Request syntax with placeholder values
#
# resp = client.list_datasets({
# next_token: "NextToken",
# max_results: 1,
# })
#
# @example Response structure
#
# resp.dataset_summaries #=> Array
# resp.dataset_summaries[0].dataset_name #=> String
# resp.dataset_summaries[0].status #=> String, one of "CREATING", "ACTIVE", "DELETING"
# resp.dataset_summaries[0].creation_time #=> Time
# resp.dataset_summaries[0].last_update_time #=> Time
# resp.dataset_summaries[0].triggers #=> Array
# resp.dataset_summaries[0].triggers[0].schedule.expression #=> String
# resp.dataset_summaries[0].triggers[0].dataset.name #=> String
# resp.dataset_summaries[0].actions #=> Array
# resp.dataset_summaries[0].actions[0].action_name #=> String
# resp.dataset_summaries[0].actions[0].action_type #=> String, one of "QUERY", "CONTAINER"
# resp.next_token #=> String
#
# @overload list_datasets(params = {})
# @param [Hash] params ({})
def list_datasets(params = {}, options = {})
req = build_request(:list_datasets, params)
req.send_request(options)
end
# Retrieves a list of data stores.
#
# @option params [String] :next_token
# The token for the next set of results.
#
# @option params [Integer] :max_results
# The maximum number of results to return in this request.
#
# The default value is 100.
#
# @return [Types::ListDatastoresResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
#
# * {Types::ListDatastoresResponse#datastore_summaries #datastore_summaries} => Array<Types::DatastoreSummary>
# * {Types::ListDatastoresResponse#next_token #next_token} => String
#
# The returned {Seahorse::Client::Response response} is a pageable response and is Enumerable. For details on usage see {Aws::PageableResponse PageableResponse}.
#
# @example Request syntax with placeholder values
#
# resp = client.list_datastores({
# next_token: "NextToken",
# max_results: 1,
# })
#
# @example Response structure
#
# resp.datastore_summaries #=> Array
# resp.datastore_summaries[0].datastore_name #=> String
# resp.datastore_summaries[0].datastore_storage.customer_managed_s3.bucket #=> String
# resp.datastore_summaries[0].datastore_storage.customer_managed_s3.key_prefix #=> String
# resp.datastore_summaries[0].datastore_storage.customer_managed_s3.role_arn #=> String
# resp.datastore_summaries[0].datastore_storage.iot_site_wise_multi_layer_storage.customer_managed_s3_storage.bucket #=> String
# resp.datastore_summaries[0].datastore_storage.iot_site_wise_multi_layer_storage.customer_managed_s3_storage.key_prefix #=> String
# resp.datastore_summaries[0].status #=> String, one of "CREATING", "ACTIVE", "DELETING"
# resp.datastore_summaries[0].creation_time #=> Time
# resp.datastore_summaries[0].last_update_time #=> Time
# resp.datastore_summaries[0].last_message_arrival_time #=> Time
# resp.datastore_summaries[0].file_format_type #=> String, one of "JSON", "PARQUET"
# resp.datastore_summaries[0].datastore_partitions.partitions #=> Array
# resp.datastore_summaries[0].datastore_partitions.partitions[0].attribute_partition.attribute_name #=> String
# resp.datastore_summaries[0].datastore_partitions.partitions[0].timestamp_partition.attribute_name #=> String
# resp.datastore_summaries[0].datastore_partitions.partitions[0].timestamp_partition.timestamp_format #=> String
# resp.next_token #=> String
#
# @overload list_datastores(params = {})
# @param [Hash] params ({})
def list_datastores(params = {}, options = {})
req = build_request(:list_datastores, params)
req.send_request(options)
end
# Retrieves a list of pipelines.
#
# @option params [String] :next_token
# The token for the next set of results.
#
# @option params [Integer] :max_results
# The maximum number of results to return in this request.
#
# The default value is 100.
#
# @return [Types::ListPipelinesResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
#
# * {Types::ListPipelinesResponse#pipeline_summaries #pipeline_summaries} => Array<Types::PipelineSummary>
# * {Types::ListPipelinesResponse#next_token #next_token} => String
#
# The returned {Seahorse::Client::Response response} is a pageable response and is Enumerable. For details on usage see {Aws::PageableResponse PageableResponse}.
#
# @example Request syntax with placeholder values
#
# resp = client.list_pipelines({
# next_token: "NextToken",
# max_results: 1,
# })
#
# @example Response structure
#
# resp.pipeline_summaries #=> Array
# resp.pipeline_summaries[0].pipeline_name #=> String
# resp.pipeline_summaries[0].reprocessing_summaries #=> Array
# resp.pipeline_summaries[0].reprocessing_summaries[0].id #=> String
# resp.pipeline_summaries[0].reprocessing_summaries[0].status #=> String, one of "RUNNING", "SUCCEEDED", "CANCELLED", "FAILED"
# resp.pipeline_summaries[0].reprocessing_summaries[0].creation_time #=> Time
# resp.pipeline_summaries[0].creation_time #=> Time
# resp.pipeline_summaries[0].last_update_time #=> Time
# resp.next_token #=> String
#
# @overload list_pipelines(params = {})
# @param [Hash] params ({})
def list_pipelines(params = {}, options = {})
req = build_request(:list_pipelines, params)
req.send_request(options)
end
# Lists the tags (metadata) that you have assigned to the resource.
#
# @option params [required, String] :resource_arn
# The ARN of the resource whose tags you want to list.
#
# @return [Types::ListTagsForResourceResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
#
# * {Types::ListTagsForResourceResponse#tags #tags} => Array<Types::Tag>
#
# @example Request syntax with placeholder values
#
# resp = client.list_tags_for_resource({
# resource_arn: "ResourceArn", # required
# })
#
# @example Response structure
#
# resp.tags #=> Array
# resp.tags[0].key #=> String
# resp.tags[0].value #=> String
#
# @overload list_tags_for_resource(params = {})
# @param [Hash] params ({})
def list_tags_for_resource(params = {}, options = {})
req = build_request(:list_tags_for_resource, params)
req.send_request(options)
end
# Sets or updates the IoT Analytics logging options.
#
# If you update the value of any `loggingOptions` field, it takes up to
# one minute for the change to take effect. Also, if you change the
# policy attached to the role you specified in the `roleArn` field (for
# example, to correct an invalid policy), it takes up to five minutes
# for that change to take effect.
#
# @option params [required, Types::LoggingOptions] :logging_options
# The new values of the IoT Analytics logging options.
#
# @return [Struct] Returns an empty {Seahorse::Client::Response response}.
#
# @example Request syntax with placeholder values
#
# resp = client.put_logging_options({
# logging_options: { # required
# role_arn: "RoleArn", # required
# level: "ERROR", # required, accepts ERROR
# enabled: false, # required
# },
# })
#
# @overload put_logging_options(params = {})
# @param [Hash] params ({})
def put_logging_options(params = {}, options = {})
req = build_request(:put_logging_options, params)
req.send_request(options)
end
# Simulates the results of running a pipeline activity on a message
# payload.
#
# @option params [required, Types::PipelineActivity] :pipeline_activity
# The pipeline activity that is run. This must not be a channel activity
# or a data store activity because these activities are used in a
# pipeline only to load the original message and to store the (possibly)
# transformed message. If a Lambda activity is specified, only
# short-running Lambda functions (those with a timeout of less than 30
# seconds or less) can be used.
#
# @option params [required, Array<String, StringIO, File>] :payloads
# The sample message payloads on which the pipeline activity is run.
#
# @return [Types::RunPipelineActivityResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
#
# * {Types::RunPipelineActivityResponse#payloads #payloads} => Array<String>
# * {Types::RunPipelineActivityResponse#log_result #log_result} => String
#
# @example Request syntax with placeholder values
#
# resp = client.run_pipeline_activity({
# pipeline_activity: { # required
# channel: {
# name: "ActivityName", # required
# channel_name: "ChannelName", # required
# next: "ActivityName",
# },
# lambda: {
# name: "ActivityName", # required
# lambda_name: "LambdaName", # required
# batch_size: 1, # required
# next: "ActivityName",
# },
# datastore: {
# name: "ActivityName", # required
# datastore_name: "DatastoreName", # required
# },
# add_attributes: {
# name: "ActivityName", # required
# attributes: { # required
# "AttributeName" => "AttributeName",
# },
# next: "ActivityName",
# },
# remove_attributes: {
# name: "ActivityName", # required
# attributes: ["AttributeName"], # required
# next: "ActivityName",
# },
# select_attributes: {
# name: "ActivityName", # required
# attributes: ["AttributeName"], # required
# next: "ActivityName",
# },
# filter: {
# name: "ActivityName", # required
# filter: "FilterExpression", # required
# next: "ActivityName",
# },
# math: {
# name: "ActivityName", # required
# attribute: "AttributeName", # required
# math: "MathExpression", # required
# next: "ActivityName",
# },
# device_registry_enrich: {
# name: "ActivityName", # required
# attribute: "AttributeName", # required
# thing_name: "AttributeName", # required
# role_arn: "RoleArn", # required
# next: "ActivityName",
# },
# device_shadow_enrich: {
# name: "ActivityName", # required
# attribute: "AttributeName", # required
# thing_name: "AttributeName", # required
# role_arn: "RoleArn", # required
# next: "ActivityName",
# },
# },
# payloads: ["data"], # required
# })
#
# @example Response structure
#
# resp.payloads #=> Array
# resp.payloads[0] #=> String
# resp.log_result #=> String
#
# @overload run_pipeline_activity(params = {})
# @param [Hash] params ({})
def run_pipeline_activity(params = {}, options = {})
req = build_request(:run_pipeline_activity, params)
req.send_request(options)
end
# Retrieves a sample of messages from the specified channel ingested
# during the specified timeframe. Up to 10 messages can be retrieved.
#
# @option params [required, String] :channel_name
# The name of the channel whose message samples are retrieved.
#
# @option params [Integer] :max_messages
# The number of sample messages to be retrieved. The limit is 10. The
# default is also 10.
#
# @option params [Time,DateTime,Date,Integer,String] :start_time
# The start of the time window from which sample messages are retrieved.
#
# @option params [Time,DateTime,Date,Integer,String] :end_time
# The end of the time window from which sample messages are retrieved.
#
# @return [Types::SampleChannelDataResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
#
# * {Types::SampleChannelDataResponse#payloads #payloads} => Array<String>
#
# @example Request syntax with placeholder values
#
# resp = client.sample_channel_data({
# channel_name: "ChannelName", # required
# max_messages: 1,
# start_time: Time.now,
# end_time: Time.now,
# })
#
# @example Response structure
#
# resp.payloads #=> Array
# resp.payloads[0] #=> String
#
# @overload sample_channel_data(params = {})
# @param [Hash] params ({})
def sample_channel_data(params = {}, options = {})
req = build_request(:sample_channel_data, params)
req.send_request(options)
end
# Starts the reprocessing of raw message data through the pipeline.
#
# @option params [required, String] :pipeline_name
# The name of the pipeline on which to start reprocessing.
#
# @option params [Time,DateTime,Date,Integer,String] :start_time
# The start time (inclusive) of raw message data that is reprocessed.
#
# If you specify a value for the `startTime` parameter, you must not use
# the `channelMessages` object.
#
# @option params [Time,DateTime,Date,Integer,String] :end_time
# The end time (exclusive) of raw message data that is reprocessed.
#
# If you specify a value for the `endTime` parameter, you must not use
# the `channelMessages` object.
#
# @option params [Types::ChannelMessages] :channel_messages
# Specifies one or more sets of channel messages that you want to
# reprocess.
#
# If you use the `channelMessages` object, you must not specify a value
# for `startTime` and `endTime`.
#
# @return [Types::StartPipelineReprocessingResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
#
# * {Types::StartPipelineReprocessingResponse#reprocessing_id #reprocessing_id} => String
#
# @example Request syntax with placeholder values
#
# resp = client.start_pipeline_reprocessing({
# pipeline_name: "PipelineName", # required
# start_time: Time.now,
# end_time: Time.now,
# channel_messages: {
# s3_paths: ["S3PathChannelMessage"],
# },
# })
#
# @example Response structure
#
# resp.reprocessing_id #=> String
#
# @overload start_pipeline_reprocessing(params = {})
# @param [Hash] params ({})
def start_pipeline_reprocessing(params = {}, options = {})
req = build_request(:start_pipeline_reprocessing, params)
req.send_request(options)
end
# Adds to or modifies the tags of the given resource. Tags are metadata
# that can be used to manage a resource.
#
# @option params [required, String] :resource_arn
# The ARN of the resource whose tags you want to modify.
#
# @option params [required, Array<Types::Tag>] :tags
# The new or modified tags for the resource.
#
# @return [Struct] Returns an empty {Seahorse::Client::Response response}.
#
# @example Request syntax with placeholder values
#
# resp = client.tag_resource({
# resource_arn: "ResourceArn", # required
# tags: [ # required
# {
# key: "TagKey", # required
# value: "TagValue", # required
# },
# ],
# })
#
# @overload tag_resource(params = {})
# @param [Hash] params ({})
def tag_resource(params = {}, options = {})
req = build_request(:tag_resource, params)
req.send_request(options)
end
# Removes the given tags (metadata) from the resource.
#
# @option params [required, String] :resource_arn
# The ARN of the resource whose tags you want to remove.
#
# @option params [required, Array<String>] :tag_keys
# The keys of those tags which you want to remove.
#
# @return [Struct] Returns an empty {Seahorse::Client::Response response}.
#
# @example Request syntax with placeholder values
#
# resp = client.untag_resource({
# resource_arn: "ResourceArn", # required
# tag_keys: ["TagKey"], # required
# })
#
# @overload untag_resource(params = {})
# @param [Hash] params ({})
def untag_resource(params = {}, options = {})
req = build_request(:untag_resource, params)
req.send_request(options)
end
# Used to update the settings of a channel.
#
# @option params [required, String] :channel_name
# The name of the channel to be updated.
#
# @option params [Types::ChannelStorage] :channel_storage
# Where channel data is stored. You can choose one of `serviceManagedS3`
# or `customerManagedS3` storage. If not specified, the default is
# `serviceManagedS3`. You can't change this storage option after the
# channel is created.
#
# @option params [Types::RetentionPeriod] :retention_period
# How long, in days, message data is kept for the channel. The retention
# period can't be updated if the channel's Amazon S3 storage is
# customer-managed.
#
# @return [Struct] Returns an empty {Seahorse::Client::Response response}.
#
# @example Request syntax with placeholder values
#
# resp = client.update_channel({
# channel_name: "ChannelName", # required
# channel_storage: {
# service_managed_s3: {
# },
# customer_managed_s3: {
# bucket: "BucketName", # required
# key_prefix: "S3KeyPrefix",
# role_arn: "RoleArn", # required
# },
# },
# retention_period: {
# unlimited: false,
# number_of_days: 1,
# },
# })
#
# @overload update_channel(params = {})
# @param [Hash] params ({})
def update_channel(params = {}, options = {})
req = build_request(:update_channel, params)
req.send_request(options)
end
# Updates the settings of a dataset.
#
# @option params [required, String] :dataset_name
# The name of the dataset to update.
#
# @option params [required, Array<Types::DatasetAction>] :actions
# A list of `DatasetAction` objects.
#
# @option params [Array<Types::DatasetTrigger>] :triggers
# A list of `DatasetTrigger` objects. The list can be empty or can
# contain up to five `DatasetTrigger` objects.
#
# @option params [Array<Types::DatasetContentDeliveryRule>] :content_delivery_rules
# When dataset contents are created, they are delivered to destinations
# specified here.
#
# @option params [Types::RetentionPeriod] :retention_period
# How long, in days, dataset contents are kept for the dataset.
#
# @option params [Types::VersioningConfiguration] :versioning_configuration
# Optional. How many versions of dataset contents are kept. If not
# specified or set to null, only the latest version plus the latest
# succeeded version (if they are different) are kept for the time period
# specified by the `retentionPeriod` parameter. For more information,
# see [Keeping Multiple Versions of IoT Analytics datasets][1] in the
# *IoT Analytics User Guide*.
#
#
#
# [1]: https://docs.aws.amazon.com/iotanalytics/latest/userguide/getting-started.html#aws-iot-analytics-dataset-versions
#
# @option params [Array<Types::LateDataRule>] :late_data_rules
# A list of data rules that send notifications to CloudWatch, when data
# arrives late. To specify `lateDataRules`, the dataset must use a
# [DeltaTimer][1] filter.
#
#
#
# [1]: https://docs.aws.amazon.com/iotanalytics/latest/APIReference/API_DeltaTime.html
#
# @return [Struct] Returns an empty {Seahorse::Client::Response response}.
#
# @example Request syntax with placeholder values
#
# resp = client.update_dataset({
# dataset_name: "DatasetName", # required
# actions: [ # required
# {
# action_name: "DatasetActionName",
# query_action: {
# sql_query: "SqlQuery", # required
# filters: [
# {
# delta_time: {
# offset_seconds: 1, # required
# time_expression: "TimeExpression", # required
# },
# },
# ],
# },
# container_action: {
# image: "Image", # required
# execution_role_arn: "RoleArn", # required
# resource_configuration: { # required
# compute_type: "ACU_1", # required, accepts ACU_1, ACU_2
# volume_size_in_gb: 1, # required
# },
# variables: [
# {
# name: "VariableName", # required
# string_value: "StringValue",
# double_value: 1.0,
# dataset_content_version_value: {
# dataset_name: "DatasetName", # required
# },
# output_file_uri_value: {
# file_name: "OutputFileName", # required
# },
# },
# ],
# },
# },
# ],
# triggers: [
# {
# schedule: {
# expression: "ScheduleExpression",
# },
# dataset: {
# name: "DatasetName", # required
# },
# },
# ],
# content_delivery_rules: [
# {
# entry_name: "EntryName",
# destination: { # required
# iot_events_destination_configuration: {
# input_name: "IotEventsInputName", # required
# role_arn: "RoleArn", # required
# },
# s3_destination_configuration: {
# bucket: "BucketName", # required
# key: "BucketKeyExpression", # required
# glue_configuration: {
# table_name: "GlueTableName", # required
# database_name: "GlueDatabaseName", # required
# },
# role_arn: "RoleArn", # required
# },
# },
# },
# ],
# retention_period: {
# unlimited: false,
# number_of_days: 1,
# },
# versioning_configuration: {
# unlimited: false,
# max_versions: 1,
# },
# late_data_rules: [
# {
# rule_name: "LateDataRuleName",
# rule_configuration: { # required
# delta_time_session_window_configuration: {
# timeout_in_minutes: 1, # required
# },
# },
# },
# ],
# })
#
# @overload update_dataset(params = {})
# @param [Hash] params ({})
def update_dataset(params = {}, options = {})
req = build_request(:update_dataset, params)
req.send_request(options)
end
# Used to update the settings of a data store.
#
# @option params [required, String] :datastore_name
# The name of the data store to be updated.
#
# @option params [Types::RetentionPeriod] :retention_period
# How long, in days, message data is kept for the data store. The
# retention period can't be updated if the data store's Amazon S3
# storage is customer-managed.
#
# @option params [Types::DatastoreStorage] :datastore_storage
# Where data in a data store is stored.. You can choose
# `serviceManagedS3` storage, `customerManagedS3` storage, or
# `iotSiteWiseMultiLayerStorage` storage. The default is
# `serviceManagedS3`. You can't change the choice of Amazon S3 storage
# after your data store is created.
#
# @option params [Types::FileFormatConfiguration] :file_format_configuration
# Contains the configuration information of file formats. IoT Analytics
# data stores support JSON and [Parquet][1].
#
# The default file format is JSON. You can specify only one format.
#
# You can't change the file format after you create the data store.
#
#
#
# [1]: https://parquet.apache.org/
#
# @return [Struct] Returns an empty {Seahorse::Client::Response response}.
#
# @example Request syntax with placeholder values
#
# resp = client.update_datastore({
# datastore_name: "DatastoreName", # required
# retention_period: {
# unlimited: false,
# number_of_days: 1,
# },
# datastore_storage: {
# service_managed_s3: {
# },
# customer_managed_s3: {
# bucket: "BucketName", # required
# key_prefix: "S3KeyPrefix",
# role_arn: "RoleArn", # required
# },
# iot_site_wise_multi_layer_storage: {
# customer_managed_s3_storage: { # required
# bucket: "BucketName", # required
# key_prefix: "S3KeyPrefix",
# },
# },
# },
# file_format_configuration: {
# json_configuration: {
# },
# parquet_configuration: {
# schema_definition: {
# columns: [
# {
# name: "ColumnName", # required
# type: "ColumnDataType", # required
# },
# ],
# },
# },
# },
# })
#
# @overload update_datastore(params = {})
# @param [Hash] params ({})
def update_datastore(params = {}, options = {})
req = build_request(:update_datastore, params)
req.send_request(options)
end
# Updates the settings of a pipeline. You must specify both a `channel`
# and a `datastore` activity and, optionally, as many as 23 additional
# activities in the `pipelineActivities` array.
#
# @option params [required, String] :pipeline_name
# The name of the pipeline to update.
#
# @option params [required, Array<Types::PipelineActivity>] :pipeline_activities
# A list of `PipelineActivity` objects. Activities perform
# transformations on your messages, such as removing, renaming or adding
# message attributes; filtering messages based on attribute values;
# invoking your Lambda functions on messages for advanced processing; or
# performing mathematical transformations to normalize device data.
#
# The list can be 2-25 `PipelineActivity` objects and must contain both
# a `channel` and a `datastore` activity. Each entry in the list must
# contain only one activity. For example:
#
# `pipelineActivities = [ { "channel": { ... } }, { "lambda": { ... } },
# ... ]`
#
# @return [Struct] Returns an empty {Seahorse::Client::Response response}.
#
# @example Request syntax with placeholder values
#
# resp = client.update_pipeline({
# pipeline_name: "PipelineName", # required
# pipeline_activities: [ # required
# {
# channel: {
# name: "ActivityName", # required
# channel_name: "ChannelName", # required
# next: "ActivityName",
# },
# lambda: {
# name: "ActivityName", # required
# lambda_name: "LambdaName", # required
# batch_size: 1, # required
# next: "ActivityName",
# },
# datastore: {
# name: "ActivityName", # required
# datastore_name: "DatastoreName", # required
# },
# add_attributes: {
# name: "ActivityName", # required
# attributes: { # required
# "AttributeName" => "AttributeName",
# },
# next: "ActivityName",
# },
# remove_attributes: {
# name: "ActivityName", # required
# attributes: ["AttributeName"], # required
# next: "ActivityName",
# },
# select_attributes: {
# name: "ActivityName", # required
# attributes: ["AttributeName"], # required
# next: "ActivityName",
# },
# filter: {
# name: "ActivityName", # required
# filter: "FilterExpression", # required
# next: "ActivityName",
# },
# math: {
# name: "ActivityName", # required
# attribute: "AttributeName", # required
# math: "MathExpression", # required
# next: "ActivityName",
# },
# device_registry_enrich: {
# name: "ActivityName", # required
# attribute: "AttributeName", # required
# thing_name: "AttributeName", # required
# role_arn: "RoleArn", # required
# next: "ActivityName",
# },
# device_shadow_enrich: {
# name: "ActivityName", # required
# attribute: "AttributeName", # required
# thing_name: "AttributeName", # required
# role_arn: "RoleArn", # required
# next: "ActivityName",
# },
# },
# ],
# })
#
# @overload update_pipeline(params = {})
# @param [Hash] params ({})
def update_pipeline(params = {}, options = {})
req = build_request(:update_pipeline, params)
req.send_request(options)
end
# @!endgroup
# @param params ({})
# @api private
def build_request(operation_name, params = {})
handlers = @handlers.for(operation_name)
tracer = config.telemetry_provider.tracer_provider.tracer(
Aws::Telemetry.module_to_tracer_name('Aws::IoTAnalytics')
)
context = Seahorse::Client::RequestContext.new(
operation_name: operation_name,
operation: config.api.operation(operation_name),
client: self,
params: params,
config: config,
tracer: tracer
)
context[:gem_name] = 'aws-sdk-iotanalytics'
context[:gem_version] = '1.80.0'
Seahorse::Client::Request.new(handlers, context)
end
# @api private
# @deprecated
def waiter_names
[]
end
class << self
# @api private
attr_reader :identifier
# @api private
def errors_module
Errors
end
end
end
end